MINT quantized Nemotron-3-Super-120B — hybrid Mamba-MoE-Attention (MLX & GGUF)
AI & ML interests
Model Quantization
Recent Activity
View all activity
SWAN quantized versions of GLM models (MLX)
MINT & SWAN quantized versions of Llama 4 Scout and Maverick (MLX & GGUF)
-
baa-ai/Llama-4-Maverick-17B-128E-Instruct-MINT-407GB-MLX
Text Generation • 401B • Updated • 252 -
baa-ai/Llama-4-Maverick-17B-128E-Instruct-MINT-407GB-GGUF
Text Generation • 401B • Updated • 43 -
baa-ai/Llama-4-Maverick-17B-128E-Instruct-SWAN-4bit-MLX
51B • Updated • 247 -
baa-ai/Llama-4-Scout-17B-16E-Instruct-MINT-117GB-MLX
Text Generation • 108B • Updated • 411
MINT quantized versions of Qwen3.5-35B-A3B at multiple budget targets (MLX & GGUF)
MINT & SWAN quantized versions of Qwen3.5-397B-A17B (MLX & GGUF)
MINT quantized Nemotron-3-Super-120B — hybrid Mamba-MoE-Attention (MLX & GGUF)
MINT & SWAN quantized versions of MiniMax-M2.5 (MLX & GGUF)
SWAN quantized versions of GLM models (MLX)
SWAN quantized versions of Llama 3.1 and 3.3 70B Instruct (MLX)
MINT & SWAN quantized versions of Llama 4 Scout and Maverick (MLX & GGUF)
-
baa-ai/Llama-4-Maverick-17B-128E-Instruct-MINT-407GB-MLX
Text Generation • 401B • Updated • 252 -
baa-ai/Llama-4-Maverick-17B-128E-Instruct-MINT-407GB-GGUF
Text Generation • 401B • Updated • 43 -
baa-ai/Llama-4-Maverick-17B-128E-Instruct-SWAN-4bit-MLX
51B • Updated • 247 -
baa-ai/Llama-4-Scout-17B-16E-Instruct-MINT-117GB-MLX
Text Generation • 108B • Updated • 411
MINT & SWAN quantized versions of Qwen3 models (MLX)
MINT quantized versions of Qwen3.5-35B-A3B at multiple budget targets (MLX & GGUF)
MINT quantized versions of Qwen3.5-122B-A10B at multiple budget targets (MLX & GGUF)
MINT & SWAN quantized versions of Qwen3.5-397B-A17B (MLX & GGUF)