Qwen2.5-Coder is the latest series of code-specific Qwen large language models covering six mainstream model sizes from 0.5 to 32 billion parameters. This instruction-tuned 32B variant brings significant improvements in code generation, code reasoning, and code fixing, with training scaled up to 5.5 trillion tokens including source code, text-code grounding, and synthetic data. The model achieves state-of-the-art open-source code LLM performance with coding abilities matching those of GPT-4o, while also maintaining strengths in mathematics and general competencies to support real-world applications such as code agents. Qwen2.5-Coder-32B-Instruct is a causal language model with 32.5 billion parameters built on a transformer architecture with RoPE, SwiGLU, RMSNorm, and attention QKV bias. It supports long-context processing up to 128K tokens and includes 64 layers with grouped query attention using 40 heads for queries and 8 heads for key-value pairs. The model combines pretraining and post-training stages to provide a comprehensive foundation for practical coding applications.
Alibaba
available local models on Mirai:
available local models on Mirai:
Name
Quantisation
Size
Qwen2.5-Coder-0.5B-Instruct
No
0.5B
Quant.
No
Size
0.5B
Qwen2.5-Coder-1.5B-Instruct
No
1.5B
Quant.
No
Size
1.5B
Qwen2.5-Coder-14B-Instruct
No
14B
Quant.
No
Size
14B
Qwen2.5-Coder-32B-Instruct
No
32B
Quant.
No
Size
32B
Qwen2.5-Coder-3B-Instruct
No
3B
Quant.
No
Size
3B
Qwen2.5-Coder-7B-Instruct
No
7B
Quant.
No
Size
7B
Qwen3-0.6B
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-4bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-8bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-1.7B
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-4bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-8bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-14B
No
14B
Quant.
No
Size
14B
Qwen3-14B-AWQ
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-4bit
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-8bit
No
14B
Quant.
No
Size
14B
Qwen3-32B
No
32B
Quant.
No
Size
32B
Qwen3-32B-AWQ
No
32B
Quant.
No
Size
32B
Qwen3-32B-MLX-4bit
No
32B
Quant.
No
Size
32B
Qwen3-4B
No
4B
Quant.
No
Size
4B
Qwen2.5-Coder is the latest series of code-specific Qwen large language models covering six mainstream model sizes from 0.5 to 32 billion parameters. This instruction-tuned 32B variant brings significant improvements in code generation, code reasoning, and code fixing, with training scaled up to 5.5 trillion tokens including source code, text-code grounding, and synthetic data. The model achieves state-of-the-art open-source code LLM performance with coding abilities matching those of GPT-4o, while also maintaining strengths in mathematics and general competencies to support real-world applications such as code agents. Qwen2.5-Coder-32B-Instruct is a causal language model with 32.5 billion parameters built on a transformer architecture with RoPE, SwiGLU, RMSNorm, and attention QKV bias. It supports long-context processing up to 128K tokens and includes 64 layers with grouped query attention using 40 heads for queries and 8 heads for key-value pairs. The model combines pretraining and post-training stages to provide a comprehensive foundation for practical coding applications.
Alibaba
available local models on Mirai:
Name
Quantisation
Size
Qwen2.5-Coder-0.5B-Instruct
No
0.5B
Quant.
No
Size
0.5B
Qwen2.5-Coder-1.5B-Instruct
No
1.5B
Quant.
No
Size
1.5B
Qwen2.5-Coder-14B-Instruct
No
14B
Quant.
No
Size
14B
Qwen2.5-Coder-32B-Instruct
No
32B
Quant.
No
Size
32B
Qwen2.5-Coder-3B-Instruct
No
3B
Quant.
No
Size
3B
Qwen2.5-Coder-7B-Instruct
No
7B
Quant.
No
Size
7B
Qwen3-0.6B
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-4bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-8bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-1.7B
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-4bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-8bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-14B
No
14B
Quant.
No
Size
14B
Qwen3-14B-AWQ
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-4bit
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-8bit
No
14B
Quant.
No
Size
14B
Qwen3-32B
No
32B
Quant.
No
Size
32B
Qwen3-32B-AWQ
No
32B
Quant.
No
Size
32B
Qwen3-32B-MLX-4bit
No
32B
Quant.
No
Size
32B
Qwen3-4B
No
4B
Quant.
No
Size
4B