Qwen3 is the latest generation of large language models in the Qwen series, offering both dense and mixture-of-experts models built upon extensive training. The series delivers significant advancements in reasoning, instruction-following, agent capabilities, and multilingual support. Qwen3-0.6B is a compact causal language model with 0.6 billion parameters featuring a unique ability to seamlessly switch between thinking mode for complex logical reasoning, mathematics, and coding tasks, and non-thinking mode for efficient general-purpose dialogue. The model includes 28 layers with grouped query attention and supports a context length of 32,768 tokens. It significantly enhances reasoning capabilities beyond previous Qwen models, excels in human preference alignment with strengths in creative writing and multi-turn dialogues, demonstrates strong agent capabilities for tool integration in both modes, and provides support for over 100 languages and dialects with strong multilingual instruction-following and translation abilities.
Alibaba
available local models on Mirai:
available local models on Mirai:
Name
Quantisation
Size
Qwen2.5-Coder-0.5B-Instruct
No
0.5B
Quant.
No
Size
0.5B
Qwen2.5-Coder-1.5B-Instruct
No
1.5B
Quant.
No
Size
1.5B
Qwen2.5-Coder-14B-Instruct
No
14B
Quant.
No
Size
14B
Qwen2.5-Coder-32B-Instruct
No
32B
Quant.
No
Size
32B
Qwen2.5-Coder-3B-Instruct
No
3B
Quant.
No
Size
3B
Qwen2.5-Coder-7B-Instruct
No
7B
Quant.
No
Size
7B
Qwen3-0.6B
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-4bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-8bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-1.7B
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-4bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-8bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-14B
No
14B
Quant.
No
Size
14B
Qwen3-14B-AWQ
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-4bit
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-8bit
No
14B
Quant.
No
Size
14B
Qwen3-32B
No
32B
Quant.
No
Size
32B
Qwen3-32B-AWQ
No
32B
Quant.
No
Size
32B
Qwen3-32B-MLX-4bit
No
32B
Quant.
No
Size
32B
Qwen3-4B
No
4B
Quant.
No
Size
4B
Qwen3 is the latest generation of large language models in the Qwen series, offering both dense and mixture-of-experts models built upon extensive training. The series delivers significant advancements in reasoning, instruction-following, agent capabilities, and multilingual support. Qwen3-0.6B is a compact causal language model with 0.6 billion parameters featuring a unique ability to seamlessly switch between thinking mode for complex logical reasoning, mathematics, and coding tasks, and non-thinking mode for efficient general-purpose dialogue. The model includes 28 layers with grouped query attention and supports a context length of 32,768 tokens. It significantly enhances reasoning capabilities beyond previous Qwen models, excels in human preference alignment with strengths in creative writing and multi-turn dialogues, demonstrates strong agent capabilities for tool integration in both modes, and provides support for over 100 languages and dialects with strong multilingual instruction-following and translation abilities.
Alibaba
available local models on Mirai:
Name
Quantisation
Size
Qwen2.5-Coder-0.5B-Instruct
No
0.5B
Quant.
No
Size
0.5B
Qwen2.5-Coder-1.5B-Instruct
No
1.5B
Quant.
No
Size
1.5B
Qwen2.5-Coder-14B-Instruct
No
14B
Quant.
No
Size
14B
Qwen2.5-Coder-32B-Instruct
No
32B
Quant.
No
Size
32B
Qwen2.5-Coder-3B-Instruct
No
3B
Quant.
No
Size
3B
Qwen2.5-Coder-7B-Instruct
No
7B
Quant.
No
Size
7B
Qwen3-0.6B
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-4bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-8bit
No
0.6B
Quant.
No
Size
0.6B
Qwen3-1.7B
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-4bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-8bit
No
1.7B
Quant.
No
Size
1.7B
Qwen3-14B
No
14B
Quant.
No
Size
14B
Qwen3-14B-AWQ
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-4bit
No
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-8bit
No
14B
Quant.
No
Size
14B
Qwen3-32B
No
32B
Quant.
No
Size
32B
Qwen3-32B-AWQ
No
32B
Quant.
No
Size
32B
Qwen3-32B-MLX-4bit
No
32B
Quant.
No
Size
32B
Qwen3-4B
No
4B
Quant.
No
Size
4B