Qwen3 is the latest generation of large language models in the Qwen series, offering a comprehensive suite of dense and mixture-of-experts models. The 8B variant is a causal language model with 8.2B parameters, 36 layers, and native support for context lengths up to 32,768 tokens, extendable to 131,072 tokens using YaRN scaling. Qwen3-8B features uniquely seamless switching between thinking mode for complex logical reasoning, mathematics, and coding, and non-thinking mode for efficient general-purpose dialogue within a single model. The model significantly enhances reasoning capabilities, surpassing previous QwQ and Qwen2.5 instruct models on mathematics, code generation, and commonsense reasoning. It demonstrates superior human preference alignment for creative writing, role-playing, multi-turn dialogues, and instruction following, while excelling in agent capabilities with precise tool integration in both thinking and non-thinking modes. The model supports over 100 languages and dialects with strong multilingual instruction following and translation capabilities.
Alibaba
available local models on Mirai:
available local models on Mirai:
Name
Size
Qwen2.5-Coder-0.5B-Instruct
0.5B
Quant.
No
Size
0.5B
Qwen2.5-Coder-1.5B-Instruct
1.5B
Quant.
No
Size
1.5B
Qwen2.5-Coder-14B-Instruct
14B
Quant.
No
Size
14B
Qwen2.5-Coder-32B-Instruct
32B
Quant.
No
Size
32B
Qwen2.5-Coder-3B-Instruct
3B
Quant.
No
Size
3B
Qwen2.5-Coder-7B-Instruct
7B
Quant.
No
Size
7B
Qwen3-0.6B
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-4bit
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-8bit
0.6B
Quant.
No
Size
0.6B
Qwen3-1.7B
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-4bit
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-8bit
1.7B
Quant.
No
Size
1.7B
Qwen3-14B
14B
Quant.
No
Size
14B
Qwen3-14B-AWQ
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-4bit
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-8bit
14B
Quant.
No
Size
14B
Qwen3-32B
32B
Quant.
No
Size
32B
Qwen3-32B-AWQ
32B
Quant.
No
Size
32B
Qwen3-32B-MLX-4bit
32B
Quant.
No
Size
32B
Qwen3-4B
4B
Quant.
No
Size
4B
Qwen3 is the latest generation of large language models in the Qwen series, offering a comprehensive suite of dense and mixture-of-experts models. The 8B variant is a causal language model with 8.2B parameters, 36 layers, and native support for context lengths up to 32,768 tokens, extendable to 131,072 tokens using YaRN scaling. Qwen3-8B features uniquely seamless switching between thinking mode for complex logical reasoning, mathematics, and coding, and non-thinking mode for efficient general-purpose dialogue within a single model. The model significantly enhances reasoning capabilities, surpassing previous QwQ and Qwen2.5 instruct models on mathematics, code generation, and commonsense reasoning. It demonstrates superior human preference alignment for creative writing, role-playing, multi-turn dialogues, and instruction following, while excelling in agent capabilities with precise tool integration in both thinking and non-thinking modes. The model supports over 100 languages and dialects with strong multilingual instruction following and translation capabilities.
Alibaba
available local models on Mirai:
Name
Size
Qwen2.5-Coder-0.5B-Instruct
0.5B
Quant.
No
Size
0.5B
Qwen2.5-Coder-1.5B-Instruct
1.5B
Quant.
No
Size
1.5B
Qwen2.5-Coder-14B-Instruct
14B
Quant.
No
Size
14B
Qwen2.5-Coder-32B-Instruct
32B
Quant.
No
Size
32B
Qwen2.5-Coder-3B-Instruct
3B
Quant.
No
Size
3B
Qwen2.5-Coder-7B-Instruct
7B
Quant.
No
Size
7B
Qwen3-0.6B
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-4bit
0.6B
Quant.
No
Size
0.6B
Qwen3-0.6B-MLX-8bit
0.6B
Quant.
No
Size
0.6B
Qwen3-1.7B
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-4bit
1.7B
Quant.
No
Size
1.7B
Qwen3-1.7B-MLX-8bit
1.7B
Quant.
No
Size
1.7B
Qwen3-14B
14B
Quant.
No
Size
14B
Qwen3-14B-AWQ
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-4bit
14B
Quant.
No
Size
14B
Qwen3-14B-MLX-8bit
14B
Quant.
No
Size
14B
Qwen3-32B
32B
Quant.
No
Size
32B
Qwen3-32B-AWQ
32B
Quant.
No
Size
32B
Qwen3-32B-MLX-4bit
32B
Quant.
No
Size
32B
Qwen3-4B
4B
Quant.
No
Size
4B