Type
Type
Local
From
From
Alibaba
Quantisation
Quantisation
uint4
Precision
Precision
No
Size
Size
14B
Qwen3-14B-MLX-4bit is a 14.8 billion parameter causal language model from the latest Qwen3 series, optimized for the MLX framework. It represents a significant advancement in reasoning, instruction-following, and multilingual capabilities with a native context length of 32,768 tokens that can be extended to 131,072 tokens using YaRN scaling. The model uniquely supports seamless switching between thinking mode for complex logical reasoning, mathematics, and coding tasks, and non-thinking mode for efficient general-purpose dialogue, all within a single model architecture. This dual-mode capability ensures optimal performance across various scenarios. Qwen3-14B excels in reasoning abilities surpassing previous QwQ and Qwen2.5 models, superior human preference alignment for creative writing and role-playing, strong agent capabilities for precise tool integration, and support for over 100 languages with multilingual instruction-following capabilities.
Qwen3-14B-MLX-4bit is a 14.8 billion parameter causal language model from the latest Qwen3 series, optimized for the MLX framework. It represents a significant advancement in reasoning, instruction-following, and multilingual capabilities with a native context length of 32,768 tokens that can be extended to 131,072 tokens using YaRN scaling. The model uniquely supports seamless switching between thinking mode for complex logical reasoning, mathematics, and coding tasks, and non-thinking mode for efficient general-purpose dialogue, all within a single model architecture. This dual-mode capability ensures optimal performance across various scenarios. Qwen3-14B excels in reasoning abilities surpassing previous QwQ and Qwen2.5 models, superior human preference alignment for creative writing and role-playing, strong agent capabilities for precise tool integration, and support for over 100 languages with multilingual instruction-following capabilities.