Type
Type
Local
From
From
Alibaba
Quantisation
Quantisation
uint4
Precision
Precision
No
Size
Size
14B
Qwen3-14B-AWQ is a 14.8 billion parameter quantized version of Qwen3, a latest generation large language model offering groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support. The model uniquely supports seamless switching between thinking mode for complex logical reasoning, mathematics, and coding, and non-thinking mode for efficient general-purpose dialogue, all within a single model architecture. This 4-bit AWQ quantized variant maintains strong performance while reducing memory requirements, supporting a native context length of 32,768 tokens, with capability to extend to 131,072 tokens using YaRN scaling. The model excels in reasoning capabilities, human preference alignment for creative writing and multi-turn dialogue, agent-based tool integration, and supports over 100 languages with strong multilingual instruction-following abilities.
Qwen3-14B-AWQ is a 14.8 billion parameter quantized version of Qwen3, a latest generation large language model offering groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support. The model uniquely supports seamless switching between thinking mode for complex logical reasoning, mathematics, and coding, and non-thinking mode for efficient general-purpose dialogue, all within a single model architecture. This 4-bit AWQ quantized variant maintains strong performance while reducing memory requirements, supporting a native context length of 32,768 tokens, with capability to extend to 131,072 tokens using YaRN scaling. The model excels in reasoning capabilities, human preference alignment for creative writing and multi-turn dialogue, agent-based tool integration, and supports over 100 languages with strong multilingual instruction-following abilities.