This is an 8-bit quantized version of the LFM2-350M model converted to MLX format for efficient inference on Apple Silicon devices. The model is based on LiquidAI's LFM2-350M and supports text generation across multiple languages including English, Arabic, Chinese, French, German, Japanese, Korean, and Spanish. It is designed for edge deployment with reduced memory requirements through 8-bit quantization while maintaining the capabilities of the original Liquid Foundation Model 2.
LiquidAI
available local models on Mirai:
available local models on Mirai:
Name
Quantisation
Size
LFM2-1.2B
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-2.6B
uint8
2.6B
Quant.
uint8
Size
2.6B
LFM2-350M
uint8
350M
Quant.
uint8
Size
350M
LFM2-700M
uint8
700M
Quant.
uint8
Size
700M
LFM2.5-1.2B-Instruct
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Instruct-MLX-4bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Instruct-MLX-8bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Thinking
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-1.2B-4bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-1.2B-8bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-2.6B-4bit
uint8
2.6B
Quant.
uint8
Size
2.6B
LFM2-2.6B-8bit
uint8
2.6B
Quant.
uint8
Size
2.6B
LFM2-350M-4bit
uint8
350M
Quant.
uint8
Size
350M
LFM2-350M-8bit
uint8
350M
Quant.
uint8
Size
350M
LFM2-700M-4bit
uint8
700M
Quant.
uint8
Size
700M
LFM2-700M-8bit
uint8
700M
Quant.
uint8
Size
700M
LFM2.5-1.2B-Thinking-4bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Thinking-8bit
uint8
1.2B
Quant.
uint8
Size
1.2B
This is an 8-bit quantized version of the LFM2-350M model converted to MLX format for efficient inference on Apple Silicon devices. The model is based on LiquidAI's LFM2-350M and supports text generation across multiple languages including English, Arabic, Chinese, French, German, Japanese, Korean, and Spanish. It is designed for edge deployment with reduced memory requirements through 8-bit quantization while maintaining the capabilities of the original Liquid Foundation Model 2.
LiquidAI
available local models on Mirai:
Name
Quantisation
Size
LFM2-1.2B
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-2.6B
uint8
2.6B
Quant.
uint8
Size
2.6B
LFM2-350M
uint8
350M
Quant.
uint8
Size
350M
LFM2-700M
uint8
700M
Quant.
uint8
Size
700M
LFM2.5-1.2B-Instruct
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Instruct-MLX-4bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Instruct-MLX-8bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Thinking
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-1.2B-4bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-1.2B-8bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2-2.6B-4bit
uint8
2.6B
Quant.
uint8
Size
2.6B
LFM2-2.6B-8bit
uint8
2.6B
Quant.
uint8
Size
2.6B
LFM2-350M-4bit
uint8
350M
Quant.
uint8
Size
350M
LFM2-350M-8bit
uint8
350M
Quant.
uint8
Size
350M
LFM2-700M-4bit
uint8
700M
Quant.
uint8
Size
700M
LFM2-700M-8bit
uint8
700M
Quant.
uint8
Size
700M
LFM2.5-1.2B-Thinking-4bit
uint8
1.2B
Quant.
uint8
Size
1.2B
LFM2.5-1.2B-Thinking-8bit
uint8
1.2B
Quant.
uint8
Size
1.2B