This model is a conversion of Google's Gemma 3 4B instruction-tuned variant to MLX format, quantized to 8-bit precision. It is a multimodal model capable of processing both images and text inputs to generate text responses. The model was converted using mlx-vlm version 0.1.18 and is optimized for use with the MLX framework on Apple Silicon hardware.
available local models on Mirai:
available local models on Mirai:
Name
Quantisation
Size
gemma-3-1b-it
uint8
1B
Quant.
uint8
Size
1B
gemma-3-27b-it
uint8
27B
Quant.
uint8
Size
27B
gemma-3-4b-it
uint8
4B
Quant.
uint8
Size
4B
gemma-3-1b-it-4bit
uint8
1B
Quant.
uint8
Size
1B
gemma-3-1b-it-8bit
uint8
1B
Quant.
uint8
Size
1B
gemma-3-27b-it-4bit
uint8
27B
Quant.
uint8
Size
27B
gemma-3-27b-it-8bit
uint8
27B
Quant.
uint8
Size
27B
gemma-3-4b-it-4bit
uint8
4B
Quant.
uint8
Size
4B
gemma-3-4b-it-8bit
uint8
4B
Quant.
uint8
Size
4B
This model is a conversion of Google's Gemma 3 4B instruction-tuned variant to MLX format, quantized to 8-bit precision. It is a multimodal model capable of processing both images and text inputs to generate text responses. The model was converted using mlx-vlm version 0.1.18 and is optimized for use with the MLX framework on Apple Silicon hardware.
available local models on Mirai:
Name
Quantisation
Size
gemma-3-1b-it
uint8
1B
Quant.
uint8
Size
1B
gemma-3-27b-it
uint8
27B
Quant.
uint8
Size
27B
gemma-3-4b-it
uint8
4B
Quant.
uint8
Size
4B
gemma-3-1b-it-4bit
uint8
1B
Quant.
uint8
Size
1B
gemma-3-1b-it-8bit
uint8
1B
Quant.
uint8
Size
1B
gemma-3-27b-it-4bit
uint8
27B
Quant.
uint8
Size
27B
gemma-3-27b-it-8bit
uint8
27B
Quant.
uint8
Size
27B
gemma-3-4b-it-4bit
uint8
4B
Quant.
uint8
Size
4B
gemma-3-4b-it-8bit
uint8
4B
Quant.
uint8
Size
4B