gemma-3-1b-it-8bit

Run locally Apple devices with Mirai

Type

Type

Local

From

From

Google

Quantisation

Quantisation

uint8

Precision

Precision

No

Size

Size

1B

Source

Source

Hugging Face Logo

This is an 8-bit quantized version of Google's Gemma 3 1B instruction-tuned model converted to MLX format for efficient inference on Apple Silicon and other supported hardware. The model is based on the original google/gemma-3-1b-it and maintains its instruction-following capabilities while being optimized for reduced memory usage and faster inference through quantization.

This is an 8-bit quantized version of Google's Gemma 3 1B instruction-tuned model converted to MLX format for efficient inference on Apple Silicon and other supported hardware. The model is based on the original google/gemma-3-1b-it and maintains its instruction-following capabilities while being optimized for reduced memory usage and faster inference through quantization.

gemma-3-1b-it-8bit

Run locally Apple devices with Mirai

Type

Local

From

Google

Quantisation

uint8

Precision

float16

Size

1B

Source

Hugging Face Logo

This is an 8-bit quantized version of Google's Gemma 3 1B instruction-tuned model converted to MLX format for efficient inference on Apple Silicon and other supported hardware. The model is based on the original google/gemma-3-1b-it and maintains its instruction-following capabilities while being optimized for reduced memory usage and faster inference through quantization.