LFM2.5-1.2B-Instruct

Run locally Apple devices with Mirai

Type

Type

Local

From

From

LiquidAI

Quantisation

Quantisation

No

Precision

Precision

No

Size

Size

1.2B

Source

Source

Hugging Face Logo

LFM2.5-1.2B-Instruct is a general-purpose instruction-tuned language model from Liquid AI designed for on-device deployment. With 1.17 billion parameters across 16 layers combining double-gated LIV convolution blocks and GQA attention, the model delivers best-in-class performance for its size, rivaling much larger models while running under 1GB of memory. Trained on 28 trillion tokens with extended pre-training and large-scale reinforcement learning, it supports a 32,768 token context length and is multilingual across English, Arabic, Chinese, French, German, Japanese, Korean, and Spanish. The model excels in fast edge inference, achieving 239 tokens per second on AMD CPUs and 82 tokens per second on mobile NPUs, with day-one support for llama.cpp, MLX, and vLLM frameworks. LFM2.5-1.2B-Instruct is recommended for agentic tasks, data extraction, and retrieval-augmented generation, and supports function calling and tool use through a ChatML-like chat template. The model is not recommended for knowledge-intensive tasks and programming.

1
Choose framework
2
Run the following command to install Mirai SDK
SPMhttps://github.com/trymirai/uzu-swift
3
Set Mirai API keyGet API Key
4
Apply code
Loading...

LFM2.5-1.2B-Instruct

Run locally Apple devices with Mirai

Type

Local

From

LiquidAI

Quantisation

No

Precision

float16

Size

1.2B

Source

Hugging Face Logo

LFM2.5-1.2B-Instruct is a general-purpose instruction-tuned language model from Liquid AI designed for on-device deployment. With 1.17 billion parameters across 16 layers combining double-gated LIV convolution blocks and GQA attention, the model delivers best-in-class performance for its size, rivaling much larger models while running under 1GB of memory. Trained on 28 trillion tokens with extended pre-training and large-scale reinforcement learning, it supports a 32,768 token context length and is multilingual across English, Arabic, Chinese, French, German, Japanese, Korean, and Spanish. The model excels in fast edge inference, achieving 239 tokens per second on AMD CPUs and 82 tokens per second on mobile NPUs, with day-one support for llama.cpp, MLX, and vLLM frameworks. LFM2.5-1.2B-Instruct is recommended for agentic tasks, data extraction, and retrieval-augmented generation, and supports function calling and tool use through a ChatML-like chat template. The model is not recommended for knowledge-intensive tasks and programming.

1
Choose framework
2
Run the following command to install Mirai SDK
SPMhttps://github.com/trymirai/uzu-swift
3
Set Mirai API keyGet API Key
4
Apply code
Loading...