Inference Engine
The fastest inference runtime for iPhone, iPad and Mac.
Inference Engine
Optimize and run your model on every Apple device.
Up to 38% faster prompt processing vs MLX.
Optimize and run your model on every Apple device. Up to 38% faster prompt processing vs MLX.


Run your model on 2 billion Apple devices. Perfect for:
Model companies.
You train and ship models. Mirai optimizes them for Apple Silicon, benchmarks on real hardware, and distributes.
AI researchers & labs.
Mirai converts your model and puts it in front of real users on Apple devices, not just leaderboards.
Independent makers.
You're fine-tuning or training from scratch. Mirai gives your model the same device reach as OpenAI and DeepSeek.
Model companies.
You train and ship models. Mirai optimizes them for Apple Silicon, benchmarks on real hardware, and distributes.
AI researchers & labs.
Mirai converts your model and puts it in front of real users on Apple devices, not just leaderboards.
Independent makers.
You're fine-tuning or training from scratch. Mirai gives your model the same device reach as OpenAI and DeepSeek.
What Apple Silicon delivers today with Mirai.
Convert. Integrate. Run.


One inference engine. Integrate from any language.
Language
Distribution
Language
Distribution
Rust
Cargo
Rust
Cargo
Swift
Swift Package Manager
Swift
Swift Package Manager
TypeScript
NPM (Node.js)
TypeScript
NPM (Node.js)
Kotlin
Coming Soon
Kotlin
Coming Soon
Python
Coming Soon
Python
Coming Soon
Same high-level API across all languages.
Full performance of the Rust core from every language.
Convert once, integrate anywhere.
Built-in features every model gets automatically:
Speculative decoding.
A draft model predicts tokens ahead, your model verifies in one pass. Up to 2x faster generation.
Structured output.
Task-specific sessions.
Built-in performance metrics.
Speculative decoding
Structured output
Task-specific sessions
Built-in performance metrics


Supported models
Supported models:
Common questions:
How does model support work?
How does model support work?
What architectures are supported?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
What architectures are supported?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
How does Mirai compare to other inference engines?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
How does Mirai compare to other inference engines?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
What is the maximum supported model size?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
What is the maximum supported model size?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
How can I run benchmarks myself?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
How can I run benchmarks myself?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
How can we discuss a specific use case?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
How can we discuss a specific use case?
Framer is a design tool that allows you to design websites on a freeform canvas, and then publish them as websites with a single click.
