About Us
We are 12-person team, who under a year, built a complete on-device inference stack, from model optimization and export tooling to a proprietary runtime and deployment layer.
On supported models, Mirai outperforms MLX and llama.cpp, while staying production-ready for real applications.
We're not building demos. We're not optimizing for benchmarks that exist in some airless vacuum. We're making local inference something you can ship.
Mirai is founded by proven entrepreneurs who built and scaled consumer AI leaders like Reface (200M+ users, backed by Andreessen Horowitz) and Prisma (100M+ users).
Our team is small (12 people), senior, and deeply technical. We ship fast and own problems end-to-end.
We’re advised by a former Apple Distinguished Engineer who worked on MLX, and backed by leading AI-focused funds and individuals.


Turn our Apple Silicon technical lead into market dominance.
We are focusing on:
Maintaining a clear performance lead over open stacks.
Expanding model support without sacrificing speed or reliability.
Building world-class developer tooling, documentation, and benchmarks.
Powering companies where latency, cost, and privacy actually matter.
For AI to work seamlessly, the core must live on device.
As AI becomes a core part of software, inference can’t rely entirely on the cloud. Latency, privacy, cost, and reliability require intelligence to run where users are.
We believe the next generation of software will be built on a new system layer. Not just models, not just runtimes, but a tightly integrated stack that makes intelligence native to the device.
Mirai is building that layer: an LLM OS that combines optimized inference, models, and deployment into a reliable on-device foundation for intelligent software.
AI in a next 10 years will make reality that everyone will make software for themselves. We will not need apps, we will need just a screen or surface where we can collaborate with machines and services.
