Worldwide
Full Time
Worldwide
Full Time
Inference Engineer
Inference Engineer
Join a small, senior team building the fastest on-device AI inference engine. Powering real products, not demos
Join a small, senior team building the fastest on-device AI inference engine. Powering real products, not demos.
About us
Mirai is building the on-device inference layer for AI.
We enable model makers and product developers to run AI models directly on edge devices, starting with Apple Silicon and expanding to Android and beyond.
Our stack spans from low-level GPU kernels to high-level model conversion tools.
We're a small team obsessed with performance, working at the intersection of systems programming and machine learning research.
The role
We're looking for engineers who can bridge the gap between ML research and high-performance inference.
You'll work across our inference engine and model conversion toolkit, implementing new model architectures, supporting new modalities, writing optimized kernels, and building a wide range of features such as function calling and batch decoding.
This role is ideal for someone who reads papers for fun, enjoys writing high-performance code, and gets excited about constant learning.
Nobody knows everything. We'd rather you know one area deeply than everything superficially. If you're good at least in a couple of these areas, you're a great fit:
JAX / Equinox / Pallas stack
Rust systems programming with a focus on developer experience
Writing Metal / Vulkan kernels
Neural codecs and voice model architectures
Trellis-based quantization approaches
Advanced speculative decoding methods, such as EAGLE
Deep understanding of Transformer / SSM / Diffusion / Vision language models
Benchmarking inference performance and model quality
Strong linear algebra, optimization methods, and probability theory
And of course, basic engineering skills, we will ship a lot of code 🙃
We welcome applications from students and early-career engineers. If you've participated in projects that demonstrate systems thinking and ML understanding, we want to hear from you!
Why us?
Founded by proven entrepreneurs who built and scaled consumer AI leaders like Reface (200M+ users) and Prisma (100M+ users).
Our team is small (12 people), senior, and deeply technical. We ship fast and own problems end-to-end.
We’re advised by a former Apple Distinguished Engineer who worked on MLX, and backed by leading AI-focused funds and individuals.

About us
Mirai is building the on-device inference layer for AI.
We enable model makers and product developers to run AI models directly on edge devices, starting with Apple Silicon and expanding to Android and beyond.
Our stack spans from low-level GPU kernels to high-level model conversion tools.
We're a small team obsessed with performance, working at the intersection of systems programming and machine learning research.
The role
We're looking for engineers who can bridge the gap between ML research and high-performance inference.
You'll work across our inference engine and model conversion toolkit, implementing new model architectures, supporting new modalities, writing optimized kernels, and building a wide range of features such as function calling and batch decoding.
This role is ideal for someone who reads papers for fun, enjoys writing high-performance code, and gets excited about constant learning.
Nobody knows everything. We'd rather you know one area deeply than everything superficially. If you're good at least in a couple of these areas, you're a great fit:
JAX / Equinox / Pallas stack
Rust systems programming with a focus on developer experience
Writing Metal / Vulkan kernels
Neural codecs and voice model architectures
Trellis-based quantization approaches
Advanced speculative decoding methods, such as EAGLE
Deep understanding of Transformer / SSM / Diffusion / Vision language models
Benchmarking inference performance and model quality
Strong linear algebra, optimization methods, and probability theory
And of course, basic engineering skills, we will ship a lot of code 🙃
We welcome applications from students and early-career engineers. If you've participated in projects that demonstrate systems thinking and ML understanding, we want to hear from you!
Why us?
Founded by proven entrepreneurs who built and scaled consumer AI leaders like Reface (200M+ users) and Prisma (100M+ users).
Our team is small (12 people), senior, and deeply technical. We ship fast and own problems end-to-end.
We’re advised by a former Apple Distinguished Engineer who worked on MLX, and backed by leading AI-focused funds and individuals.

Why join us?
Impactful Work
You’ll work on core infrastructure that directly shapes how AI runs on billions of devices. Not demos, not prototypes, but production systems.
Career Growth
You’ll take ownership of complex, low-level systems early, and grow alongside a team that has already shipped and scaled AI products.
Collaborative Team
We’re a small, highly collaborative team. No silos, no layers. Just smart people solving hard problems together.
Technology
You’ll work on model optimization, inference runtimes, deployment tooling, and performance-critical systems. Setting new standards for on-device AI.