Remote / SF / Europe
Full Time
Models & Research
Remote / SF / Europe
Full Time
Models & Research
Machine Learning Engineer
Machine Learning Engineer
Join a small, senior team building the fastest on-device AI inference engine. Powering real products, not demos
Join a small, senior team building the fastest on-device AI inference engine. Powering real products, not demos.
The role
We’re hiring a Machine Learning Engineer to work on lightweight text classification and embedding models operating on edge devices under tight memory and latency constraints. You will own the entire model development lifecycle, from data labeling to model training and deployment.
What you’ll do
Design evaluation metrics and benchmarks.
Build data processing and labeling pipelines.
Experiment with model architectures and training methods.
Contribute to our training and model optimization framework
What we’re looking for
Strong background in Deep Learning, with deep knowledge of modern text foundation models
Solid grasp of statistics and probability theory
Understanding of ML performance optimization landscape
Experience building complex training pipelines from scratch
Very strong software engineering skills.
Bonus:
Experience with JAX
Experience with GPU programming
Knowledge of Rust
You are probably a good fit if you have a thorough understanding of such concepts as:
muP
Muon
FSDP
FlashAttention
SigLIP
Why join?
You’ll work on applied research that directly impacts how AI systems operate in real-world environments. Not just benchmarks.
We value technical depth, fast iteration, and clarity of thought.
Competitive compensation + meaningful equity.
About us
Mirai builds advanced on-device AI systems. We operate at the intersection of model design and hardware-aware optimization.
Our Models & Research team focuses on making small, efficient models reliable under strict real-world constraints.
Why us?
Founded by proven entrepreneurs who built and scaled consumer AI leaders like Reface (200M+ users) and Prisma (100M+ users).
Our team is small (14 people), senior, and deeply technical. We ship fast and own problems end-to-end.
We’re advised by a former Apple Distinguished Engineer who worked on MLX, and backed by leading AI-focused funds and individuals.

The role
We’re hiring a Machine Learning Engineer to work on lightweight text classification and embedding models operating on edge devices under tight memory and latency constraints. You will own the entire model development lifecycle, from data labeling to model training and deployment.
What you’ll do
Design evaluation metrics and benchmarks.
Build data processing and labeling pipelines.
Experiment with model architectures and training methods.
Contribute to our training and model optimization framework
What we’re looking for
Strong background in Deep Learning, with deep knowledge of modern text foundation models
Solid grasp of statistics and probability theory
Understanding of ML performance optimization landscape
Experience building complex training pipelines from scratch
Very strong software engineering skills.
Bonus:
Experience with JAX
Experience with GPU programming
Knowledge of Rust
You are probably a good fit if you have a thorough understanding of such concepts as:
muP
Muon
FSDP
FlashAttention
SigLIP
Why join?
You’ll work on applied research that directly impacts how AI systems operate in real-world environments. Not just benchmarks.
We value technical depth, fast iteration, and clarity of thought.
Competitive compensation + meaningful equity.
About us
Mirai builds advanced on-device AI systems. We operate at the intersection of model design and hardware-aware optimization.
Our Models & Research team focuses on making small, efficient models reliable under strict real-world constraints.
Why us?
Founded by proven entrepreneurs who built and scaled consumer AI leaders like Reface (200M+ users) and Prisma (100M+ users).
Our team is small (14 people), senior, and deeply technical. We ship fast and own problems end-to-end.
We’re advised by a former Apple Distinguished Engineer who worked on MLX, and backed by leading AI-focused funds and individuals.

Interested?
Join a small, senior team building the fastest on-device AI inference engine. Powering real products, not demos.