Mirai
Category: AI Infrastructure
An on-device AI inference platform and SDK that enables developers to deploy and run AI models directly on Apple devices (iPhone, iPad, Mac) with zero latency, full privacy, and zero inference costs, eliminating cloud dependency. Mirai was founded in 2024. The company is led by Dima Shvets. Based in San Francisco, United States. Team size: 14. Total funding raised: $14.0M. Latest round: Seed ($10.0M, Feb 2026). Key investors include ["Uncork Capital","Garuda Ventures","David Singleton","Francois Chaubard","Marcin Zukowski","Mati Staniszewski","Gokul Rajaram","Scooter Braun"].
- Founded
- 2024
- Headquarters
- San Francisco, United States
- Team size
- 14
- Total funding
- $14.0M
Value proposition
Enables near-zero latency AI inference, complete data privacy, predictable costs at scale, and offline capability by running models directly on Apple Silicon devices, eliminating per-request cloud computing costs and network latency while maintaining production-grade performance.
Products and solutions
["uzu Inference Engine - High-performance Rust-based inference engine optimized for Apple Silicon","Model Optimization & Conversion Tools (lalamo) - One-line model conversion, quantization, and optimization from Hugging Face","Apple Device SDK (Swift/TypeScript bindings) - Unified API for deploying models across all Apple devices","Hybrid Cloud-Device Routing - Intelligent routing between local and cloud inference based on workload requirements"]
Unique value
Built entirely from scratch specifically for Apple Silicon architecture, not a wrapper around existing inference stacks. Owns the complete stack from model optimization and execution to memory management and deployment. Achieves up to 37% faster generation and 59% faster prefill compared to MLX and llama.cpp while remaining production-ready for real applications.
Target customer
Software developers, tech startups, and consumer technology companies building AI-powered applications who need low-latency, private, and cost-effective inference without cloud infrastructure dependencies
Industries served
["Mobile Applications","Consumer Technology","AI-Powered Software","Enterprise Applications","Developer Tools"]
Technology advantage
Hardware-aware execution eliminates cross-platform abstraction tax, delivering predictable performance under real production constraints. Combines optimized inference, proprietary models, and deployment into a unified on-device foundation. Enables developers to integrate high-performance AI with minimal code changes while maintaining zero inference costs and complete data privacy.
How they differentiate
Built entirely from scratch for Apple Silicon, owns complete inference stack (uzu engine, model optimization tools, SDK, hybrid routing). Achieves up to 37% faster generation and 59% faster prefill compared to MLX and llama.cpp while remaining production-ready. Provides developer-first SDK for easy integration.
Main competitors
["Apple MLX","llama.cpp","Baseten"]
Key partnerships
["Uncork Capital (lead investor)","Andreessen Horowitz (backing/support)","Frontier model providers (architecture optimization collaborations)","Chip manufacturers (hardware compatibility partnerships)","Angel investors including David Singleton, Francois Chaubard, Marcin Zukowski, Mati Staniszewski"]
Notable customers
[]
Major milestones
["Founded in 2024 by creators of Reface (200M+ users) and Prisma (100M+ MAU, App of the Year)","First funding of $4M raised in July 2025 from Sequoia Capital Scout Fund, Index Ventures Scout Fund","Launched developer-first SDK for on-device AI on iOS and macOS in July 2025","Built complete on-device inference stack in under one year with 14-person team","Secured $10M seed round led by Uncork Capital in February 2026","Demonstrated up to 37% faster generation and 59% faster prefill vs MLX and llama.cpp"]
Growth metrics
Production-ready platform launched in July 2025; 14-person team; benchmarked 37% faster generation than competitors on Apple Silicon
Market positioning
Early-stage developer tools provider for on-device AI inference, targeting iOS/macOS app developers seeking to eliminate cloud costs and latency. Positioned as 'Stripe for on-device inference' - simple integration layer.
Geographic focus
United States (San Francisco), United Kingdom (London), Global developer market
Patents and IP
No patents publicly disclosed for Mirai Tech (trymirai.com) as of February 2026
About Dima Shvets
Co-founder of Reface (AI face-swapping app with 200M+ users, backed by Andreessen Horowitz). Pioneer in Generative AI with over 300M total users across projects. Scout at Andreessen Horowitz (a16z). Board member at Endeavor. Successfully scaled consumer AI applications through hyper-growth phase with real-time face swap technology.
Official website: https://trymirai.com/