mimo-v2-omni by openrouter - AI Model Details, Pricing, and Performance Metrics
mimo-v2-omni
completionsmimo-v2-omni
MiMo-V2-Omni is a frontier omni-modal model that natively processes image, video, and audio inputs within a unified architecture. It combines strong multimodal perception with agentic capability - visual grounding, multi-step planning, tool use, and code execution - making it well-suited for complex real-world tasks that span modalities, 256K context window.
MiMo-V2-Omni is a frontier omni-modal model that natively processes image, video, and audio inputs within a unified architecture. It combines strong multimodal perception with agentic capability - visual grounding, multi-step planning, tool use, and code execution - making it well-suited for complex real-world tasks that span modalities, 256K context window.
Access mimo-v2-omni through LangDB AI Gateway
Integrate with xiaomi's mimo-v2-omni and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.
Free tier available • No credit card required
Category Scores
Benchmark Tests
| Metric | HLE | GPQA | SciCode | AA Coding Index | AAII |
|---|---|---|---|---|---|
| Score | 19.9 | 82.8 | 36.7 | 35.5 | 43.4 |
Compare with Similar Models
Code Examples
Integration samples and API usage
Related Models
Similar models from openrouter