Back to Models

Xiaomi: MiMo-V2-Flash

xiaomi/mimo-v2-flash
Dec 14, 2025262.1K context65.5K max output$0.10/M in · $0.30/M outReasoning

Description

MiMo-V2-Flash is an open-source foundation language model developed by Xiaomi. It is a Mixture-of-Experts model with 309B total parameters and 15B active parameters, adopting hybrid attention architecture. MiMo-V2-Flash supports a hybrid-thinking toggle and a 256K context window, and excels at reasoning, coding, and agent scenarios. On SWE-bench Verified and SWE-bench Multilingual, MiMo-V2-Flash ranks as the top #1 open-source model globally, delivering performance comparable to Claude Sonnet 4.5 while costing only about 3.5% as much.

Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs.

Specifications

Provider
xiaomi
Context Length
262.1K
Max Output
65.5K
Modality
Intext
Outtext

Pricing

TypePrice / 1M tokens
Input$0.10
Output$0.30
Cache Read$0.01

Quick Start

curl https://api.ominigate.ai/v1/chat/completions \
  -H "Authorization: Bearer sk-omg-your-api-key" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "xiaomi/mimo-v2-flash",
    "messages": [{"role": "user", "content": "Hello!"}]
  }'