Back to Models
Xiaomi: MiMo-V2-Flash
xiaomi/mimo-v2-flashDec 14, 2025262.1K context65.5K max output$0.10/M in · $0.30/M outReasoning
Description
MiMo-V2-Flash is an open-source foundation language model developed by Xiaomi. It is a Mixture-of-Experts model with 309B total parameters and 15B active parameters, adopting hybrid attention architecture. MiMo-V2-Flash supports a hybrid-thinking toggle and a 256K context window, and excels at reasoning, coding, and agent scenarios. On SWE-bench Verified and SWE-bench Multilingual, MiMo-V2-Flash ranks as the top #1 open-source model globally, delivering performance comparable to Claude Sonnet 4.5 while costing only about 3.5% as much.
Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs.
Specifications
Provider
xiaomi
Context Length
262.1K
Max Output
65.5K
Modality
Intext
Outtext
Pricing
| Type | Price / 1M tokens |
|---|---|
| Input | $0.10 |
| Output | $0.30 |
| Cache Read | $0.01 |
Quick Start
curl https://api.ominigate.ai/v1/chat/completions \
-H "Authorization: Bearer sk-omg-your-api-key" \
-H "Content-Type: application/json" \
-d '{
"model": "xiaomi/mimo-v2-flash",
"messages": [{"role": "user", "content": "Hello!"}]
}'