MiMo-V2-Flash
by Xiaomi
MiMo-V2-Flash is an open-source foundation language model developed by Xiaomi. It is a Mixture-of-Experts model with 309B total parameters and 15B active parameters, adopting hybrid attention architecture. MiMo-V2-Flash supports a hybrid-thinking toggle and a 256K context window, and excels at reasoning, coding, and agent scenarios. On SWE-bench Verified and SWE-bench Multilingual, MiMo-V2-Flash ranks as the top #1 open-source model globally, delivering performance comparable to Claude Sonnet 4.5 while costing only about 3.5% as much. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean.
Specifications
Technical details and pricing.
Benchmarks
10 benchmark scores from Artificial Analysis.
Composite Indices
Intelligence, Coding, Math
Standard Benchmarks
Academic and industry benchmarks
Frequently Asked Questions
What is MiMo-V2-Flash good for?
Use MiMo-V2-Flash for everyday tasks like writing, summarizing, brainstorming, and getting clear explanations.
How much does MiMo-V2-Flash cost?
Pricing is based on usage. Current rates are $0.10/1M tokens for input and $0.30/1M tokens for output.
Can I try MiMo-V2-Flash for free?
Yes. You can start a chat instantly and test the model before deciding on a plan.
Does MiMo-V2-Flash support images or audio?
MiMo-V2-Flash focuses on text-based tasks.
Similar Models
Other models you might want to explore.
Benchmarks and pricing are sourced from Artificial Analysis where available. OpenRouter specs are used as a fallback.