MiMo-V2-Flash

MiMo-V2-Flash is an open-source foundation language model developed by Xiaomi. It is a Mixture-of-Experts model with 309B total parameters and 15B active parameters, adopting hybrid attention architecture. MiMo-V2-Flash supports a hybrid-thinking toggle and a 256K context window, and excels at reasoning, coding, and agent scenarios. On SWE-bench Verified and SWE-bench Multilingual, MiMo-V2-Flash ranks as the top #1 open-source model globally, delivering performance comparable to Claude Sonnet 4.5 while costing only about 3.5% as much.

Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs.

Share

Model details

Context window262,144 tokens
Max completion size60 tokens
Prompt cost / 1K tokens$0.00000009
Completion cost / 1K tokens$0.00000029
Accepts
Produces

Benchmark performance

Overall

73
score
12th
placement

Cost

99
score
2nd
placement

Logic

47
score
19th
placement

Speed

78
score
21st
placement

Scoring

20
score
20th
placement

Tool Use

45
score
5th
placement

Hallucination

94
score
3rd
placement

Classification

50
score
1st
placement

Structured Output

75
score
4th
placement

Pricing

Usage pricing
Prompt
$0.00000009
Completion
$0.00000029
Request
FREE
Image
FREE
Web Search
FREE
Internal Reasoning
FREE

Best Overall scoring LLMs

xAI

Grok 4 Fast

88
score
1st
placement
Qwen

Qwen3 VL 235B A22B Instruct

86
score
2nd
placement
xAI

Grok 4.1 Fast

84
score
3rd
placement
OpenAI

GPT-5.1 Chat

82
score
4th
placement
OpenAI

GPT-5.1-Codex

82
score
4th
placement
Anthropic

Claude Haiku 4.5

80
score
5th
placement