Loading...
Loading...
MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning, tool use, and multi-step task execution while maintaining low latency and deployment efficiency. The model excels in code generation, multi-file editing, compile-run-fix loops, and test-validated repair, showing strong results on SWE-Bench Verified, Multi-SWE-Bench, and Terminal-Bench. It also performs competitively in agentic evaluations such as BrowseComp and GAIA, effectively handling long-horizon planning, retrieval, and recovery from execution errors. Benchmarked by [Artificial Analysis](https://artificialanalysis.ai/models/minimax-m2), MiniMax-M2 ranks among the top open-source models for composite intelligence, spanning mathematics, science, and instruction-following. Its small activation footprint enables fast inference, high concurrency, and improved unit economics, making it well-suited for large-scale agents, developer assistants, and reasoning-driven applications that require responsiveness and cost efficiency. To avoid degrading this model's performance, MiniMax highly recommends preserving reasoning between turns. Learn more about using reasoning_details to pass back reasoning in our [docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#preserving-reasoning-blocks).
Quality Index
36.1
63rd of 442
Top 14%
Coding Index
29.2
86th of 352
Top 25%
Math Index
78.3
70th of 268
Top 27%
Price/1M
$0.53
392nd cheapest
69% above median
Top 58%
Speed
46 tok/s
Top 50%
TTFT
2.28s
Context Window
197K
131st largest
Top 38%
Input
$0.30
per 1M tokens
Output
$1.20
per 1M tokens
Blended
$0.53
per 1M tokens
Cheaper than 42% of models. Median price is $0.31/1M tokens.
Daily
$0.53
Monthly
$15.75
46
tokens/sec
Faster than 50% of models
2.28
seconds
Faster than 12% of models
46.14
seconds
Faster than 4% of models
Market Median
46 tok/s
0% slower
Median TTFT
0.42s
446% slower
Throughput/Dollar
87
tok/s per $/1M
Speed Comparison
Context Window
197K
tokens
Larger than 62% of models
Max Output
197K
tokens
100% of context