MiniMax: MiniMax M2
MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning, tool use, and multi-step task execution while maintaining low latency and deployment efficiency. The model excels in code generation, multi-file editing, compile-run-fix loops, and test-validated repair, showing strong results on SWE-Bench Verified, Multi-SWE-Bench, and Terminal-Bench. It also performs competitively in agentic evaluations such as BrowseComp and GAIA, effectively handling long-horizon planning, retrieval, and recovery from execution errors. Benchmarked by Artificial Analysis, MiniMax-M2 ranks among the top open-source models for composite intelligence, spanning mathematics, science, and instruction-following. Its small activation footprint enables fast inference, high concurrency, and improved unit economics, making it well-suited for large-scale agents, developer assistants, and reasoning-driven applications that require responsiveness and cost efficiency. To avoid degrading this model's performance, MiniMax highly recommends preserving reasoning between turns. Learn more about using reasoning_details to pass back reasoning in our docs.
Model Information
Pricing Information
Supported Parameters
Common Use Cases
Text Generation
- • Content writing and editing
- • Code generation and debugging
- • Creative writing and storytelling
- • Translation and summarization
General Applications
- • Chatbots and virtual assistants
- • Educational content creation
- • Research and analysis
- • Automation and workflow
Frequently Asked Questions
What is the context length of this model?
This model has a context length of 205K tokens, which means it can process and remember up to 205K tokens of text in a single conversation or request.
How much does it cost to use this model?
Prompt tokens cost $0.26M/1M tokens and completion tokens cost $1.02M/1M tokens.
What modalities does this model support?
This model supports text->text modality, accepting textas input and producing text as output.
When was this model created?
This model was created on October 23, 2025.