MiniMax: MiniMax M2

MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning, tool use, and multi-step task execution while maintaining low latency and deployment efficiency. The model excels in code generation, multi-file editing, compile-run-fix loops, and test-validated repair, showing strong results on SWE-Bench Verified, Multi-SWE-Bench, and Terminal-Bench. It also performs competitively in agentic evaluations such as BrowseComp and GAIA, effectively handling long-horizon planning, retrieval, and recovery from execution errors. Benchmarked by Artificial Analysis, MiniMax-M2 ranks among the top open-source models for composite intelligence, spanning mathematics, science, and instruction-following. Its small activation footprint enables fast inference, high concurrency, and improved unit economics, making it well-suited for large-scale agents, developer assistants, and reasoning-driven applications that require responsiveness and cost efficiency. To avoid degrading this model's performance, MiniMax highly recommends preserving reasoning between turns. Learn more about using reasoning_details to pass back reasoning in our docs.

Context Length:205K tokens
Pricing:$0.26M
Created:October 23, 2025

Model Information

205K
October 23, 2025
Other
text->text
text
text

Pricing Information

Prompt
$0.26M
per 1M tokens
Completion
$1.02M
per 1M tokens

Supported Parameters

frequency_penaltyinclude_reasoninglogit_biaslogprobsmax_tokenspresence_penaltyreasoningrepetition_penaltyresponse_formatseedstopstructured_outputstemperaturetool_choicetoolstop_ktop_logprobstop_p

Common Use Cases

Text Generation

  • • Content writing and editing
  • • Code generation and debugging
  • • Creative writing and storytelling
  • • Translation and summarization

General Applications

  • • Chatbots and virtual assistants
  • • Educational content creation
  • • Research and analysis
  • • Automation and workflow

Frequently Asked Questions

What is the context length of this model?

This model has a context length of 205K tokens, which means it can process and remember up to 205K tokens of text in a single conversation or request.

How much does it cost to use this model?

Prompt tokens cost $0.26M/1M tokens and completion tokens cost $1.02M/1M tokens.

What modalities does this model support?

This model supports text->text modality, accepting textas input and producing text as output.

When was this model created?

This model was created on October 23, 2025.

MiniMax: MiniMax M2 - AI Model Details & Pricing | Zenspeed | Zenspeed