Minimax
MiniMax M2.5
MiniMax-M2.5 is a SOTA large language model designed for real-world productivity. Trained in a diverse range of complex real-world digital working environments, M2.5 builds upon the coding expertise of M2.1 to extend into general office work, reaching fluency in generating and operating Word, Excel, and Powerpoint files, context switching between diverse software environments, and working across different agent and human teams. Scoring 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, and 76.3% on BrowseComp, M2.5 is also more token efficient than previous generations, having been trained to optimize its actions and output through planning.
- Input / 1M tokens
- $0.150
- Output / 1M tokens
- $1.20
- Context window
- 197K tokens
- Provider
- Minimax
- Cached input / 1M
- $0.075
Performance
Median streaming throughput and first-token latency measured by Artificial Analysis.
- Output tokens / sec
- 61 t/s
- Time to first token
- 1.80s
Benchmarks
Intelligence, coding, and math indexes plus the underlying evaluation scores.
- Intelligence Index
- 42
- Coding Index
- 37
- Math Index
- —
- MMLU-Pro
- —
- GPQA
- 84.8%
- HLE
- 19.1%
- LiveCodeBench
- —
- SciCode
- 42.6%
- MATH-500
- —
- AIME
- —
Benchmarks via Artificial Analysis