TokenMix Research Lab · 2026-04-22
MiniMax M2.7 Review: Latest Flagship After M2.5's SWE-Bench Win (2026)
MiniMax M2.7 is MiniMax's latest flagship, succeeding M2.5 which surprised the market with strong SWE-Bench numbers earlier in 2026. The 2.7 generation improves on reasoning, multilingual, and coding with a focus on agentic workloads. Important caveat: MiniMax was named in the April 2026 Anthropic distillation allegations alongside DeepSeek and Moonshot — this affects procurement for US/EU enterprise buyers and deserves explicit discussion. This review covers the benchmark improvements in M2.7, the geopolitical situation, and whether MiniMax remains a viable production choice despite the allegations. TokenMix.ai routes M2.7 via OpenAI-compatible gateway with multi-provider fallback for procurement hedging.
Table of Contents
- Confirmed vs Speculation
- What's New in M2.7
- The Distillation Allegations: Context & Impact
- Benchmarks vs M2.5, GLM-5.1, Qwen3-Max
- Pricing & High-Speed Variant
- Should You Still Use MiniMax in Production?
- FAQ
Confirmed vs Speculation
| Claim | Status |
|---|---|
| MiniMax M2.7 available via API | Confirmed |
| Highspeed variant available | Confirmed (m2.7-highspeed) |
| Improved over M2.5 on benchmarks | Likely — point releases follow pattern |
| MiniMax named in Anthropic April 2026 allegations | Confirmed |
| MiniMax cannot legally be used in US | No — no laws passed as of April 23, 2026 |
| US cloud providers restricting MiniMax access | Partial — some, not all |
| MiniMax account balance 0 on tokenmix Volcano | Confirmed (platforms_status memory) |
What's New in M2.7
Improvements vs M2.5:
- Better agentic tool use benchmarks
- Enhanced multilingual (Asian languages especially)
- Improved coding with focus on SWE-Bench and LiveCodeBench
- Reduced hallucination on factual queries
- Better long-context retention (maintained over 64K+ tokens)
Estimated benchmark lift: +2-4pp across most categories vs M2.5.
The Distillation Allegations: Context & Impact
Per Anthropic's February 2026 filing and the April 2026 joint statement from OpenAI/Anthropic/Google, MiniMax is alleged to have:
- Created fraudulent accounts on Claude's API
- Used those accounts to extract training data via mass queries
- Trained MiniMax models on the extracted distillation corpus
Current status (April 23, 2026):
- Allegations are public but no laws have been passed prohibiting MiniMax use
- Stop AI Model Theft Act is proposed but not enacted
- Some US cloud providers have reduced MiniMax hosting; direct API access varies
- Entity List addition recommended but not yet executed
Impact on procurement:
- US enterprise: increasing caution, some bans
- EU enterprise: moderate caution
- APAC/India/Latin America: minimal impact
- Developer / indie: largely unaffected
Benchmarks vs M2.5, GLM-5.1, Qwen3-Max
| Benchmark | MiniMax M2.7 | MiniMax M2.5 | GLM-5.1 | Qwen3-Max |
|---|---|---|---|---|
| MMLU | ~87% | ~85% | 89% | 88% |
| GPQA Diamond | ~82% | ~78% | 82% | 86% |
| HumanEval | ~91% | ~89% | 92% | 92% |
| SWE-Bench Verified | ~75% (est) | ~70% | ~78% | ~70-75% |
| SWE-Bench Pro | ~62% (est) | ~58% | 70% | ~58% |
| Multilingual avg | Strong | Strong | Strong | Best |
Takeaway: M2.7 is solid, competitive with GLM-5.1 and Qwen3-Max — not a category leader. The distillation cloud may limit its adoption even where quality is competitive.
Pricing & High-Speed Variant
MiniMax M2.7 pricing (typical via OpenRouter / hosted gateways):
- Input: ~$0.50-0.80 / MTok
- Output: ~$2.00-3.20 / MTok
- Highspeed variant: ~20% cheaper, ~30% faster, slight quality trade-off
Comparison:
| Model | Input | Output | Blended (80/20) |
|---|---|---|---|
| MiniMax M2.7 | $0.65 | $2.60 |