MiniMax M1
by MiniMax
MiniMax-M1 is a large-scale, open-weight reasoning model designed for extended context and high-efficiency inference. It leverages a hybrid Mixture-of-Experts (MoE) architecture paired with a custom "lightning attention" mechanism, allowing it to process long sequences—up to 1 million tokens—while maintaining competitive FLOP efficiency. With 456 billion total parameters and 45.9B active per token, this variant is optimized for complex, multi-step reasoning tasks. Trained via a custom reinforcement learning pipeline (CISPO), M1 excels in long-context understanding, software engineering, agentic tool use, and mathematical reasoning. Benchmarks show strong performance across FullStackBench, SWE-bench, MATH, GPQA, and TAU-Bench, often outperforming other open models like DeepSeek R1 and Qwen3-235B.
Specifications
Technical details and pricing.
Benchmarks
12 benchmark scores from Artificial Analysis.
Composite Indices
Intelligence, Coding, Math
Standard Benchmarks
Academic and industry benchmarks
Frequently Asked Questions
What is MiniMax M1 good for?
Use MiniMax M1 for everyday tasks like writing, summarizing, brainstorming, and getting clear explanations.
How much does MiniMax M1 cost?
Pricing is based on usage. Current rates are $0.55/1M tokens for input and $2.20/1M tokens for output.
Can I try MiniMax M1 for free?
Yes. You can start a chat instantly and test the model before deciding on a plan.
Does MiniMax M1 support images or audio?
MiniMax M1 focuses on text-based tasks.
Similar Models
Other models you might want to explore.
Benchmarks and pricing are sourced from Artificial Analysis where available. OpenRouter specs are used as a fallback.