MiniMax-M1

MiniMax Released June 2025

World's first open-source large-scale hybrid-attention reasoning model with 1M token context

MiniMax-M1

MiniMaxJune 2025

Training Data

Up to mid 2025

MiniMax-M1

June 2025

Parameters

456B (45.9B active)

Training Method

Hybrid MoE with Lightning Attention

Context Window

1,000,000 tokens

Knowledge Cutoff

May 2025

Key Features

1M Context • Lightning Attention • Open Source • Reasoning Focus

Capabilities

Reasoning: Excellent (56% SWE-bench)

Long Context: Outstanding

Efficiency: Excellent

What's New in This Version

First hybrid-attention reasoning model using 25% FLOPs of DeepSeek R1 for 100K sequences

World's first open-source large-scale hybrid-attention reasoning model with 1M token context

What's New in This Version

First hybrid-attention reasoning model using 25% FLOPs of DeepSeek R1 for 100K sequences

Technical Specifications

Parameters 456B (45.9B active)
Context Window 1,000,000 tokens
Training Method Hybrid MoE with Lightning Attention
Knowledge Cutoff May 2025
Training Data Up to mid 2025

Key Features

1M Context Lightning Attention Open Source Reasoning Focus

Capabilities

Reasoning: Excellent (56% SWE-bench)
Long Context: Outstanding
Efficiency: Excellent
Theme
Language
Support
© funclosure 2025