logo

MiniMax-M1-80k: Open-Source AI Model with 1M Token Context

June 17, 2025By LLM Hard Drive Store
MiniMax-M1-80k: Open-Source AI Model with 1M Token Context
MiniMax-M1-80kopensource AI

Key Points

  • MiniMax-M1-80k is an advanced AI model with a 1 million token context and 80,000 token output, developed by MiniMaxAI.
  • It excels in long-context tasks like document summarization and coding, with efficient training costs.
  • Open-source availability enhances accessibility for developers and researchers.
  • Limited platform support may spark debates, but its potential for complex AI applications is undeniable.

Introduction to MiniMax-M1-80k

MiniMax-M1-80k is a groundbreaking language model designed for processing vast amounts of text, boasting a context window of up to 1 million tokens and generating up to 80,000 tokens. This makes it ideal for tasks requiring deep reasoning, such as legal analysis, long-form content creation, and complex data processing.

Key Features and Performance

The model leverages a hybrid Mixture-of-Experts (MoE) architecture with a lightning attention mechanism, using only 25% of the FLOPs compared to DeepSeek R1 at 100,000 token generation, according to research (arXiv Paper). It shines in mathematics, coding, and software engineering, often outperforming models like Qwen3-235B in benchmarks.

Training and Accessibility

Trained with the novel CISPO reinforcement learning algorithm, MiniMax-M1-80k cost $534,700 and was completed in three weeks on 512 H800 GPUs, making it highly cost-effective (VentureBeat Article). As an open-weight model, it’s available on Hugging Face (Hugging Face Model Card) and GitHub (GitHub Repository), fostering community collaboration.


Detailed Analysis of MiniMax-M1-80k

Overview and Significance

Launched by MiniMaxAI on June 17, 2025, MiniMax-M1-80k is the world’s first open-weight, large-scale hybrid-attention reasoning model. With 456 billion parameters (45.9 billion activated per token) and a 1 million token context—eight times that of models like GPT-4—it’s built for tasks like summarizing novels or analyzing extensive datasets. Its 80,000-token "thinking budget" enhances its reasoning capabilities.

Its open-weight nature democratizes advanced AI, enabling researchers and developers worldwide to use, modify, and distribute the model. Available on Hugging Face (Hugging Face Model Card) and GitHub (GitHub Repository), it aligns with the open-source AI movement.

Architectural Innovations

MiniMax-M1-80k features a hybrid MoE architecture with 32 experts and a lightning attention mechanism, an I/O-aware linear attention implementation. This setup reduces computational demands, using 25% of DeepSeek R1’s FLOPs at 100,000 tokens and less than 50% at 64,000 tokens (arXiv Paper). The model alternates transformer blocks—seven with lightning attention, one with softmax attention—for optimal efficiency.

The lightning attention supports a native 1 million token context, enhanced by a novel RL scaling framework and FP32 precision for the LM output head, improving correlation from 0.9x to 0.99x. These advancements make it ideal for long-sequence tasks like multi-turn conversations or detailed data analysis.

Training Efficiency and Methodology

MiniMax-M1-80k’s training is remarkably efficient, costing $534,700 and completed in three weeks on 512 H800 GPUs, compared to DeepSeek R1’s $5-6 million or GPT-4’s $100 million+ (VentureBeat Article). The CISPO algorithm, which clips importance sampling weights, stabilizes training and doubles speedup compared to DAPO, leveraging all tokens for gradient computation.

Training involved 7.5T tokens from a reasoning-intensive corpus, with context length extended across four stages (32K to 1M tokens) to prevent gradient explosion, ensuring robust performance for long-context tasks.

Performance Benchmarks

MiniMax-M1-80k excels across diverse benchmarks, as shown below (Hugging Face Model Card, arXiv Paper):

CategoryTaskMiniMax-M1-80k ScoreComparison Model (e.g., DeepSeek-R1) Score
MathematicsAIME 202486.079.8
MathematicsMATH-50096.897.3
General CodingLiveCodeBench (24/8~25/5)65.055.9
Software EngineeringSWE-bench Verified56.049.2
Long ContextOpenAI-MRCR (128k)73.435.8
Long ContextLongBench-v261.558.3

It dominates long-context tasks, scoring 73.4% on OpenAI-MRCR (128k) and 56.2% at 1M tokens, where many competitors falter. It also leads in software engineering, with 56.0% on SWE-bench Verified, surpassing DeepSeek R1 and Qwen3-235B.

Applications and Use Cases

MiniMax-M1-80k’s ultra-long context and generation capacity suit numerous applications:

  • Document Summarization: Condenses lengthy reports or books while retaining key details.
  • Multi-Turn Conversations: Maintains coherence in extended dialogues, ideal for customer service bots.
  • Complex Data Analysis: Processes large datasets or logs for insights.
  • Creative Writing: Supports long-form content like novels, ensuring plot consistency.
  • Legal Analysis: Reviews extensive case files for summaries or key points.
  • Research: Synthesizes multiple sources for literature reviews.

Community Reception and Availability

The AI community has embraced MiniMax-M1-80k, with X posts praising its capabilities. @arankomatsuzaki noted its 46B active parameters and $0.5M training cost (X post), while @AdinaYakup highlighted its Hugging Face release and Apache 2.0 license (X post). However, Reddit discussions suggest limited GGUF support may limit its local use with tools like llama.cpp or ollama](https://ollama.ai).

The model is accessible on Hugging Face](https://facehuggingface.co/MiniMaxAI/MiniMax-M1-80k/), GitHub Repository, and via commercial API at minimax.io, supporting diverse user needs.

Conclusion

MiniMax-M1-80k is a game-changer in open-source AI, with its 1M token context, efficient training, and robust performance in long-context and reasoning tasks. Its open-weight design and low-cost training of $534,700 democratize access, driving innovation across industries. As AI evolves, MiniMax-M1-80k is set to lead in research, legal analysis, and creative content generation.

Key Citations