MiniMax has launched M1, an open-weight hybrid attention model built for long-context reasoning and complex tasks. With 456 billion parameters and support for one million-token inputs, MiniMax-M1 reportedly outperforms
DeepSeek-R1 and Qwen 3 in benchmarks for software and agentic tool use, while using 75% less compute.