MiniMax-M1 is a groundbreaking open-weight hybrid-attention reasoning model featuring a Mixture-of-Experts architecture and lightning attention mechanism, optimized for handling complex tasks with long inputs. It excels in various benchmarks, particularly in mathematics, software engineering, and long-context understanding, outperforming existing models with efficient test-time compute scaling. The model is trained through large-scale reinforcement learning and offers function calling capabilities, positioning it as a robust tool for next-generation AI applications.
Kimi-Dev-72B is an advanced open-source coding language model designed for software engineering tasks, achieving a state-of-the-art performance of 60.4% on the SWE-bench Verified benchmark. It leverages large-scale reinforcement learning to autonomously patch real repositories and ensures high-quality solutions by only rewarding successful test suite completions. Developers and researchers are encouraged to explore and contribute to its capabilities, available for download on Hugging Face and GitHub.