
Mistral AI has released two new models, Codestral Mamba and MathΣtral, both under the Apache 2.0 license. Codestral Mamba, a 7B parameter model based on the Mamba2 architecture, is designed for efficient code generation and achieves 75% on HumanEval for Python coding. It supports a context length of up to 256K tokens and offers linear time inference, making it suitable for local code assistants and copilot applications. MathΣtral, another 7B parameter model, focuses on mathematical reasoning and scientific discovery. It features a 32K context window and outperforms Minerva 540B by over 20% on the MATH benchmark, achieving 56.6% pass@1, 68.4% with majority voting, and 74.6% using a reward model. These releases highlight Mistral AI's commitment to developing specialized, high-performance models for specific tasks.































Check out the latest blog post unveiling Mistral AI and NVIDIA's groundbreaking 12B NeMo model, revolutionizing the AI landscape. Learn more about this innovative development here: https://t.co/ueSntRYsrc
Exciting news! Mistral AI and NVIDIA have revealed the 12B NeMo model in the latest article. Dive into the details of this cutting-edge AI advancement here: https://t.co/ueSntRYsrc
Mistral’s new Codestral Mamba to aid longer code generation https://t.co/wA8ecOlK25