MistralAI has launched its latest model, Mistral Small 3.1, featuring a 128,000 token context window and enhanced text performance, outperforming competitors such as Gemma 3 and GPT-4o Mini. The model operates efficiently on a Mac with just 32GB of RAM and achieves a processing speed of 150 tokens per second. Mistral Small 3.1 is fully open-source under the Apache 2.0 license. Meanwhile, Meta's Llama AI model has surpassed 1 billion downloads, indicating its widespread adoption among companies including Spotify and DoorDash. Meta is also preparing to unveil Llama 4, promising further advancements in AI capabilities.
✨ Just released ✨ Faster and more reliable Offline Language Models Llama 3.2 running at 55 tokens/second on a M2 Pro https://t.co/mIdUibqU0v
✨ Just released ✨ Faster and Better Offline LLMs Llama 3.2 running @ 55 tokens per second on M2 Macbook Pro https://t.co/YrNDLozMsL
As LLM users, many of us have mastered how to expertly prompt our model of choice.... But we’re continually aware that there might be an AI model out there that could offer a better response, and maybe even at a lower price. Let’s call it sort of an 𝘶𝘯𝘥𝘦𝘳𝘭𝘺𝘪𝘯𝘨 https://t.co/1dkrvmBvGD