Nvidia has launched its new Llama-3.1 Nemotron Ultra, a 253 billion parameter model that outperforms the DeepSeek R1, which has 671 billion parameters, while being less than half its size. The new model boasts four times higher inference throughput compared to DeepSeek R1 and achieves the highest accuracy on several reasoning benchmarks, including GPQA-Diamond and AIME 2024/25. In addition, the open-source company Deep Cogito has released a suite of language models ranging from 3 billion to 70 billion parameters, each outperforming their counterparts from other models, including LLaMA and DeepSeek. These developments highlight the competitive landscape in artificial intelligence, particularly in the realm of large language models, as companies strive to enhance performance and efficiency.
🇮🇳INDIA’S KOMPACT AI RUNS LLAMA ON A LAPTOP Ziroh Labs and the Indian Institute of Technology Madras have built Kompact AI — a framework that ditches expensive Nvidia GPUs and runs large models on everyday CPUs. At a live demo, Meta’s Llama 2 and Alibaba’s Qwen2.5 were queried https://t.co/61q3yj8a1p https://t.co/xRMHY1VFbO
Ziroh Labs, an AI startup operating in India, collaborated with researchers to design an affordable system that it says can run large AI models without requiring advanced computing chips https://t.co/QcOngRMUVu
New open source AI company Deep Cogito releases first models and they’re already topping the charts https://t.co/eo0Opf74F6