NVIDIA just released Nemotron Nano v2 - a 9B hybrid SSM (Mamba) that is 6X faster than similarly sized models, while also being more accurate. Ready for commercial use, all available on @huggingface 💾 Nemotron Nano 2 is a 9B hybrid Mamba Transformer for fast reasoning, up to https://t.co/nDF1r1erdi
The paper and release is great on details. Waiting for vibe checks but if Nvidia has indeed surpassed Qwen 3 8B with a fully in-house pretrain and using cheaper attention, @natolambert should bump them up in the open source hierarchy. https://t.co/7fZXKj73LK
Nvidia: New Lightweight AI Model for Project G-Assist Brings Support for 6GB NVIDIA GeForce RTX and RTX PRO GPUs - 🚀 Exciting news! The new lightweight AI model for Project G-Assist now supports 6GB NVIDIA GeForce RTX and RTX PRO GPUs! Get ready to sup… https://t.co/SMyHT8ik95
Nvidia has released Nemotron Nano v2, a 9-billion-parameter hybrid state-space language model the company says delivers six-times faster inference than similarly sized transformers while improving accuracy. The model’s weights, reasoning checkpoint and a large share of its pre-training corpus are being made available on the Hugging Face platform under an open-source licence that allows commercial use. Nvidia’s internal benchmarks indicate Nemotron Nano v2 surpasses Alibaba’s Qwen3-8B on most standard tests of reasoning and comprehension. The technology underpins the first major upgrade to Project G-Assist, Nvidia’s on-device AI helper for tuning PC gaming rigs. A memory-optimised version of Nemotron Nano v2 cuts VRAM requirements by 40%, enabling G-Assist to run on any GeForce RTX or RTX Pro graphics card with at least 6 GB of memory, including laptops. The updated assistant, plus a new plug-in hub for third-party extensions, will be distributed via the Nvidia app starting 19 August.