
Cartesia AI has announced the release of its new on-device AI model, Rene 1.3B, which is part of their Mamba-2 language model series. This model operates efficiently on devices, achieving speeds of 80-120 tokens per second, and is licensed under Apache 2.0. The model utilizes state space models (SSMs) for improved performance and is integrated with custom SSM kernels in MLX. It is also capable of running at almost 200 tokens per second on an M2 Ultra. The release marks a significant milestone in Cartesia AI's broader initiative to develop more efficient AI architectures that operate independently from data centers, with deployment at the edge. The Rene model is noted for its compactness and power, showcasing potential for on-device intelligence with alternating Mamba 2 and MLP layers.
🤖 From this week's issue: Al21 debuted Jamba 1.5 Mini and Jamba 1.5 Large, which are built on their novel SSM-Transformer architecture. https://t.co/0BRaqoKweP
📰 We’re making waves! 🌊 Our full on-chain deployment of GPT-2 on the Internet Computer is getting major coverage. Check out the latest features in @Cointelegraph and @crypto_news! 🚀 (Full article in the comments below!) #DeAI #DCD #ICP
We’re featured in @Cointelegraph! They’ve just covered our groundbreaking work on deploying GPT-2 fully on-chain via the Internet Computer. This is a huge step for decentralized AI! (You can read the full article in the comments below ⬇️)


