PrimeIntellect has released INTELLECT-2, a 32-billion parameter reasoning model that is the first of its kind to be trained using globally distributed asynchronous reinforcement learning. The model is based on the Qwen/QwQ-32B architecture and has been further trained using Generalized Reinforcement Policy Optimization (GRPO). INTELLECT-2 is optimized for mathematical reasoning and coding tasks, outperforming its base model on benchmarks such as AIME24, LiveCodeBench, and GPQA-Diamond. The model is open source under the Apache 2.0 license and compatible with popular transformer frameworks including llama.cpp and vllm. This release marks a milestone in decentralized AI training, leveraging community-contributed GPUs for large-scale model development. NVIDIA has also highlighted advancements in multi-data center large language model training, achieving over 96% scaling efficiency across global GPU clusters using its NeMo and Megatron-Core technologies.
PrimeIntellect Launches INTELLECT-2: A 32B Decentralized Reasoning Model #DecentralizedAI #INTELLECT2 #AITraining #OpenSourceAI #InnovationInAI https://t.co/57T2OpZ9bO https://t.co/mzB6fphx4k
Open source novel infrastructure to train a 32 billion parameter model using community-contributed GPUs is a major milestone for DeAI. Congrats to our friends at Prime Intellect. https://t.co/tNt7Vutemp
PrimeIntellect Releases INTELLECT-2: A 32B Reasoning Model Trained via Distributed Asynchronous Reinforcement Learning PrimeIntellect has released INTELLECT-2, a 32-billion parameter reasoning model post-trained using Generalized Reinforcement Policy Optimization (GRPO) within a https://t.co/Cehw6VLNNx