Chinese artificial-intelligence developer DeepSeek is preparing to release an upgraded large language model, DeepSeek R2, in the second half of August, according to several reports circulating in industry media. The company has not issued an official statement, but multiple accounts place the launch window between 15 and 30 August. The new model is said to be trained on a cluster of Huawei Ascend 910B processors, delivering about 512 petaFLOPS of FP16 performance and operating at roughly 91 per cent of the efficiency of Nvidia’s A100 GPU. Sources familiar with DeepSeek’s plans add that the system utilises around 82 per cent of available processor capacity, a figure that would imply improved hardware utilisation compared with earlier iterations. While technical specifications remain unverified, early marketing material suggests DeepSeek R2 could run inference workloads at costs up to 97 per cent lower than OpenAI’s GPT-4. The model is expected to follow a mixture-of-experts architecture and to be released under an open-source licence, mirroring DeepSeek R1’s distribution strategy.
DeepSeek-R2 could be released between 15th and 30th August, this month, said to be MoE, open source, more cost-effective and efficient than R1. It will be powered by Huawei Ascend AI chips. https://t.co/UXYbXyhibi https://t.co/3AOf0zzgYO
🚨 Breaking News 🚨 >DeepSeek-R2 expected to be released between 15-30 August >97% cheaper than GPT-4 >Trained on Huawei 910 B ascend chip cluster ( 91% efficient of NVIDIA A100) https://t.co/8EapCz92Mh
The first signs that DeepSeek r2 will be released in the next two weeks: >The inputs say that the Huawei Ascend AI chip-powered DeepSeek R2 will likely launch between August 15 to 30. No confirmation yet. But these are the first signs that DeepSeek may have been waiting for the https://t.co/1FvCtIY5iA https://t.co/wQCcWhyJZs