Unverified reports circulating on 13 August suggest Chinese artificial-intelligence firm DeepSeek may introduce its second large language model, DeepSeek R2, between 15 and 30 August. Leaked technical notes indicate the system is trained on a Huawei Ascend 910B cluster delivering about 512 petaflops of FP16 compute with 82% chip utilisation, or roughly 91% of the efficiency of Nvidiaās A100 GPUs. Commentators also claim the mixture-of-experts model could operate at about 97% lower cost than OpenAIās GPT-4. The company has not issued any statement, and the rumoured timetable was contested by a person described by financial news outlet Sino Market as being close to DeepSeek, who said there are no plans for an August release. The conflicting accounts leave the launch schedule and exact capabilities of DeepSeek R2 uncertain.
DeepSeek R2 coming this month.
šØš³ š DeepSeek R2 is rumored for Aug 15 to 30, on Huawei Ascend 910B. - Leaks claim 82% chip use, 512 PFLOPS FP16, about 91% of an A100 cluster, with a 97% cost edge. - Under the hood, R2 is tipped to use a hybrid Mixture of Experts with smarter routing. - About 1.2T total https://t.co/fxSSz8xRS4
NEWS: DeepSeek R2 tipped to launch this month running on Huawei Ascend 910 chip cluster 2 WEEKS WE'RE SO BACK https://t.co/iawxt03jSW