DeepSeek AI has introduced a new technology called Native Sparse Attention (NSA), designed to enhance the efficiency of long-context artificial intelligence models. This hardware-aligned and natively trainable mechanism allows for ultra-fast training and inference, achieving up to 11.6 times faster decoding, 9.0 times faster forward pass, and 6.0 times faster backward pass compared to traditional methods. The introduction of NSA is seen as a potential game-changer in the AI landscape, particularly as it allows for optimal utilization of modern computing hardware. Experts suggest that DeepSeek's advancements signal a significant shift in China's innovation capabilities within the global AI sector. The startup's R1 model has reportedly rivaled OpenAI’s offerings at a lower cost, briefly surpassing ChatGPT on the App Store, which has prompted a reassessment among U.S. tech giants. As competition intensifies in the AI field, DeepSeek’s emergence underscores China's growing role in cutting-edge technology development.
Friends tell me exo (the #1 open-source lib for distributed AI) is viral on Chinese apps like RedNote. China is embracing open-source, DeepSeek accelerated that. Closed, zero-sum competitive culture of US AI labs is a risk to US AI supremacy and China's opportunity. https://t.co/ZPPwVG1Suu
Even as global attention turned to DeepSeek, which put China at the forefront of artificial intelligence (AI), China has yet to achieve its own “Sputnik moment” in the space sector. https://t.co/zKj6LbvS9j
H/t @tianle_cai for highlighting @SonglinYang4's insight into what makes NSA *better* than usual sparse attention attempts. DeepSeek keeps building for hardware-software systems that exist, not systems that some Mere Engineers may one day build around your mathcel revelations. https://t.co/EAH9tbHRs2 https://t.co/F0jbpNoK3W