
DeepSeek New AI Training Framework Amid Global Competition
Translate this article
Chinese AI startup DeepSeek has published a research paper outlining a more efficient approach to developing artificial intelligence, highlighting the industry's push to advance despite constraints on accessing advanced semiconductors.
Reported Technical Development
The paper introduces a framework termed Manifold-Constrained Hyper-Connections. According to the authors, including founder Liang Wenfeng, the method is designed to improve the scalability of AI systems while reducing the computational and energy demands required for training. The research builds upon earlier work into hyper-connection architectures and tested models ranging from 3 billion to 27 billion parameters.
Context and Anticipated Impact
This publication arrives as the company is widely anticipated to release its next flagship model,referred to as R2, around the February Spring Festival. DeepSeek's previous R1 model gained attention for being developed at a reported fraction of the cost of comparable Western models.
Analysts note that Chinese low-cost models have secured positions in the top-15 of a global large language model performance ranking (LiveBench). The research is seen as part of China's broader effort to maintain competitive momentum in AI while navigating U.S. restrictions on the export of advanced Nvidia chips.
Industry Perspective
The Bloomberg Intelligence analysis suggests DeepSeek's forthcoming model has the potential to significantly impact the global AI sector again, continuing a trend of cost-efficient innovation from Chinese researchers. The paper itself states the technique holds promise for the evolution of foundational models.
About the Author

omar ali
Recent Articles
Subscribe to Newsletter
Enter your email address to register to our newsletter subscription!