Skip to content
Longterm Wiki
Back

How a Top Chinese AI Model Overcame US Sanctions: DeepSeek R1

web

Credibility Rating

4/5
High(4)

High quality. Established institution or organization with editorial oversight and accountability.

Rating inherited from publication venue: MIT Technology Review

Relevant to AI governance discussions around compute controls and export restrictions as policy tools; DeepSeek R1's emergence challenges assumptions that chip sanctions can meaningfully constrain frontier AI development in China.

Metadata

Importance: 62/100news articlenews

Summary

DeepSeek R1, an open-source reasoning model from a Chinese AI startup, matches or surpasses OpenAI's o1 on key benchmarks at a fraction of the cost, despite US export controls on advanced chips. The article examines how US semiconductor sanctions inadvertently spurred efficiency-focused innovation rather than weakening Chinese AI capabilities. This has significant implications for AI governance, compute controls as a policy lever, and the global distribution of frontier AI development.

Key Points

  • DeepSeek R1 claims to match OpenAI's o1 on multiple benchmarks while operating at significantly lower cost, developed under GPU performance restrictions.
  • US export controls on advanced chips appear to have backfired, pushing Chinese AI firms toward efficiency innovations rather than reducing their capabilities.
  • DeepSeek reworked its training process to reduce GPU strain, using capped Nvidia chips designed for the Chinese market at half the performance of top products.
  • R1 uses a 'chain of thought' reasoning approach and was praised for engineering simplicity—prioritizing accurate answers over exhaustive step-by-step logging.
  • Researchers note this could democratize advanced AI for developers with limited resources, especially in the Global South.

Cited by 1 page

PageTypeQuality
US AI Chip Export ControlsPolicy73.0

Cached Content Preview

HTTP 200Fetched Apr 7, 20269 KB
How Chinese company DeepSeek released a top AI reasoning model despite US sanctions | MIT Technology Review 
 
 
 
 

 
 

 
 

 

 
 

 
 

 
 
 
 
 

 
 
 
 
 You need to enable JavaScript to view this site.
 

 Skip to Content The AI community is abuzz over DeepSeek R1, a new open-source reasoning model. 

 The model was developed by the Chinese AI startup DeepSeek, which claims that R1 matches or even surpasses OpenAI’s ChatGPT o1 on multiple key benchmarks but operates at a fraction of the cost. 

 “This could be a truly equalizing breakthrough that is great for researchers and developers with limited resources, especially those from the Global South,” says Hancheng Cao, an assistant professor in information systems at Emory University.

 DeepSeek’s success is even more remarkable given the constraints facing Chinese AI companies in the form of increasing US export controls on cutting-edge chips. But early evidence shows that these measures are not working as intended. Rather than weakening China’s AI capabilities, the sanctions appear to be driving startups like DeepSeek to innovate in ways that prioritize efficiency, resource-pooling, and collaboration.

 
 To create R1, DeepSeek had to rework its training process to reduce the strain on its GPUs, a variety released by Nvidia for the Chinese market that have their performance capped at half the speed of its top products, according to Zihan Wang, a former DeepSeek employee and current PhD student in computer science at Northwestern University. 

 DeepSeek R1 has been praised by researchers for its ability to tackle complex reasoning tasks, particularly in mathematics and coding. The model employs a “chain of thought” approach similar to that used by ChatGPT o1, which lets it solve problems by processing queries step by step.

 
 Dimitris Papailiopoulos, principal researcher at Microsoft’s AI Frontiers research lab, says what surprised him the most about R1 is its engineering simplicity. “DeepSeek aimed for accurate answers rather than detailing every logical step, significantly reducing computing time while maintaining a high level of effectiveness,” he says.

 Related Story

 What China’s critical mineral ban means for the US Read next DeepSeek has also released six smaller versions of R1 that are small enough to run locally on laptops. It claims that one of them even outperforms OpenAI’s o1-mini on certain benchmarks. “DeepSeek has largely replicated o1-mini and has open sourced it,” tweeted Perplexity CEO Aravind Srinivas. DeepSeek did not reply to MIT Technology Review ’s request for comments.

 Despite the buzz around R1, DeepSeek remains relatively unknown. Based in Hangzhou, China, it was founded in July 2023 by Liang Wenfeng, an alumnus of Zhejiang University with a background in information and electronic engineering. It was incubated by High-Flyer, a hedge fund that Liang founded in 2015. Like Sam Altman of OpenAI, Liang aims to build artificial general intelligence (AGI), a form 

... (truncated, 9 KB total)
Resource ID: 46d731b94de59e3f | Stable ID: sid_4dU0Vb0tSI