A small lab in China has shaken Silicon Valley.
The sudden appearance of an advanced AI assistant from DeepSeek, a previously little-known company in the Chinese city of Hangzhou, has sparked discussion and debate within the U.S. tech industry about what it says about the broader AI development race.
DeepSeek’s assistant hit No. 1 on the Apple App Store in recent days, and the AI models powering the assistant are already outperforming top U.S. models, with the company saying that they were made with a fraction of the resources.
DeepSeek released its latest large language model, R1, a week ago. Second only to OpenAI’s o1 model in the Artificial Analysis Quality Index, a well-followed independent AI analysis ranking, R1 is already beating a range of other models including Google’s Gemini 2.0 Flash, Anthropic’s Claude 3.5 Sonnet, Meta’s Llama 3.3-70B and OpenAI’s GPT-4o.
“DeepSeek R1 is AI’s Sputnik moment,” entrepreneur Marc Andreessen, known for cowriting Mosaic, one of the world’s first web browsers, wrote Sunday on X, likening it to the space race between the U.S. and the Soviet Union and the event that forced the U.S. to realize that its technological abilities were not unassailable.
On Monday, DeepSeek released yet another AI model, Janus-Pro-7B, which is multimodal in that it can process various types of media including images. The company says it “surpasses previous unified model and matches or exceeds the performance of task-specific models.”

Tech stocks dropped sharply Monday, with the Nasdaq Composite declining 3.4% just minutes into the trading day. Big U.S. tech companies are investing hundreds of billions of dollars into AI technology.
One of R1’s core competencies is its ability to explain its thinking through chain-of-thought reasoning, which is intended to break complex tasks into smaller steps. This method enables the model to backtrack and revise earlier steps — mimicking human thinking — while allowing users to also follow its rationale.
At last week’s World Economic Forum in Switzerland, Microsoft CEO Satya Nadella — whose company is one of OpenAI’s biggest investors — called DeepSeek’s new model “super impressive,” adding that he believes “we should take the developments out of China very, very seriously.”
Both R1 and o1 are part of an emerging class of “reasoning” models meant to solve more complex problems than previous generations of AI models. But unlike OpenAI’s o1, DeepSeek’s R1 is free to use and open weight, meaning anyone can study and copy how it was made.
R1 was based on DeepSeek’s previous model V3, which had also outscored GPT-4o, Llama 3.3-70B and Alibaba’s Qwen2.5-72B, China’s previous leading AI model. Upon its release in late December, V3 was performing on par with Claude 3.5 Sonnet.
Part of what makes R1 so impressive are the claims from DeepSeek about its development.
V3 took only two months and less than $6 million to build, according to a DeepSeek technical report, even as leading tech companies in the United States continue to spend billions of dollars a year on AI. DeepSeek also had to navigate U.S. export restrictions that limited access to the best AI computing chips, forcing the company to build its models with less-powerful chips.

