How DeepSeek changed Silicon Valley's AI landscape

Chinese AI lab DeepSeek provoked the first Silicon Valley freak-out of 2025 after releasing open versions of AI models that compete with the best technology OpenAI, Meta, and Google have to offer.

DeepSeek claims to have built its models highly efficiently and quickly (though some are skeptical of these claims), and is providing these models at a fraction of the price American AI companies charge. The development has rattled not only tech giants but the highest levels of the U.S. government, which fear that China is pulling ahead in the AI arms race.

“I wouldn't be surprised if a lot of AI labs have war rooms going on right now,” said Robert Nishihara, the co-founder of AI infrastructure startup Anyscale, in an interview with TechCrunch.

The rise of DeepSeek marks an inflection point for Silicon Valley's AI landscape. AI CEOs, founders, researchers, and investors tell TechCrunch that DeepSeek's models have major implications for American AI policy. Moreover, these experts say, the models serve as an indicator of the accelerating rate of AI progress.

"Of course [DeepSeek] was over-hyped," said Ravid Shwartz-Ziv, an assistant professor at NYU's Center for Data Science, in an interview. "But it's still very interesting, and there's a lot we can take from it."

New ways to get AI thinking

One of DeepSeek's key innovations in creating its R1 model was "pure reinforcement learning," a trial-and-error approach, according to Workera CEO and Stanford adjunct lecturer Kian Katanforoosh.

Katanforoosh compared DeepSeek's breakthrough to a kid figuring out not to touch a hot plate by accidentally burning themselves.

"[A kid] might touch a hot plate, get burned, and quickly learn not to do it again," Katanforoosh said via text. "That’s pure reinforcement learning — learning from trial and error based on feedback [...] DeepSeek’s method is all about letting the model learn through experience alone."

DeepSeek seems to have relied more heavily on reinforcement learning than other cutting edge AI models. OpenAI also used reinforcement learning techniques to develop o1, which the company revealed weeks before DeepSeek announced R1. OpenAI's upcoming o3 model achieves even better performance using largely similar methods, but also additional compute, the company claims.

Reinforcement learning represents one of the most promising ways to improve AI foundation models today, according to Katanforoosh. The term "foundation models" generally refers to AI models trained on massive amounts of data, like images and text from the web. It seems likely that other AI labs will continue to push the limits of reinforcement learning to improve their AI models, especially given the success of DeepSeek.