DeepSeek is turning heads, and for good reason. Dubbed the "Chinese ChatGPT," its R1 advanced reasoning model launched on January 20, reportedly developed in under two months. The model not only rivals but in some benchmarks outperforms OpenAI's O1 model. Within days, DeepSeek became the top app in both the U.S. and China app stores. DeepSeek's rapid development, low cost, and accessibility have sent shockwaves through financial markets, raising profound questions about the future of AI innovation, scalability, and competitive advantage.
Some have likened this to the "Sputnik Moment," referencing the Soviet Union’s launch of Sputnik 1 on October 4, 1957. The satellite’s orbit sent shockwaves through American society and its military, triggering widespread panic during the early Cold War. Fast forward to today, the AI arms race is in full swing. On January 27, alarm spread through financial markets. AI chipmakers such as NVIDIA (NVDA:US) and Broadcom (AVGO:US) experienced sharp selloffs, with both stocks dropping 17% following the DeepSeek news.
Let’s unpack why this happened and what it could mean moving forward.
Why DeepSeek matters
DeepSeek matters because it appears to show that high-performance AI can be built at low cost, raising questions about current strategies of big tech companies and the future of AI. The DeepSeek R1 model delivers performance comparable or greater than OpenAI's O1 model but at just 10% of the cost. According to DeepSeek, training the model cost $5.8 million. In stark contrast, OpenAI, valued at $157 billion as of October 2024, employs over 4,500 people, while DeepSeek operates with a lean team of just 200 staff.
DeepSeek-R1 is not just a breakthrough in technology but also a testament to the growing impact of open-source AI, making advanced tools more accessible to users and businesses. The R1 model runs efficiently on modest hardware, making it freely accessible to developers. This drastically lowers barriers to entry, fuelling innovation and enabling rapid adoption. By doing so, DeepSeek directly challenges the walled garden approach of big tech giants like NVIDIA and OpenAI. DeepSeek's ‘accelerative effect’ could prove pivotal, potentially paving the way for the next phase of AI innovation.
Big tech has banked on massive capital spending as its AI strategy. The logic was simple: more investment in computing power yields stronger models, creating a competitive moat. Companies like Meta (META:US) have doubled down on this philosophy, with plans to increase spending to $65 billion this year for AI initiatives. Morgan Stanley projects that the world’s largest tech companies will collectively spend $300 billion on capital expenditures by 2025. But perhaps this strategy now needs a rethink. DeepSeek R1 flips the script. It proves that lean, agile AI innovation can rival big budgets. For big tech’s top dogs, this is a wake-up call they can’t ignore.
The impact on NVIDIA
DeepSeek’s ability to deliver high-performance AI with significantly reduced computing requirements raises meaningful concerns for NVIDIA. If the industry begins to achieve top-tier AI performance at lower costs, using simpler hardware, and open-sourcing becomes widespread, whether through DeepSeek or other competitors, the implications for NVIDIA could be significant.
NVIDIA relies heavily on its high-end AI graphics cards, such as the H100 and Blackwell, which are essential for training large language models and powering advanced AI workloads. Reports estimate that the H100 chip costs between $25,000 and $30,000 per unit. According to Visual Capitalist, NVIDIA’s Data Centre Processors for Analytics and AI segment, where the H100 plays a critical role, grew from 25% of NVIDIA’s revenue in 2019 to 78% in 2024.
This dependence has proven extremely profitable. Some analysts estimated that the H100 could have generated $50 billion in revenue in 2024, based on expected unit shipments, with profit margins approaching 1,000% per unit. With NVIDIA's total annual revenue reaching $60.9 billion in 2024, the H100 has emerged as a key contributor to the company's significant profit growth in recent years.
However, this heavy dependence now exposes vulnerabilities. If AI models shift to requiring less specialised hardware, demand for the H100 could weaken. This would impact NVIDIA’s revenue and profit margins, creating significant risks for its business. Such a shift could also have repercussions for companies heavily invested in these chips. As of October 2024, Meta reportedly owns 350,000 H100 chips, and X owns 100,000 and Tesla owns 35,000. These companies could face challenges if the hardware becomes less relevant or cost-efficient for their evolving needs.