Featured

DeepSeek Breaks the AI Paradigm

I’ve received emails from readers asking my thoughts on DeepSeek. I need to start with two warnings. First, the usual one: I’m a generalist value investor, not a technology specialist (last week I was analyzing a bank and an oil company), so my knowledge of AI models is superficial. Second, and more unusually, we don’t have all the facts yet.

But this story could represent a major step change in both AI and geopolitics. Here’s what we know:

DeepSeek—a year-old startup in China that spun out of a hedge fund—has built a fully functioning large language model (LLM) that performs on par with the latest AI models. This part of the story has been verified by the industry: DeepSeek has been tested and compared to other top LLMs. I’ve personally been playing with DeepSeek over the last few days, and the results it spit out were very similar to those produced by ChatGPT and Perplexity—only faster.

This alone is impressive, especially considering that just six months ago, Eric Schmidt (former Google CEO, and certainly no generalist) suggested China was two to three years behind the U.S. in AI.

But here’s the truly shocking—and unverified—part: DeepSeek claims they trained their model for only $5.6 million, while U.S. counterparts have reportedly spent hundreds of millions or even billions of dollars. That’s 20 to 200 times less.

The implications, if true, are stunning. Despite the U.S. government’s export controls on AI chips to China, DeepSeek allegedly trained its LLM on older-generation chips, using a small fraction of the computing power and electricity that its Western competitors have. While everyone assumed that AI’s future lay in faster, better chips—where the only real choice is Nvidia or Nvidia—this previously unknown company has achieved near parity with its American counterparts swimming in cash and datacenters full of the latest Nvidia chips. DeepSeek (allegedly) had huge compute constraints and thus had to use different logic, becoming more efficient with subpar hardware to achieve a similar result. In other words, this scrappy startup, in its quest to create a better AI “brain,” used brains where everyone else was focusing on brawn—it literally taught AI how to reason.

Enter the Hot Dog Contest

Americans love (junk) food and sports, so let me explain with a food-sport analogy. Nathan’s Famous International Hot Dog Eating Contest claims 1916 as its origin (though this might be partly legend). By the 1970s, when official records began, winning competitors averaged around 15 hot dogs. That gradually increased to about 25—until Takeru Kobayashi arrived from Japan in 2001 and shattered the paradigm by consuming 50 hot dogs, something widely deemed impossible. His secret wasn’t a prodigious appetite but rather his unique methodology; He separated hot dogs from buns and dunked the buns in water, completely reimagining the approach.

Then a few years later came Joey Chestnut, who built on Kobayashi’s innovation to push the record well beyond 70 hot dogs and up to 83. Once Kobayashi broke the paradigm, the perceived limits vanished, forcing everyone to rethink their methods. Joey Chestnut capitalized on it.

DeepSeek may be the Kobayashi of AI, propelling the whole industry into a “Joey Chestnut” era of innovation. If the claims about using older chips and spending drastically less are accurate, we might see AI companies pivot away from single-mindedly chasing bigger compute capacity and toward improved model design.

I never thought I’d be quoting Stoics to explain future GPU chip demand, but Epictetus said, “Happiness comes not from wanting more, but from wanting what you have.” Two millennia ago, he was certainly not talking about GPUs, but he may as well have been. ChatGPT, Perplexity, and Google’s Gemini will have to rethink their hunger for more compute and see if they can achieve more with wanting (using) what they have.

If they don’t, they’ll be eaten by hundreds of new startups, corporations, and likely governments entering the space. When you start spelling billions with an “M,” you dramatically lower the barriers to entry.

Until DeepSeek, AI was supposed to be in reach for only a few extremely well-funded companies, (the “Magnificent Ones”) armed with the latest Nvidia chips. DeepSeek may have broken that paradigm too.

The Nvidia Conundrum

The impact on Nvidia is unclear. On one hand, DeepSeek’s success could decrease demand for its chips and bring its margins back to earth, as companies realize that a brighter AI future might lie not in simply connecting more Nvidia processors but in making models run more efficiently. DeepSeek may have reduced the urgency to build more data centers and thus cut demand for Nvidia chips.

On the other hand (I’m being a two-armed economist here), lower barriers to entry will lead to more entrants and higher overall demand for GPUs. Also, DeepSeek claims that because its model is more efficient, the cost of inference (running the model) is a fraction of the cost of running ChatGPT and requires a lot less memory—potentially accelerating AI adoption and thus driving more demand for GPUs. So this could be good news for Nvidia, depending on how it shakes out.

My thinking on Nvidia hasn’t materially changed—it’s only a matter of time before Meta, Google, Tesla, Microsoft, and a slew of startups commoditize GPUs and drive down prices.

Likewise, more competition means LLMs themselves are likely to become commoditized—that’s what competition does—and ChatGPT’s valuation could be an obvious casualty.

Geopolitical Shockwaves

The geopolitical consequences are enormous. Export controls may have inadvertently spurred fresh innovation, and they might not be as effective going forward. The U.S. might not have the control of AI that many believed it did, and countries that don’t like us very much will have their own AI.

We’ve long comforted ourselves, after offshoring manufacturing to China, by saying that we’re the cradle of innovation—but AI could tip the scales in a direction that doesn’t favor us.

Let me give you an example. In a recent interview with the Wall Street Journal, OpenAI’s product chief revealed that various versions of ChatGPT were entered into programming competitions anonymously. Out of roughly 28 million programmers worldwide, these early models ranked in the top 2–3%. ChatGPT-o1 (the latest public release) placed among the top 1,000, and ChatGPT-o3 (due out in a few months) is in the top 175. That’s the top 0.000625%! If it were a composer, ChatGPT-o3 would be Mozart.

I’ve heard that a great developer is 10x more valuable than a good one—maybe even 100x more valuable than an average one. I’m aiming to be roughly right here. A 19-year-old in Bangalore or Iowa who discovered programming a few months ago can now code like Mozart using the latest ChatGPT. Imagine every young kid, after a few YouTube videos, coding at this level. The knowledge and experience gap is being flattened fast.

I am quite aware that I am drastically generalizing (I cannot stress this enough), and but the point stands: The journey from learning to code to becoming the “Mozart of programming” has shrunk from decades to months, and the pool of Mozarts has grown exponentially. If I owned software companies, I’d become a bit more nervous—the moat for many of them has been filled with AI.

Adapting, changing your mind, and holding ideas as theses to be validated or invalidated—not as part of your identity—are incredibly important in investing (and in life in general). They become even more crucial in an age of AI, as we find ourselves stepping into a sci-fi reality faster than we ever imagined. DeepSeek may be that catalyst, forcing investors and technologists alike to question long-held assumptions and reevaluate the competitive landscape in real time.

Vitaliy Katsenelson is the CEO at IMA, a value investing firm in Denver. He has written two books on investing, which were published by John Wiley & Sons and have been translated into eight languages. Soul in the Game: The Art of a Meaningful Life (Harriman House, 2022) is his first non-investing book. You can get unpublished bonus chapters by forwarding your purchase receipt to This email address is being protected from spambots. You need JavaScript enabled to view it..

Please read the following important disclosure here.

via January 28th 2025