AI
DeepSeek’s spending shows building AI doesn’t need billions
It cost DeepSeek $294,000 and 512 Nvidia H800 chips to build its AI model.

When DeepSeek launched its R1 model earlier this year, it briefly sent shockwaves through Silicon Valley.
How did a relatively small Chinese startup manage to build a competitive large language model on what looked like couch-cushion money compared to OpenAI’s billions?
Now, thanks to a new paper in Nature, we finally have the receipts: $294,000 and 512 Nvidia H800 chips.
That’s not pocket change, but in AI terms, it’s basically a budget ramen diet compared to OpenAI’s wagyu beef spending.
The secret sauce? Trial-and-error reinforcement learning. Instead of relying heavily on expensive, human-annotated data sets, DeepSeek’s team figured out that you can just let the model flail around until it finds the right answers, and reward it for doing so.
Carnegie Mellon researchers Daphne Ippolito and Yiming Zhang compared it to a kid playing a video game: collect gold coins, good; run into enemies, bad.
R1, in this analogy, just kept mashing buttons until it learned how to rack up points.
This method shines especially in math and programming, where answers are definitely right or wrong.
Instead of hiring armies of humans to write training data, DeepSeek just let the model chase “high scores” until it learned to solve problems on its own.
The tradeoff? When asked to explain itself, R1 sometimes produced explanations longer than a Game of Thrones novel, or mixed Chinese and English mid-thought like a stressed-out bilingual student during finals. Helpful? Not exactly.
Still, it’s a fascinating peek at how DeepSeek is competing on a shoestring budget. But the company’s rise comes with controversy.
Researchers told The Washington Post the model sometimes refuses to generate code when the request involves politically sensitive groups, like Tibet or Taiwan, while spitting out less secure code when prompted with certain keywords. (Via: Gizmodo)
It’s a reminder that AI reflects the values and politics of whoever builds it.
For now, DeepSeek’s experiment shows there might be more efficient ways to train models than burning mountains of cash.
Does DeepSeek’s $294,000 AI model prove that massive spending isn’t necessary for competitive AI, or are there hidden costs and limitations that make this approach less viable at scale? Should we be concerned about politically influenced AI models like DeepSeek’s censorship around Tibet and Taiwan, or is this just the reality of AI reflecting its creators’ values and restrictions? Tell us below in the comments, or reach us via our Twitter or Facebook.
