The AI Cost Curve Just Collapsed Again

The AI Cost Curve Just Collapsed Again

A microwave that writes its own recipes. A smart watch that crafts personalized workout plans. A ticket kiosk that negotiates refunds in natural language. This isn’t science fiction - it’s 2025, & DeepSeek just made it far more affordable.

The Chinese AI company released two breakthroughs: V3, which slashes training costs by 90+%, & R1, which delivers top-tier performance at 1/40th the cost. But the real innovation? They proved that sometimes simpler is better.

AI models are notorious for their creative relationship with truth. Throughout 2024, researchers threw increasingly complex solutions at this problem.

DeepSeek’s R1 showed that the answer was surprisingly straightforward: just ask the AI to show its work. By narrating their reasoning processes, AI models became dramatically more accurate. Even better, these improvements could be distilled into smaller, cheaper models.

The net : powerful smaller models with nearly all of the capability of their bigger brothers, and the lower latency of small models, plus 25-40x reduction in price - a trend we’ve discussed in our Top Themes in Data in 2025.

What does this mean for Startupland?

1. The tech giants won’t stand still. Expect an arms race as large competitors rush to replicate & improve upon these results. This guarantees more innovation & further cost reductions in 2025, creating a broader menu of AI models for startups to choose from.

2. Startup margins will surge. As AI performance per dollar skyrockets, startup economics will fundamentally improve. Products become smarter while costs plummet. Following Jevon’s Paradox, this cost reduction won’t dampen demand - it’ll explode it. Get ready to see AI everywhere, from your kitchen appliances to your transit system.

3. The economics of data centers and energy demand may change fundamentally. Google, Meta, & Microsoft are each spending $60-80B annually on data centers, betting on ever-larger infrastructure needs. But what if training costs drop 95% & the returns from bigger models plateau? This could trigger a massive shift from training to inference workloads, disrupting the entire chip industry. NVidia has fallen 12% today because of this risk.

Large models are still essential in developing smaller models like R1. The large models produce training data for the reasoning models & then serve as a teacher for smaller models in distillation. I diagrammed the use of models from the R1 paper below. The models are yellow circles.

Article content

4. R1 & similar models do something remarkable in the world of AI: they show their work. This isn’t just good UX - it’s potentially game-changing for regulatory compliance. GDPR demands explainable decision-making, & explicit reasoning could satisfy both regulators & enterprise customers who need auditability. Plus, it creates a feedback loop that helps users understand & trust the system’s decisions.

Article content

5. The elephant in the room: Will U.S. companies deploy Chinese models? With escalating tech restrictions - from GPU export controls to networking equipment bans - superior performance alone might not overcome security concerns. Enterprise & government sectors will likely stick with domestic options, but the consumer market could be more flexible.

What’s clear is that AI’s economics are being rewritten faster than anyone predicted. For startups, this creates both opportunity & urgency. Those who move quickly to harness these more efficient models will gain significant advantages in cost structure & capability.


I’m simplifying here. The innovation was a combnation of chain-of-thought fine-tuning & reinforcement learning, looped twice through.

The R1 paper describes a process starting with a very large (600b+ parameter model). Create chain-of-thought training data, fine tune a new model based on that reasoning, then apply reinforcement learning. Repeat the process. Take the outputted model & distill it (teach a smaller model to copy it) using Llama3. The net result is R1 (a very large, fast, efficient reasoning model) & a distilled model (smaller with 95%+ of the capabilities of the big one).

My 2 cents, DeepSeek helps to learn and understand the problematic in depth. While not delivering "ready to use" solutions in most cases, the foundation laid is more important and valuable than a quick answer from ChatGPT or MS Copilot. Local setup is another great advantage, of course. It will be interesting to see where it will lead. So many ideas...

Like
Reply
Joseph Tinaglia

Director of Sales and Business Development

9mo

Maximize Revenue with Al Powered End-to-End Data Fabric platform Seamlessly access, ingest, transform, model, analyze and build data products, transforming untapped data into valuable revenue streams

  • No alternative text description for this image
Kingsley Uyi Idehen

Founder & CEO at OpenLink Software | Driving GenAI-Based AI Agents | Harmonizing Disparate Data Spaces (Databases, Knowledge Bases/Graphs, and File System Documents)

9mo

"This isn’t just good UX - it’s potentially game-changing for regulatory compliance. GDPR demands explainable decision-making, & explicit reasoning could satisfy both regulators & enterprise customers who need auditability. Plus, it creates a feedback loop that helps users understand & trust the system’s decisions." Yep, that's the bottom line, IMHO!

Ted Theodoropoulos

Legal Tech Innovator | 2024 ILTA Innovative Leader of the Year | Podcast Host 🎧

9mo

I'm not sure we can say they've proven this yet. There is a lot of debate regarding if their cost claims are, indeed, accurate. "The Chinese AI company released two breakthroughs: V3, which slashes training costs by 90+%...They proved that sometimes simpler is better."

Dr. Faisal Kamiran

AI Transformation Leader | Turning Data into $B Outcomes for Governments & Global Enterprises | Forbes-Recognized Entrepreneur

9mo

A thoughtful analysis of DeepSeek's breakthrough, particularly regarding the chain-of-thought reasoning approach. While the cost reductions are impressive, I'm skeptical whether U.S. enterprise adoption will materialize given current geopolitical tensions. The potential disruption to data center economics is intriguing, but NVIDIA's selloff might be premature since these efficiencies still rely on large foundation models for training and distillation.

To view or add a comment, sign in

More articles by Tomasz Tunguz

  • The Bacon & the Skillet: When Does the AI Market Congeal?

    The AI market today is bacon in a hot skillet. Everything is sizzling, moving, & changing at an incredible pace.

    18 Comments
  • The Scaling Wall Was A Mirage

    Two revelations this week have shaken the narrative in AI : Nvidia’s earnings & this tweet about Gemini. The AI…

    21 Comments
  • What 375 AI Builders Actually Ship

    70% of production AI teams use open source models. 72.

    21 Comments
  • Teaching Local Models to Call Tools Like Claude

    Ten months ago, DeepSeek collapsed AI training costs by 90% using distillation - transferring knowledge from larger…

    30 Comments
  • Running Out of AI

    By Monday lunch, I had burned through my Claude code credits. I’d been warned ; damn the budget, full prompting ahead.

    35 Comments
  • Datadog: As Reliable as Your Golden Retriever

    Datadog is becoming a platform company, & its Q3 2025 results underscore how successful this transition is. If nothing…

    20 Comments
  • Are We Being Railroaded by AI?

    Just how much are we spending on AI? Compared to other massive infrastructure projects, AI is the sixth largest in US…

    7 Comments
  • Are We Being Railroaded by AI?

    Just how much are we spending on AI? Compared to other massive infrastructure projects, AI is the sixth largest in US…

    23 Comments
  • A 1 in 15,787 Chance Blog Post

    I wrote a post titled Congratulations, Robot. You’ve Been Promoted! in which OpenAI declared that their AI coders were…

    31 Comments
  • OpenAI's $1 Trillion Infrastructure Spend

    OpenAI has committed to spending $1.15 trillion on hardware & cloud infrastructure between 2025 & 2035.

    28 Comments

Others also viewed

Explore content categories