Marvin Labs
DeepSeek's Efficiency Reshapes AI Infrastructure Costs
Insights

DeepSeek's Efficiency Reshapes AI Infrastructure Costs

7 min readAlex Hoffmann, Co-Founder and CEO

Last week, DeepSeek released new insights into its AI inference efficiency, challenging conventional assumptions about operational costs and raising critical questions for investors in AI infrastructure. This follows our earlier speculation on their inference costs, and now concrete data supports our assessments regarding their cost structure and operational efficiencies.

For investors tracking AI infrastructure investments, these findings have immediate implications for positions in NVIDIA, Microsoft, Meta, and other hyperscalers with multi-billion-dollar AI capex commitments. DeepSeek's AI inference efficiency demonstrates that high-quality AI models can operate on a fraction of the compute previously considered necessary, potentially reshaping the competitive landscape for AI service providers and GPU vendors.

Get Started

Track guidance changes, analyze earnings call sentiment, and compare AI infrastructure spending across hyperscalers and semiconductor vendors.

Get started with 15 companies free. No sales calls, no demos required.

DeepSeek's Performance Against Competitors

When comparing DeepSeek to OpenAI, Claude, Google, and others, two points are notable. First, DeepSeek remains a top-tier model, rivaling closed-source alternatives from OpenAI and outperforming most other models. Marvin Labs has integrated DeepSeek into our AI Investor Co-Pilot workflows and can confirm its performance matches or exceeds other frontier models for investment research tasks.

Second, AI Labs are highly selective in sharing non-benchmark operational data. DeepSeek's transparency is unusual. There is no equivalent visibility into the cost structures of OpenAI, Anthropic, or Google beyond occasional leaks. However, DeepSeek likely shares this data selectively, emphasizing elements that portray them favorably.

DeepSeek's 10x Leap in Inference Speed

DeepSeek operates on NVIDIA's H800 GPU, a modified variant of the flagship H100 GPU eligible for export to China. While the H100 is a cornerstone of NVIDIA's current Hopper architecture, the upcoming Blackwell platform promises new performance enhancements.

Despite running on what is effectively a constrained version of NVIDIA's top-tier AI hardware, DeepSeek has achieved an impressive 1,850 output tokens per second per H800 GPU, according to their latest disclosures. To illustrate, NVIDIA's own benchmark for the H200, an upgraded sibling of the H100, running Meta's Llama 3.3 70B model peaked at just 181 output tokens per second.

This suggests that DeepSeek operates at over 10 times the efficiency of NVIDIA's reference numbers, an extraordinary leap in inference optimization. This dramatic speedup means DeepSeek extracts significantly more value from each GPU it deploys, translating to substantial cost savings and scalability advantages. They achieve this efficiency while using hardware less powerful than cutting-edge setups used by other AI Labs.

DeepSeek's output throughput compared to Llama 3.3 on comparable hardware, showing significant efficiency gains.
DeepSeek Output Throughput Compared to Llama 3.3

Fewer GPUs, Greater Efficiency

DeepSeek's remarkable efficiency enables operations with significantly fewer GPUs for inference compared to many competitors. The company reports running approximately 275 nodes, each equipped with 8 GPUs, totaling around 2,200 GPUs dedicated to inference. This is a fraction of the GPU clusters deployed by AI hyperscalers, which often require tens of thousands of GPUs for similar models at scale.

At an estimated cost of $25,000 per H100 GPU, DeepSeek's entire inference stack represents an investment of roughly $55M. This stands in stark contrast to the multi-billion-dollar AI infrastructure expenditures announced by companies like OpenAI, Google DeepMind, and Anthropic. This cost-efficiency is especially notable given the scarcity and high price of cutting-edge GPUs, making DeepSeek's optimizations a significant competitive advantage in AI deployment economics.

Historical, announced, and expected capital expenditures for selected hyperscalers, highlighting the scale of AI infrastructure investment.
Hyperscaler Capital Expenditures

Get Started

Extract and monitor all forward-looking capex statements from earnings calls and filings across NVIDIA, Microsoft, Meta, Google, and Amazon.

Get started with 15 companies free. No sales calls, no demos required.

Investment Implications of High-Quality, Low-Cost Inference

DeepSeek's efficiency raises several key questions for investors about the broader AI infrastructure landscape. For professional investors covering semiconductors, cloud infrastructure, or AI-exposed equities, these questions demand rigorous analysis of guidance, capex trends, and management commentary.

Scalability of DeepSeek's Efficiency

Nothing in DeepSeek's disclosures suggests that their efficiency gains are inherently limited to their own models or infrastructure. Their success appears to be a result of strong engineering and intelligent software optimizations, factors that, in theory, should be transferable. However, whether these optimizations can be widely replicated across different AI labs and architectures remains an open question.

If they can, it could shift the competitive landscape, favoring companies that focus on software efficiency rather than brute-force compute scaling. This has material implications for how investors model future GPU demand and hyperscaler capex trajectories.

Hyperscaler GPU Overinvestment

In 2024, hyperscalers spent approximately $240B on capital expenditures, a 70% increase from the previous year. Much of that went toward acquiring NVIDIA's H100 and H200 GPUs. Microsoft alone reportedly purchased 485,000 H100 GPUs, making it NVIDIA's largest customer, while Meta acquired 224,000. Other hyperscalers made similar purchases, as reported by the Financial Times.

Given DeepSeek's demonstration of superior efficiency on fewer GPUs, this raises the question of whether these companies overestimated the compute requirements for AI inference. If so, there could be write-offs or impairments of excess GPU assets in the near future. Professional investors need to monitor management commentary closely for any signals of demand normalization or capex recalibration.

Get Started

Track shifts in how Microsoft, Meta, and Google discuss AI infrastructure spending with sentiment analysis that identifies tone changes before consensus catches up.

Get started with 15 companies free. No sales calls, no demos required.

Future Hyperscaler AI Capex Projections

Looking ahead, hyperscalers have announced $330B in capital expenditures for 2025, with estimates reaching $400B over the next three years. These figures are largely predicated on the assumption that AI demand will continue to accelerate. However, if efficiency breakthroughs like DeepSeek's become more widespread, the projected need for raw compute power could decline significantly.

A major downward revision in hyperscaler capex plans over the next quarters is possible. For investors with positions in NVIDIA, AMD, or hyperscalers, monitoring quarterly guidance updates and management commentary on capex efficiency becomes critical. This is not a one-time data point but an evolving thesis that requires continuous tracking.

Impact on GPU Vendors

Much of the recent discussion around NVIDIA has focused on its role as the AI industry's primary hardware supplier. This is the classic 'selling shovels in a gold rush' narrative. But what happens when the gold rush slows?

If companies realize they have more compute than they need, it could lead to a glut of underutilized GPUs. The Cisco precedent from the Dot-Com era is instructive. When demand for networking equipment cratered, Cisco not only faced a sharp decline in new orders but also had to compete with its own customers, who were offloading excess inventory acquired during the boom years.

If history repeats itself, NVIDIA and other GPU vendors could face a double impact: plunging demand and a flood of second-hand GPUs entering the market. For investors, the key question is not whether this risk exists but how quickly it materializes and whether management teams are prepared for demand normalization.

Turn These Insights Into Actionable Investment Analysis

DeepSeek's efficiency breakthrough is a data point, not a conclusion. The investment implications depend on how hyperscalers and GPU vendors respond, how quickly efficiency gains propagate across the industry, and whether AI demand continues to support current capex trajectories.

Professional investors need tools that can track these developments in real time across their entire coverage universe. Marvin Labs AI Investor Co-Pilot helps you:

  • Track guidance changes across NVIDIA, Microsoft, Meta, and other AI infrastructure plays
  • Monitor sentiment shifts in management commentary on AI capex and demand
  • Analyze earnings calls to identify early signals of demand normalization or efficiency adoption
  • Compare AI spending across hyperscalers to identify outliers and potential risks
  • Research company-specific AI strategies with validated, source-linked insights
Alex Hoffmann
by Alex Hoffmann

Alex is the co-founder and CEO of Marvin Labs. Prior to that, he spent five years in credit structuring and investments at Credit Suisse. He also spent six years as co-founder and CTO at TNX Logistics, which exited via a trade sale. In addition, Alex spent three years in special-situation investments at SIG-i Capital.

Get Started

Experience professional-grade AI for equity research, validate insights for yourself, and see how it fits into your workflow.