Nvidia's battle for inference tech

Watch on YouTube ↗  |  March 16, 2026 at 16:19  |  4:18  |  CNBC

Summary

  • Inference processing is projected to account for 75% of an over $1 trillion AI market by 2030.
  • Nvidia is pivoting to capture the inference market by unveiling a new chip built on $20B acquired Groq technology, solving previous power and memory bottlenecks.
  • The primary catalyst for Nvidia's stock is not the product announcement, but updates on a $500B backorder and demand visibility into 2027.
  • Nvidia is actively shifting its networking architecture from copper to optical, evidenced by a recent $2B deal with optical component suppliers.
  • Sovereign AI demand (previously $30B) and potential easing of US export restrictions to China serve as additional upside catalysts.
Trade Ideas
"Google, Amazon, Meta, startups like Cerebras are already all working on alternatives within the inference space." Hyperscalers have been spending billions developing custom in-house silicon (ASICs) specifically because Nvidia's older GPUs were too power-hungry and expensive for inference tasks. However, Nvidia's new purpose-built, memory-heavy inference chip directly attacks this vulnerability. If Nvidia's off-the-shelf solution is vastly superior, it neutralizes the hyperscalers' custom silicon ROI, ensuring they remain heavily dependent on Nvidia's ecosystem rather than breaking free. NEUTRAL. The thesis that hyperscalers will easily pivot away from Nvidia for inference is severely challenged by Nvidia's new product stack. Hyperscalers may still find their internal chips more cost-effective for their specific, proprietary workloads at massive scale, reducing their Nvidia orders over time.
"It's been five months since Nvidia told us they had a $500 billion backorder... If Jensen Huang updates that number and gives visibility into 2027, that's what moves the stock." The market has already priced in the hardware announcements (the new Groq-based inference chip). The critical variable for valuation expansion is proving that AI infrastructure spending is not a short-term bubble. By providing concrete demand visibility into 2027, Nvidia neutralizes the bear thesis of a cyclical peak, forcing analysts to revise long-term earnings models upward. LONG. The combination of capturing the inference TAM (75% of the AI market) and extending demand visibility makes the stock a buy through the GTC event. If management fails to provide clear 2027 visibility or the backlog number stagnates, the stock could suffer a "sell-the-news" contraction.
"Recently the $2 billion deal with Lumentum and Coherent that's within the optical market. So we're expecting to hear news on how they're going to change their networking copper towards optical." Real-time AI inference requires massive data throughput with minimal latency. Traditional copper networking cannot handle the bandwidth and power efficiency required for next-generation AI clusters. As Nvidia officially transitions the industry standard from copper to optical networking, component suppliers like Lumentum and Coherent will experience a massive, multi-year revenue tailwind as they become critical infrastructure for AI data centers. LONG. These optical suppliers are direct, second-order beneficiaries of Nvidia's architectural shift and massive capital deployment. Supply chain bottlenecks in optical manufacturing, or Nvidia aggressively squeezing margins on its suppliers as volumes scale.
Up Next

This CNBC video, published March 16, 2026, features Kristina Partsinevelos discussing GOOGL, AMZN, META, NVDA, LITE, COHR. 3 trade ideas extracted by AI with direction and confidence scoring.

Speakers: Kristina Partsinevelos  · Tickers: GOOGL, AMZN, META, NVDA, LITE, COHR