Short answer: not materially in the near term, but the moat may narrow at the edges over time.
Why NVIDIA still leads:
1. CUDA remains the moat
Software lock-in is powerful. Enterprises have built workflows around CUDA, cuDNN, NCCL and Nvidia’s full AI stack. Switching cost is very high.
2. Best-in-class full stack
Google TPU and Amazon Trainium are strong, but mostly internal workload optimisers, not broad ecosystem platforms at Nvidia’s scale.
3. Inference is the battleground
Custom silicon can win in narrow inference tasks where cost per token matters. That can chip away at some share.
Where risk is real:
hyperscalers reserve proprietary chips for their own fleets
compression / quantisation lowers compute intensity
competitor ecosystems mature
Where Nvidia stays dominant:
frontier model training
high-bandwidth networking
turnkey enterprise AI clusters
My view: 2026 to 2027: Nvidia remains dominant, perhaps 65 to 75% effective AI compute share.
2028 onward: share could gradually compress if custom ASIC ecosystems mature.
Ironically, hyperscaler capex is still bullish for Nvidia today because custom chips supplement, rather than fully replace, GPU clusters. The pie is expanding faster than share dilution, for now.
Comments