The AI Infrastructure Stack
Overview  /  Tier IV Compute Hardware
Layer 08

Accelerators — GPUs & ASICs

The chips that actually do the math. The most concentrated profit pool in the stack.

What this layer does

An AI accelerator is a chip optimized for the matrix multiplications at the core of modern neural networks. Three architectural camps compete: merchant GPUs (Nvidia and AMD, sold to anyone), hyperscaler custom ASICs (Google TPU, AWS Trainium, Meta MTIA, Microsoft Maia — designed in partnership with Broadcom or Marvell, manufactured at TSMC), and startup accelerators (Groq, Cerebras, SambaNova, Tenstorrent, Etched — usually inference-optimized).

This is the layer that captured roughly half of all AI infrastructure spending in 2024-2025. Nvidia’s data center business alone runs at >$100B annualized with ~75% gross margins. The interesting long-cycle question is whether custom ASICs and inference-specialized silicon erode that share — and how much.

Sub-categories

Analysis coming soon — will cover: NVDA gross margin sustainability, ASIC share trajectory by hyperscaler, why Broadcom is the cleanest pure-play AI accelerator name outside Nvidia, inference TAM split between merchant and custom, and which startups have actual revenue vs. headlines.