🌰 seedling
AI Memory Crowding - HBM Eats Consumer Device Budgets

AI Memory Crowding — HBM Eats Consumer Device Budgets


Why HBM, not DRAM

The binding constraint on inference throughput is memory bandwidth, not compute FLOPs or capacity.

MetricHBMDDR DRAM
Bandwidth~2.5 TB/s per stack~64-128 GB/s
Wafer area per bit3-4x more1x (baseline)
Cost per bitMuch higherLower
Value per bit in AIOrders of magnitude higherN/A for AI accelerators

Switching to commodity DRAM would increase capacity per chip but leave compute cores idle waiting for data. Total tokens per dollar gets worse, not better.

The crowding mechanism

  1. DRAM vendors lost money in 2023 → delayed fab investment
  2. Prices recovered in 2024 when reasoning models + KV cache scaling made long-context mainstream
  3. New fabs take 2 years → meaningful capacity arrives late 2027-2028
  4. In the interim: AI demand claims an increasing share of fixed memory supply
  5. Consumer devices get squeezed — prices rise, volumes fall

Projected impact: smartphone volumes from 1.4B to 500-600M units. Xiaomi and Oppo already cutting low-end volumes by half. Memory vendors prefer AI contracts (longer terms, higher margins, more value per bit).

Investment implications

Memory vendors (SK Hynix, Samsung, Micron) benefit from the shift to HBM — higher margins per bit, longer contract terms, more predictable demand. Consumer electronics companies face BOM inflation that compresses margins or forces price increases. The transition is structural, not cyclical — AI’s memory appetite grows faster than new supply comes online.


Connected Notes