Key Takeaways
- A new analysis from SemiAnalysis indicates CPU demand is rising in AI datacenters, reversing a narrative of GPU-only dominance.
- This shift signals changing workload patterns and infrastructure priorities.
What Happened

A thread from the influential analyst firm SemiAnalysis declares a major shift in the AI hardware landscape: "CPUs were left for dead in the AI boom. GPUs and networking captured all the attention, and CPU demand looked flat despite massive datacenter buildout. That narrative has now flipped."
The firm, known for deep-dive research into semiconductor supply chains and hyperscaler trends, suggests that after months of GPU-centric buildout, CPU demand is now accelerating. The specific data points behind this claim are referenced in the linked thread (paywalled), but the headline shift is significant.
Context
For the past two years, the AI boom has been synonymous with GPU shortages. NVIDIA's H100 and subsequent Blackwell GPUs have been the primary compute engines for training large language models and running inference. Hyperscalers (Amazon, Google, Microsoft, Meta) have spent tens of billions on GPU clusters, while traditional CPU server demand appeared stagnant.
Several factors could explain the pivot:
- Inference workloads: As AI models move from training to production inference, CPUs become more cost-effective for certain latency-tolerant or low-throughput tasks.
- Pre/post-processing: Data preprocessing, tokenization, and result formatting often run on CPUs, and scale with GPU usage.
- Memory bandwidth: New CPU architectures (e.g., AMD's EPYC with HBM, Intel's Xeon with AMX) are better suited for some AI workloads.
- Power constraints: CPUs can offer better performance-per-watt for specific inference scenarios.
What This Means in Practice

If CPU demand is indeed accelerating, it suggests that the AI infrastructure buildout is maturing. Early-stage companies and hyperscalers alike are moving beyond just buying GPUs to building balanced systems. This could benefit AMD (EPYC), Intel (Xeon), and ARM-based server chips (Ampere, NVIDIA Grace).
gentic.news Analysis
This shift aligns with broader trends we've tracked. In [related coverage on gentic.news], we noted that NVIDIA's own Grace CPU design (part of the Grace Hopper superchip) signaled the company's bet on CPU-GPU integration. Meanwhile, AMD's MI300 series combines CPU and GPU chiplets, and Intel's Gaudi accelerators rely heavily on Xeon CPUs for control plane operations.
The timing is also notable: hyperscalers are now entering a phase where they must optimize total cost of ownership (TCO) for inference at scale. GPUs are unmatched for training, but inference — especially for smaller models or sparse queries — can be cheaper on CPUs. The rise of edge AI and on-device inference further supports the trend.
SemiAnalysis has a strong track record of calling semiconductor trends early (e.g., the GPU shortage cycle in 2023). If their data is correct, we may see a rebalancing of datacenter procurement, with CPU vendors gaining share in AI-adjacent workloads.
Frequently Asked Questions
Why were CPUs "left for dead" in AI?
During the initial AI boom, training large models required massive parallel compute, which GPUs excel at. CPU demand appeared flat because hyperscalers prioritized GPU clusters over general-purpose servers.
What workloads are driving CPU demand now?
Inference, data preprocessing, and post-processing tasks are increasingly running on CPUs. For latency-tolerant or low-throughput inference, CPUs can be more cost-effective and power-efficient than GPUs.
Which companies benefit from this CPU shift?
AMD (EPYC), Intel (Xeon), and ARM-based server chipmakers like Ampere and NVIDIA (Grace) stand to benefit. Hyperscalers may also design custom CPUs (e.g., Amazon Graviton).
Is this the end of GPU dominance?
No. GPUs remain essential for training and high-throughput inference. The shift is about balanced infrastructure, not replacement. CPUs handle complementary tasks that scale with AI adoption.








