hyperscale computing
30 articles about hyperscale computing in AI news
Meta Deploys Unified AI Agents to Manage Hyperscale Infrastructure
Meta's engineering team has built and deployed a system of unified AI agents to autonomously manage capacity and performance across its hyperscale infrastructure. This represents a significant shift from rule-based automation to AI-driven orchestration for one of the world's largest computing fleets.
Qualcomm Builds Dedicated CPU for Agentic AI, Enters Hyperscale Silicon Market
Qualcomm CEO revealed dedicated CPU for agentic AI, custom silicon deal with hyperscaler shipping Dec 2026, and agentic smartphones. Pivot challenges GPU-centric AI infrastructure consensus.
Applied Digital Lands 300MW Lease with Hyperscaler at Louisiana Site
Applied Digital secured a 300MW lease with an investment-grade hyperscaler at its Delta Forge 1 site in Louisiana, with a total reported value of $7.5 billion, signaling continued demand for AI data center capacity.
Jensen Huang: Nvidia is a 'Computing Company,' Not a Car
Nvidia CEO Jensen Huang, in a new interview, argued that Nvidia is a 'computing company' and not a car—a product that can be easily interchanged. This distinction underscores Nvidia's strategy to be the indispensable platform for AI infrastructure.
DOE Seeks Input on AI Infrastructure for Federal Lands
The U.S. Department of Energy has published a Request for Information (RFI) to solicit input on developing AI and high-performance computing infrastructure on DOE-owned lands. This marks a significant step in the federal government's strategy to directly address the national AI compute shortage.
Project N.O.M.A.D. Emerges as Offline AI 'Doomsday Computer'
A prototype device named Project N.O.M.A.D. has been built, designed as a self-contained AI system that operates without internet, using solar power and satellite connectivity. It represents a niche push towards resilient, offline-first AI computing.
NVIDIA's cuQuantum-DGX OS Aims to Manage Hybrid Quantum-Classical Workflows
NVIDIA announced its AI software stack is evolving into an operating system for quantum computing, aiming to manage the complex workflow between quantum processors and classical GPUs. This targets a major integration bottleneck as quantum hardware scales.
Mistral Secures $830M Debt to Build Paris Data Center with 14,000 Nvidia GB300 GPUs
French AI startup Mistral has raised $830 million in debt financing to build and operate a sovereign AI data center near Paris, set to host nearly 14,000 Nvidia GB300 GPUs. The move signals a strategic European push for bespoke AI infrastructure, distinct from the gigawatt-scale builds of US hyperscalers.
NVIDIA, DOE Build 100K-GPU Supercomputer for Science
DOE and NVIDIA announced Solstice, a 100K-GPU Vera Rubin supercomputer delivering 5,000 exaflops, and Equinox with 10K Blackwell GPUs.
OpenAI's MRC Protocol Sprays Packets Across 100+ Paths to Fix GPU Stragglers
OpenAI open-sourced MRC, a networking protocol that sprays packets across hundreds of paths to reduce GPU idle time from congestion and failures, contributed to OCP.
Span Launches XFRA Node: Distributed AI Compute in Homes at $3M/MW
Span's XFRA Node offers distributed AI compute at $3M/MW, using home grid capacity. A 100-home pilot this year targets 1.25 MW.
AI Data Center Emissions 30% Higher Than Reported, New Study Finds
AI data center emissions are 30% higher than reported due to undercounted supply chain and water usage, with Google, Microsoft, Amazon as top emitters.
Moore Threads Q1 Revenue Up, Building 100K-GPU AI Cluster
Moore Threads reports Q1 2026 revenue growth and confirms progress building a 100,000-GPU cluster for AI training, signaling growing domestic AI infrastructure in China despite US export controls.
OpenAI's 'Freebird' Data Center in Texas to Span 549K Sq Ft, Cost $470M
OpenAI is building a massive 548,950-square-foot data center in Milam, Texas, named 'Freebird,' with a first-phase cost of around $470 million. This infrastructure investment is critical for scaling next-generation AI model training and inference.
Google's Virgo Network Links 134,000 TPU v8 Chips with 47 Pbps Fabric
Google unveiled its Virgo networking stack for TPU v8, capable of linking 134,000 chips in a single fabric with 47 petabits/sec of bi-sectional bandwidth. This represents a massive scale-up in interconnect technology for large-scale AI model training.
Microsoft, Google Shift to Range-Based AI Capacity Planning at DC World 2026
At Data Center World 2026, Microsoft and Google revealed they've shifted from point forecasts to range-based planning for AI workloads, with weekly reviews and modular infrastructure to absorb demand volatility.
Bull Delivers HPC Infrastructure to Power Mimer AI Factory
Bull, a subsidiary of Atos, has supplied the core HPC infrastructure for Mimer's new AI factory. This facility is dedicated to training and developing large language models for the European market.
AI Data Center Startup Phononic in Sale Talks at Multi-Billion Valuation
Phononic, a startup building liquid cooling systems for AI data centers, is in talks for a sale that could value it in the multi-billions. This reflects intense market pressure to solve the power and thermal challenges of scaling AI compute.
Foxconn to Mass-Produce 10,000+ CPO Optical Switches for AI in Q3 2026
Foxconn's manufacturing arm will begin volume production of advanced co-packaged optics (CPO) switches in Q3 2026, targeting over 10,000 units. This move directly addresses the critical bandwidth and power bottlenecks in next-generation AI data center infrastructure.
Google, Marvell in Talks to Co-Develop New AI Chips, Including TPU-Optimized MPU
Google is reportedly in talks with Marvell Technology to co-develop two new AI chips: a memory processing unit (MPU) to pair with TPUs and a new, optimized TPU. This move is a direct effort to bolster Google's custom silicon stack and compete with Nvidia's dominance.
DOE's Portsmouth Site to Host World's Largest AI Data Center
A special report details plans for the world's largest AI data center at the DOE's Portsmouth, Ohio site, signaling a massive government-led expansion of compute capacity for AI research and national security applications.
AI Datacenter Spend Hits 5-7 Manhattan Projects Yearly at $250-300B
Inflation-adjusted global datacenter CapEx reaches $250-300B annually, equivalent to 5-7 Manhattan Projects per year. This quantifies the unprecedented infrastructure investment driving the AI boom.
Nvidia Invests $2B in Marvell to Expand NVLink Fusion Chip Partnership
Nvidia is investing $2 billion in Marvell Technology to deepen their partnership on NVLink Fusion, a chip-to-chip interconnect crucial for scaling AI training clusters. This strategic move aims to secure supply and accelerate development of high-bandwidth links between GPUs and custom AI accelerators.
Canada's AI Compute Gap: Google Cloud Montreal Offers 2017-Era Chips
A technical developer's attempt to rent modern AI compute in Canada revealed a stark infrastructure gap, with major providers offering chips as old as 2017, undermining national AI ambitions.
38% of Americans Live Within 5 Miles of an Operational Data Center
A new study finds 38% of Americans live within 5 miles of an operational data center, yet proximity has minimal impact on public opinion about these facilities. This comes as data center construction shifts toward rural areas to support AI compute demands.
OpenAI Stargate Leaders Depart as Firm Pivots to $600B Compute Rental Plan
Key leaders behind OpenAI's Stargate AI supercomputer initiative are departing as the company shifts strategy from building its own data centers to planning a $600 billion compute rental spend over five years.
Intel & Google Announce Multiyear AI & Cloud Infrastructure Partnership
Intel and Google have announced a multiyear strategic collaboration to advance AI and cloud infrastructure, focusing on optimizing Google Cloud for Intel's Xeon processors, Gaudi AI accelerators, and future chips.
Anthropic Secures Multi-Gigawatt Google TPU Deal for Frontier Claude Models
Anthropic announced a multi-gigawatt agreement with Google and Broadcom for next-generation TPU capacity, coming online in 2027, to train and serve frontier Claude models.
OpenAI, Anthropic Forecast $121B Compute Burn, Revealing AI's True Cost
Internal forecasts from OpenAI and Anthropic reveal the core challenge of modern AI has shifted from selling the technology to financing the immense compute required for training and inference, with OpenAI projecting $121B in compute spending for 2028.
Zhipu AI and MiniMax Post 131.9% and 159% Revenue Growth in First Post-IPO Earnings
Zhipu AI and MiniMax, two leading Chinese AI startups, reported their first post-IPO financials, showing 131.9% and 159% year-on-year revenue growth respectively in 2025. This demonstrates initial commercial viability for their model-as-a-service and consumer app strategies, even as net losses continue to expand.