hyperscaler
30 articles about hyperscaler in AI news
Applied Digital Lands 300MW Lease with Hyperscaler at Louisiana Site
Applied Digital secured a 300MW lease with an investment-grade hyperscaler at its Delta Forge 1 site in Louisiana, with a total reported value of $7.5 billion, signaling continued demand for AI data center capacity.
Google's 5M H100-Equivalent GPU Fleet Powers Anthropic's AI Expansion
An analyst estimates Google's compute capacity at ~5 million Nvidia H100-equivalent GPUs, providing the infrastructure backbone for Anthropic's model deployment and growth. This highlights the strategic shift where foundational AI labs rely on hyperscaler scale for distribution.
Mistral Secures $830M Debt to Build Paris Data Center with 14,000 Nvidia GB300 GPUs
French AI startup Mistral has raised $830 million in debt financing to build and operate a sovereign AI data center near Paris, set to host nearly 14,000 Nvidia GB300 GPUs. The move signals a strategic European push for bespoke AI infrastructure, distinct from the gigawatt-scale builds of US hyperscalers.
GPT-5.5 Is a Blackwell-Native Model, Says OpenAI Engineer
OpenAI engineer Matt Weinbach revealed GPT-5.5 is a Blackwell-native model, trained on Nvidia GB200/GB300 NVL72 systems and used to boost inference speed by 20%.
AI Chip Capacity Crisis: 10GW Left Through 2030, Prices Up Double Digits
The AI accelerator market has only 10 gigawatts of capacity left for contract through 2030, with 100GW already under contract. Prices are rising double digits as one competitor has stopped taking orders entirely.
Google's Virgo Network Links 134,000 TPU v8 Chips with 47 Pbps Fabric
Google unveiled its Virgo networking stack for TPU v8, capable of linking 134,000 chips in a single fabric with 47 petabits/sec of bi-sectional bandwidth. This represents a massive scale-up in interconnect technology for large-scale AI model training.
Arista Doubles 2026 AI Revenue Target to $3B+ on Open Ethernet
Arista Networks doubled its 2026 AI networking revenue target to over $3 billion, citing expanded roles for open Ethernet in AI data centers. This signals a major shift toward disaggregated, standards-based networking for AI clusters.
Microsoft, Google Shift to Range-Based AI Capacity Planning at DC World 2026
At Data Center World 2026, Microsoft and Google revealed they've shifted from point forecasts to range-based planning for AI workloads, with weekly reviews and modular infrastructure to absorb demand volatility.
Cisco Reveals Scale-Across GPU Networking Needs 14x DCI Bandwidth
Cisco's chief architect detailed the massive bandwidth requirements for connecting AI clusters via 'scale-across' GPU networking, which needs 14x the capacity of traditional data center interconnects. This shift is creating a multi-billion dollar market for 800G coherent pluggables and deep-buffered switches.
Microsoft's 2000 Nvidia Veto Rights Resurface Amid AI Chip Wars
A 2000 investment deal granted Microsoft veto rights over any acquisition of Nvidia. This historical clause gains new relevance as Nvidia's AI dominance makes it a potential target in the ongoing semiconductor consolidation.
Bull Delivers HPC Infrastructure to Power Mimer AI Factory
Bull, a subsidiary of Atos, has supplied the core HPC infrastructure for Mimer's new AI factory. This facility is dedicated to training and developing large language models for the European market.
AI Data Center Startup Phononic in Sale Talks at Multi-Billion Valuation
Phononic, a startup building liquid cooling systems for AI data centers, is in talks for a sale that could value it in the multi-billions. This reflects intense market pressure to solve the power and thermal challenges of scaling AI compute.
Anthropic Hiring Data Center Leasing Principals in Europe & Australia
Anthropic is actively hiring for data center leasing roles in Europe and Australia, revealing a strategic push to build out its own compute infrastructure as it scales its AI models.
Google, Marvell in Talks to Co-Develop New AI Chips, Including TPU-Optimized MPU
Google is reportedly in talks with Marvell Technology to co-develop two new AI chips: a memory processing unit (MPU) to pair with TPUs and a new, optimized TPU. This move is a direct effort to bolster Google's custom silicon stack and compete with Nvidia's dominance.
AMD Backs UALink Open Interconnect to Challenge NVIDIA NVLink in AI
AMD is supporting the newly formed UALink Consortium, which aims to create an open standard for connecting AI accelerators. This move challenges NVIDIA's control over the critical NVLink technology that underpins its AI data center systems.
Jensen Huang: Nvidia is a 'Computing Company,' Not a Car
Nvidia CEO Jensen Huang, in a new interview, argued that Nvidia is a 'computing company' and not a car—a product that can be easily interchanged. This distinction underscores Nvidia's strategy to be the indispensable platform for AI infrastructure.
DOE Seeks Input on AI Infrastructure for Federal Lands
The U.S. Department of Energy has published a Request for Information (RFI) to solicit input on developing AI and high-performance computing infrastructure on DOE-owned lands. This marks a significant step in the federal government's strategy to directly address the national AI compute shortage.
DOE's Portsmouth Site to Host World's Largest AI Data Center
A special report details plans for the world's largest AI data center at the DOE's Portsmouth, Ohio site, signaling a massive government-led expansion of compute capacity for AI research and national security applications.
IOWN Forum Pushes All-Photonic WAN for AI Neocloud Interconnects
The IOWN Global Forum is focusing its optical networking tech on datacenter interconnects, aiming to let GPU 'neoclouds' and financial firms use cheaper, remote facilities without latency penalties for AI workloads.
Nvidia Invests $2B in Marvell to Expand NVLink Fusion Chip Partnership
Nvidia is investing $2 billion in Marvell Technology to deepen their partnership on NVLink Fusion, a chip-to-chip interconnect crucial for scaling AI training clusters. This strategic move aims to secure supply and accelerate development of high-bandwidth links between GPUs and custom AI accelerators.
Meta Deploys Unified AI Agents to Manage Hyperscale Infrastructure
Meta's engineering team has built and deployed a system of unified AI agents to autonomously manage capacity and performance across its hyperscale infrastructure. This represents a significant shift from rule-based automation to AI-driven orchestration for one of the world's largest computing fleets.
Canada's AI Compute Gap: Google Cloud Montreal Offers 2017-Era Chips
A technical developer's attempt to rent modern AI compute in Canada revealed a stark infrastructure gap, with major providers offering chips as old as 2017, undermining national AI ambitions.
Cloud GPU vs. Colocation: H100 Costs $8k/Month on Google Cloud vs. $1k Colo
A technical founder highlights the stark economics: renting one H100 on Google Cloud costs ~$8,000/month, while the retail hardware is ~$30,000. At that rate, 4 months of cloud rental equals the cost of outright ownership, making colocation at ~$1k/month a compelling alternative for sustained AI workloads.
Meta Expands Broadcom Partnership for Next-Gen AI Infrastructure
Meta is expanding its partnership with semiconductor giant Broadcom to co-develop its next-generation AI infrastructure. This move signals a continued, long-term commitment to custom silicon for AI training and inference.
Mac Studio AI Hardware Shortage Signals Shift to Cloud Rentals
Developers report a global shortage of high-memory Apple Silicon Macs, with 128GB Mac Studios unavailable worldwide. This pushes practitioners toward renting cloud H100 GPUs at ~$3/hr, marking a shift from the recent local AI trend.
NVIDIA's cuQuantum-DGX OS Aims to Manage Hybrid Quantum-Classical Workflows
NVIDIA announced its AI software stack is evolving into an operating system for quantum computing, aiming to manage the complex workflow between quantum processors and classical GPUs. This targets a major integration bottleneck as quantum hardware scales.
Cloudflare Agent Cloud Integrates OpenAI GPT-5.4 & Codex for Enterprise AI
Cloudflare has integrated OpenAI's GPT-5.4 and Codex models into its Agent Cloud platform. This allows enterprises to build, deploy, and scale AI agents for production workflows with built-in security and performance.
Altimeter's Gerstner: AI Economics Shift to Owned Compute for Fixed Costs
Altimeter Capital's Brad Gerstner states the fundamental economics of AI have flipped, where companies owning their compute infrastructure lock in fixed costs while AI-driven revenue scales, creating a powerful advantage.
OpenAI Forecasts $121B in AI Hardware Costs for 2028
OpenAI is forecasting its own AI research hardware costs will reach $121 billion in 2028, according to a WSJ report. This figure highlights the extreme capital intensity required to compete at the frontier of AI.
Michael Burry Sees Anthropic as Direct Threat to Palantir's AI Narrative
Investor Michael Burry stated that Anthropic's ascent is emerging as a direct threat to Palantir's AI narrative, signaling a potential re-evaluation of competitive dynamics in the enterprise AI space.