datacenter
19 articles about datacenter in AI news
CPU Demand Flipping the AI Narrative as Datacenter Growth Shifts
A new analysis from SemiAnalysis indicates CPU demand is rising in AI datacenters, reversing a narrative of GPU-only dominance. This shift signals changing workload patterns and infrastructure priorities.
AI Datacenter Spend Hits 5-7 Manhattan Projects Yearly at $250-300B
Inflation-adjusted global datacenter CapEx reaches $250-300B annually, equivalent to 5-7 Manhattan Projects per year. This quantifies the unprecedented infrastructure investment driving the AI boom.
IOWN Forum Pushes All-Photonic WAN for AI Neocloud Interconnects
The IOWN Global Forum is focusing its optical networking tech on datacenter interconnects, aiming to let GPU 'neoclouds' and financial firms use cheaper, remote facilities without latency penalties for AI workloads.
Utah Hyperscale Data Center to Exceed State Power Use
A hyperscale data center in Box Elder County, Utah, developed by Kevin O'Leary's O'Leary Digital, is set to generate and consume more power than the state itself, moving toward final approval.
Cisco Reveals Scale-Across GPU Networking Needs 14x DCI Bandwidth
Cisco's chief architect detailed the massive bandwidth requirements for connecting AI clusters via 'scale-across' GPU networking, which needs 14x the capacity of traditional data center interconnects. This shift is creating a multi-billion dollar market for 800G coherent pluggables and deep-buffered switches.
Microsoft's Fairwater AI Data Center Launches Early, Boosts Azure Capacity
Microsoft has launched its Fairwater AI data center ahead of schedule. The facility adds significant high-performance computing capacity to Azure's AI infrastructure, crucial for training and running large models.
DOE's Portsmouth Site to Host World's Largest AI Data Center
A special report details plans for the world's largest AI data center at the DOE's Portsmouth, Ohio site, signaling a massive government-led expansion of compute capacity for AI research and national security applications.
Meta Deploys Unified AI Agents to Manage Hyperscale Infrastructure
Meta's engineering team has built and deployed a system of unified AI agents to autonomously manage capacity and performance across its hyperscale infrastructure. This represents a significant shift from rule-based automation to AI-driven orchestration for one of the world's largest computing fleets.
Google, CoreWeave Sell Record $5.7B in Junk Bonds for AI Data Centers
Google and its partner CoreWeave sold a record $5.7 billion in high-yield bonds to fund AI data center expansion. The deal was oversubscribed, showing strong investor appetite for AI infrastructure debt.
Nvidia: Cost Per Token Is the Only AI Infrastructure Metric That Matters
Nvidia asserts that total cost of ownership for AI infrastructure must be measured in cost per delivered token, not raw compute metrics. This shift is critical for scaling profitable agentic AI applications.
Elice Group Expands AI Infrastructure with Modular Data Centers, Plans IPO
Elice Group, a Korean AI and EdTech company, is accelerating its AI infrastructure expansion using modular data centers and preparing for an initial public offering in 2026 to fuel growth.
Nvidia to Ship 1.19 Exabytes of HBM in 2026, Apple iPhone Memory 2x Larger
An analysis projects Nvidia will ship ~1.19 exabytes of HBM memory in 2026 for AI infrastructure, while Apple will ship ~2.4 exabytes of LPDDR5 for iPhones, putting AI's massive hardware scale in consumer market perspective.
Mac Studio AI Hardware Shortage Signals Shift to Cloud Rentals
Developers report a global shortage of high-memory Apple Silicon Macs, with 128GB Mac Studios unavailable worldwide. This pushes practitioners toward renting cloud H100 GPUs at ~$3/hr, marking a shift from the recent local AI trend.
OpenAI Finishes GPT-5.5 'Spud' Pretraining, Halts Sora for Compute
OpenAI has finished pretraining its next major model, codenamed 'Spud' (likely GPT-5.5), built on a new architecture and data mix. The company reportedly halted its Sora video generation project entirely, sacrificing a $1B Disney investment, to prioritize compute for Spud's launch.
OpenAI Expands Funding Round to $120B, Adds Andreessen Horowitz, TPG Ahead of Potential 2026 IPO
OpenAI has added $10B to its record funding round, bringing the total to over $120B. CFO Sarah Friar told CNBC this could be the company's final private raise before a potential IPO later this year.
OpenAI Shifts Sora Team to World-Model Research, Reportedly Cancels Video Model for Compute
A report claims OpenAI has redirected its Sora team to focus on world-model research for robotics and canceled the video model to free compute for a new, powerful LLM codenamed 'Spud.'
Jensen Huang Claims NVIDIA Has 'Achieved AGI' in Lex Fridman Interview, Sparking Industry Debate
NVIDIA CEO Jensen Huang stated in a Lex Fridman podcast interview that he believes his company has 'achieved AGI.' The brief, unverified claim has ignited immediate discussion about the definition and benchmarks for artificial general intelligence.
arXiv Survey Maps KV Cache Optimization Landscape: 5 Strategies for Million-Token LLM Inference
A comprehensive arXiv review categorizes five principal KV cache optimization techniques—eviction, compression, hybrid memory, novel attention, and combinations—to address the linear memory scaling bottleneck in long-context LLM inference. The analysis finds no single dominant solution, with optimal strategy depending on context length, hardware, and workload.
The Proxy-Free Web Scraping Revolution: How AI APIs Are Changing Data Collection
A new generation of web scraping APIs eliminates the need for manual proxy management, handling thousands of pages automatically while avoiding blocks. This represents a major shift toward AI-driven data collection infrastructure.