chips
30 articles about chips in AI news
Google's Virgo Network Links 134,000 TPU v8 Chips with 47 Pbps Fabric
Google unveiled its Virgo networking stack for TPU v8, capable of linking 134,000 chips in a single fabric with 47 petabits/sec of bi-sectional bandwidth. This represents a massive scale-up in interconnect technology for large-scale AI model training.
Google, Marvell in Talks to Co-Develop New AI Chips, Including TPU-Optimized MPU
Google is reportedly in talks with Marvell Technology to co-develop two new AI chips: a memory processing unit (MPU) to pair with TPUs and a new, optimized TPU. This move is a direct effort to bolster Google's custom silicon stack and compete with Nvidia's dominance.
Canada's AI Compute Gap: Google Cloud Montreal Offers 2017-Era Chips
A technical developer's attempt to rent modern AI compute in Canada revealed a stark infrastructure gap, with major providers offering chips as old as 2017, undermining national AI ambitions.
Anthropic Considers Custom AI Chips, Following Google & OpenAI
Anthropic is reportedly considering developing custom AI chips, a strategic move to gain control over its compute infrastructure and reduce costs. This follows similar initiatives by Google, Amazon, and OpenAI.
DeepSeek V4 to Run on Huawei Ascend 950PR Chips, Sparking 20% Price Surge
DeepSeek's anticipated V4 model will be powered by Huawei's Ascend 950PR chips, with Alibaba, ByteDance, and Tencent stockpiling hundreds of thousands of units ahead of launch. This has driven chip prices up approximately 20% in recent weeks.
AWS Commits 2 Gigawatts of Trainium Capacity to OpenAI, Reveals 1.4 Million Chips Deployed
Amazon's $50B OpenAI deal includes a 2-gigawatt commitment of Trainium computing capacity. AWS disclosed 1.4 million Trainium chips are deployed, with over 1 million Trainium2 chips running Anthropic's Claude.
Google's $1.9 Trillion Vertical Integration Strategy: Building an AI Empire from Chips to Power Grid
Google is investing $1.9 trillion over the next decade to control every layer of the AI stack, from custom TPU chips to power infrastructure. This vertical integration strategy creates a competitive moat that could reshape the entire AI industry landscape.
Broadcom to Manufacture Google TPU Chips in Foundry Partnership
Google has licensed its Tensor Processing Unit (TPU) intellectual property to Broadcom for chip fabrication. This allows Google to earn from its IP while Broadcom manages the complex hardware build and networking integration.
Apple's Neural Engine Jailbroken: Researchers Unlock Full Training Capabilities on M-Series Chips
Security researchers have reverse-engineered Apple's Neural Engine, bypassing private APIs to enable full neural network training directly on ANE hardware. This breakthrough unlocks 15.8 TFLOPS of compute previously restricted to inference-only operations across all M-series devices.
Meta to Cut 8,000 Jobs in May, Redirecting Capital to AI Infrastructure
Meta is reportedly planning to lay off 8,000 employees in May, the first round of major cuts this year. The move signals a capital shift from general operations to concentrated investment in AI infrastructure like chips and data centers.
TSMC's $56B 2026 CapEx Fuels AI Chip Race with 22 New Fabs
TSMC is constructing up to 22 advanced semiconductor fabs simultaneously, backed by a $52–56 billion capital expenditure plan for 2026. This unprecedented manufacturing scale is critical for producing the 2nm-and-below chips required by next-generation AI models.
Samsung Projects Record $14.6B Q1 Profit on 300% DRAM Price Surge
Samsung Electronics expects a record Q1 operating profit of 20 trillion won (~$14.6B), nearly triple YoY, fueled by soaring AI-driven demand and a 300% price increase for DRAM chips.
AWS CEO: All Latest Anthropic Models Trained on Amazon Trainium
Amazon Web Services CEO Matt Garman stated that all of Anthropic's latest AI models are trained on AWS's custom Trainium chips. This confirms the deepening technical and strategic integration between the AI lab and its primary cloud investor.
Intel & Google Announce Multiyear AI & Cloud Infrastructure Partnership
Intel and Google have announced a multiyear strategic collaboration to advance AI and cloud infrastructure, focusing on optimizing Google Cloud for Intel's Xeon processors, Gaudi AI accelerators, and future chips.
U.S. AI Data Center Builds Face 50% Delay Risk on China Power Gear
Electrical infrastructure, not chips or capital, is becoming the critical bottleneck for AI data center deployment. U.S. projects face 5-year transformer lead times while depending on China for 30-40% of key components.
Google's AICore Beta Enables On-Device Gemini Nano 4 Downloads for Android Phones
A new beta of Google's AICore system service enables users to download Gemini Nano 4 Full and Gemini Nano 4 Fast models directly onto compatible Android phones, including those with Snapdragon 8 Elite Gen 5 chips. This moves beyond pre-installed AI to user-initiated model management.
Sam Altman Steps Down from OpenAI Safety Oversight, Shifts Focus to Fundraising & Infrastructure
OpenAI CEO Sam Altman has reportedly stopped overseeing safety efforts at the company. His focus is now on fundraising, securing AI chips, and building data centers.
Apple's Private Cloud Compute: Leak Suggests 4x M2 Ultra Cluster for On-Device AI Offload
A leak suggests Apple's Private Cloud Compute for AI may be built on clusters of four M2 Ultra chips, potentially offering high-performance, private server-side processing for iPhone AI tasks. This would mark Apple's strategic move into dedicated, privacy-focused AI infrastructure.
AI Data Centers Now Consume 10% of US Electricity, With Single Facilities Reaching 400+ Megawatt Loads
Data centers powering AI and cloud computing now account for 10% of total U.S. electricity consumption, with individual facilities reaching 400+ megawatt capacities. New half-mile-long structures require advanced water-cooling systems to manage chips generating 2kW of heat each.
Nvidia Projects $1 Trillion in AI Chip Revenue Through 2027, According to Analyst
An analyst report, shared by Rohan Paul, projects Nvidia will generate $1 trillion in cumulative revenue from AI chips between 2024 and 2027. This forecast underscores the scale of infrastructure investment required for the current AI boom.
ASML's €350M EUV Lithography Machines Are the Unmatched Bottleneck for AI Chip Production
ASML's monopoly on Extreme Ultraviolet lithography machines, costing ~€350M each, is the critical enabler for advanced AI chips like the NVIDIA H100. Without its ~200 operational EUV systems, production of leading-edge semiconductors for models like GPT-4 and data centers would halt.
Nvidia's Jensen Huang Dismisses Custom AI Chip Threat: 'Science Projects' Versus 'AI Factories'
Nvidia CEO Jensen Huang confidently dismissed concerns about custom AI chips challenging Nvidia's dominance, framing competitors' efforts as 'science projects' while Nvidia builds revenue-generating 'AI factories' with a complete platform approach.
Nvidia's $2B Nebius Bet: Chip Giant Doubles Down on AI Infrastructure Empire
Nvidia will invest $2 billion in AI cloud specialist Nebius Group NV, expanding its strategic investments in companies that build data centers using its chips. The partnership aims to deploy over 5 gigawatts of AI-optimized data center capacity by 2030, equivalent to powering 4 million U.S. households.
Jensen Huang's '5-Layer Cake': Nvidia CEO Redefines AI as Industrial Infrastructure
Nvidia CEO Jensen Huang introduces a revolutionary framework positioning AI as essential infrastructure spanning energy, chips, infrastructure, models, and applications. This industrial perspective reshapes how we understand AI's technological and economic foundations.
The Invisible Dance: How AI Chip Manufacturing Relies on Microscopic Wire Bonding
High-speed semiconductor wire bonding creates thousands of electrical connections per minute using ultra-fine 25-micron wires. This critical but often overlooked process enables the AI chips powering today's most advanced systems.
China's Semiconductor Leaders Rally for National AI Chip Alliance Amid Tech War Escalation
China's top semiconductor executives have issued an unprecedented public call for a consolidated national effort to build AI chips, signaling a strategic shift toward self-reliance as U.S. export controls tighten. This coordinated push represents China's most direct response yet to technological containment efforts.
Apple's M5 Pro and Max: Fusion Architecture Redefines AI Computing on Silicon
Apple unveils M5 Pro and M5 Max chips with groundbreaking Fusion Architecture, merging two 3nm dies into a single SoC. The chips deliver up to 30% faster CPU performance and over 4x peak GPU compute for AI workloads compared to previous generations.
Firmus Secures Major AI Infrastructure Deal, Signaling Confidence in Australian Tech Market
Australian AI startup Firmus Technologies has signed a multi-billion dollar contract with a global tech firm for a Melbourne data center, deploying 18,400 Nvidia chips. The deal comes as the Nvidia-backed company prepares for an IPO later this year.
DeepSeek's Blackwell Training Exposes Critical Gaps in US Chip Export Controls
Chinese AI startup DeepSeek reportedly trained its latest model on Nvidia's restricted Blackwell chips, challenging US export controls. The development reveals significant loopholes in semiconductor restrictions amid escalating AI competition.
Nvidia's Record Earnings Mask China Dilemma: H200 Sales Frozen Amid AI Boom
Nvidia reported record quarterly revenue of $68.1 billion, up 73% year-over-year, driven by surging demand for data center processors. However, the company has generated zero revenue from its H200 chips in China and faces ongoing uncertainty about future sales in the critical market.