Skip to content
gentic.news — AI News Intelligence Platform
Connecting to the Living Graph…

Listen to today's AI briefing

Daily podcast — 5 min, AI-narrated summary of top stories

A vast data center interior with rows of server racks and cooling pipes, likely the Colossus 1 supercomputer, leased…
Big TechBreakthroughScore: 92

Anthropic Leases xAI's Colossus 1 After Mixed-Architecture Flaw Blocked

Anthropic leased xAI's 220K-GPU Colossus 1 after its mixed architecture failed to train Grok. Musk builds Blackwell-only Colossus 2 for training and IPO.

·21h ago·2 min read··1 views·AI-Generated·Report error
Share:
Source: tomshardware.comvia tomshardwareSingle Source
Why did Anthropic lease xAI's Colossus 1 supercomputer?

Anthropic leased xAI's entire 220,000-GPU Colossus 1 supercomputer from SpaceX for inference, after the mixed-architecture design proved too inefficient to train Grok. Musk is building a Blackwell-only Colossus 2 for frontier training and a potential IPO.

TL;DR

Anthropic leased xAI's 220K-GPU Colossus 1. · Colossus 1's mixed architecture couldn't train Grok. · Musk plans unified Blackwell-only Colossus 2.

Anthropic leased xAI's entire 220,000-GPU Colossus 1 supercomputer from SpaceX for inference. The mixed-architecture design was too inefficient to train Grok, per Tom's Hardware.

Key facts

  • 220,000 GPUs in Colossus 1 cluster.
  • Mixed H100/H200/MI300X architecture caused training bottlenecks.
  • Colossus 2 will be unified Blackwell-only.
  • Anthropic raised $11.5B+ total funding.
  • xAI considering IPO with Colossus 2 as asset.

Anthropic has leased xAI's entire 220,000-GPU Colossus 1 supercomputer from SpaceX to run Claude inference workloads, according to Tom's Hardware. The deal addresses Anthropic's growing compute bottlenecks as Claude usage scales—Claude Code alone appeared in 693 prior articles on gentic.news—but the arrangement reveals deeper structural flaws in the cluster's design.

Why Colossus 1 Failed at Training

The mixed-architecture design—combining NVIDIA H100s and H200s with AMD MI300X accelerators—produced inter-GPU communication bottlenecks that made it unsuitable for training large frontier models like Grok. [According to Tom's Hardware], the inefficiency was so severe that xAI couldn't effectively train its own model on the cluster. Instead, the 220,000 GPUs now handle inference workloads for Anthropic's Claude models, including Claude Opus 4.6 and Claude Sonnet 4.6.

Colossus 2: A Clean-Slate Bet

Musk is now preparing Colossus 2, a unified Blackwell-only cluster designed specifically for training frontier models and supporting a potential xAI IPO. [The source reports] that the new cluster will avoid the heterogeneity issues that plagued Colossus 1. The move aligns with xAI's broader infrastructure ambitions, including orbital data centers, as noted in the knowledge graph.

Strategic Implications

For Anthropic, the Colossus 1 deal signals willingness to bypass traditional cloud providers for raw compute capacity. The company, which has raised over $11.5B and is projected to surpass OpenAI in ARR by mid-2026, now has a dedicated inference farm outside AWS/GCP. But the arrangement also ties Anthropic's inference infrastructure to a competitor's hardware—a risk if xAI reprioritizes Colossus 1 for its own needs.

The Colossus 1 cluster is reportedly powered by mobile gas turbines and diesel generators, raising sustainability questions that may conflict with Anthropic's stated safety and ethics commitments.

What to watch

Watch for xAI's Colossus 2 deployment timeline and whether Anthropic extends the Colossus 1 lease beyond inference workloads. Also track any IPO filing from xAI that discloses Colossus 2 as a key asset.


Sources cited in this article

Source: gentic.news · · author= · citation.json

AI-assisted reporting. Generated by gentic.news from 3 verified sources, fact-checked against the Living Graph of 4,300+ entities. Edited by Ala SMITH.

Following this story?

Get a weekly digest with AI predictions, trends, and analysis — free.

AI Analysis

The Colossus 1 lease is a pragmatic but risky bet for Anthropic. On one hand, it secures massive inference capacity outside the Big Three cloud providers—critical as Claude usage explodes (Claude Code appeared in 693 articles on this site alone). On the other, it ties inference infrastructure to a direct competitor's hardware, giving xAI leverage if Musk decides to reprioritize. The mixed-architecture failure is instructive: AI infrastructure is still so immature that even a 220K-GPU cluster can be rendered useless for training by poor integration choices. Colossus 2's unified Blackwell approach suggests xAI learned the lesson, but the real test will be whether NVIDIA's NVLink can scale to that size without similar bottlenecks. The sustainability angle—gas turbines and diesel generators—creates a tension with Anthropic's safety-first branding that will likely surface in future policy discussions.
Compare side-by-side
Anthropic vs xAI
Enjoyed this article?
Share:

AI Toolslive

Five one-click lenses on this article. Cached for 24h.

Pick a tool above to generate an instant lens on this article.

Related Articles

From the lab

The framework underneath this story

Every article on this site sits on top of one engine and one framework — both built by the lab.

More in Big Tech

View all