Overview
On May 6, 2026, Anthropic packaged two announcements together: (1) higher usage limits across Claude Code and the Claude API, and (2) a new compute partnership with SpaceX. The second causes the first. The headline reads “higher limits,” but the real story is that Anthropic has leased the entire Colossus 1 supercomputer — originally built by direct rival xAI — and is converting that capacity into raised user limits within a month.
What Changed — Three Limit Bumps
The announcement lists three changes, all effective immediately:
| Change | Detail |
|---|---|
| Claude Code 5-hour rate limit | Doubled for Pro, Max, Team, and seat-based Enterprise plans |
| Claude Code peak-hour throttle | Removed for Pro and Max accounts |
| Claude API rate limits | Substantially raised for Opus models — see the API rate-limits docs |
Note that the API bump is scoped to Opus. Sonnet and Haiku are not called out. Opus is the most expensive line and the one used for frontier reasoning workloads — so the freshly-arrived GPUs are being routed first to unlock the most expensive inference, not to relax limits across the board.
The New Compute — All of Colossus 1
The headline numbers:
- 300MW+ of new capacity
- 220,000+ NVIDIA GPUs — mix of H100, H200, and next-gen GB200 accelerators
- Online within the month
- Location: the former Electrolux factory in Memphis’s Boxtown district
That cluster was originally stood up in record time by xAI to train Grok. The same-day SpaceXAI counterpart announcement confirms the framing:
“SpaceXAI has signed an agreement with Anthropic to provide access to Colossus 1… Anthropic plans to use this additional compute to directly improve capacity for Claude Pro and Claude Max subscribers.”
In effect, xAI is pivoting to Colossus 2 and handing first-gen Colossus to a direct competitor. Elon Musk’s public comment: “No one set off my evil detector.”
Anthropic’s Full Compute Portfolio
The SpaceX deal is the latest piece in a six-month run of megadeals.
| Partner | Scale | Timing | Source |
|---|---|---|---|
| Amazon (Trainium) | up to 5GW, ~1GW new by end of 2026 | In progress | official |
| Google (TPU) + Broadcom | 5GW, coming online 2027 | Future | official |
| Microsoft + NVIDIA | $30B of Azure capacity | Strategic | official |
| Fluidstack (US infra) | $50B Anthropic-funded | Multi-year | official |
| SpaceX / xAI | 300MW+, 220K GPUs | Immediate (~1 month) | official |
The official post explicitly names three accelerator families — AWS Trainium, Google TPU, and NVIDIA GPUs — for training and serving Claude. The implicit thesis is that single-silicon lock-in is the biggest infrastructure risk, and the SpaceX deal pads out the NVIDIA leg immediately.
How Rate Limits Are Layered — Where the Bump Lands
It helps to remember Anthropic’s API limit structure before reading the announcement. The rate-limits docs split it into two:
- Spend limits — monthly cap. Tier 1 ($100) → Tier 2 ($500) → Tier 3 ($1,000) → Tier 4 ($200,000) → Monthly Invoicing (no cap).
- Rate limits — per-minute RPM / TPM, model-by-model.
On top, Service Tiers layer a separate availability dimension:
- Priority Tier — committed spend buys SLA-grade availability and predictable pricing. Surfaced via headers like
anthropic-priority-input-tokens-limit. - Standard — default.
- Batch — async workloads that can run outside normal capacity.
What this announcement actually moved: Standard Tier Opus RPM/TPM and Claude Code’s 5-hour window. Priority Tier itself is not called out as changed — Priority already had reserved capacity, so the freshly-landed GPUs appear to be allocated first to lifting the Standard-tier ceiling that most subscribers actually hit.
Alongside — How Rivals Do This
Frontier LLM vendors using capacity announcements as marketing assets isn’t new.
- OpenAI × Microsoft — the Stargate Project, joined by Oracle and SoftBank, pursuing tens of gigawatts.
- OpenAI × AMD — multi-year GPU supply with AMD share warrants.
- OpenAI × Broadcom — co-developing a custom AI accelerator.
The grammar is consistent across these: (a) gigawatt-scale numbers, (b) multi-year commitments, (c) explicit promises of improved end-user experience. Anthropic’s announcement follows the same template with one twist — renting a rival’s existing frontier cluster wholesale instead of building net-new.
What This Is and Isn’t
It is:
- Proof that a market exists for taking over a competitor’s frontier supercomputer at month-scale notice. AI infrastructure is starting to trade like a vendor-neutral commodity.
- Speed news. 300MW typically takes 18-24 months to bring online from scratch; this lands in one.
- An explicit four-leg compute strategy: Trainium + TPU + NVIDIA + flexible leased capacity.
It isn’t:
- A model upgrade. Opus, Sonnet, Haiku are untouched.
- A price change. Pricing is the same.
- A new enterprise SKU. Priority Tier terms aren’t called out as changed.
Orbital Compute — One More Line
The Anthropic post closes with a line about “expressed interest in partnering with SpaceX to develop multiple gigawatts of orbital AI compute capacity.” The SpaceXAI side is more direct:
“SpaceX is the only organization with the launch cadence, mass-to-orbit economics, and constellation operations experience to make orbital compute a near-term engineering program rather than a research concept.”
Not a near-term deliverable. But it’s the first time both sides have put orbital AI compute — sidestepping terrestrial power/cooling/siting limits via Starlink-adjacent infrastructure — into a joint official document.
Takeaways
One-line summary: “To raise subscriber limits, Anthropic rented a rival’s entire supercomputer.”
Three implications:
- AI capacity is starting to trade like a commodity. A running, frontier-class cluster — GPUs, power, cooling, networking all already wired — can be taken over by a rival on month-scale terms. That’s a market-maturity signal.
- Multi-silicon strategy is now table stakes. Anthropic has four legs: Trainium, TPU, NVIDIA, and leased capacity. The redundancy reduces single-incident risk and provides routing flexibility — whichever leg comes online fastest gets translated directly into user-visible limit bumps.
- For end users, it’s simple. Pro / Max subscribers get more Claude Code uninterrupted: doubled 5-hour window, no peak-hours throttle, and bigger Opus API ceilings, all landing together.
Signals to watch next: (a) whether the Standard-tier RPM/TPM tables in the docs actually update with new numbers, (b) whether Priority Tier sees matching capacity bumps, (c) when “orbital compute” turns from intent into a dated roadmap.
References
Primary announcements
- Anthropic: Higher usage limits for Claude and a compute deal with SpaceX
- xAI/SpaceXAI: New Compute Partnership with Anthropic
Anthropic compute megadeal series
- Anthropic × Amazon — up to 5GW
- Anthropic × Google × Broadcom — 5GW
- Anthropic × Microsoft × NVIDIA strategic partnerships
- Anthropic’s $50B US AI infrastructure investment with Fluidstack
- Covering data-center-driven electricity price increases
Anthropic platform docs
- API Rate Limits · Service Tiers (Priority/Standard/Batch)
- Pricing · Enterprise plan · Max plan · Team plan
- Claude Code · Claude Code Enterprise
- Models: Opus · Sonnet · Haiku
Colossus 1 / Memphis background
- Tom’s Hardware: SpaceX rents Colossus to rival Anthropic; Musk’s “evil detector”
- DCD: Anthropic to use all of SpaceX-xAI’s Colossus 1 capacity
- Capacity: Anthropic secures full capacity of Memphis data centre
- Wikipedia: Colossus (supercomputer)
Comparison — competitor megadeals
