Inside Anthropic's $200B Google Cloud Compute Bet

Anthropic has committed $200 billion to Google Cloud over five years - the largest cloud contract in AI history - alongside a 3.5 GW TPU capacity deal with Google and Broadcom coming online in 2027.

Inside Anthropic's $200B Google Cloud Compute Bet

The number is $200 billion. Five years. One counterparty: Google. Anthropic's compute commitment, reported Tuesday by The Information, is the largest cloud contract in AI history - and the structure behind it shows what happens when a company's revenue growth outpaces its ability to provision hardware fast enough.

TL;DR

  • Anthropic commits $200B to Google Cloud and chips over 5 years - more than 40% of Google's disclosed cloud revenue backlog
  • A separate Broadcom SEC filing confirms 3.5 GW of next-gen TPU capacity reserved for Anthropic, coming online 2027 (tripling the 1 GW deal signed in October 2025)
  • Anthropic's run-rate revenue has gone from $9B at year-end 2025 to roughly $40B by late April 2026 - demand is the easy part; supply is the problem
  • Anthropic trains across AWS Trainium, Google TPUs, and NVIDIA GPUs simultaneously - a three-hardware strategy at a scale very few companies have ever attempted

The Compute Math

$200B over five years

Spread evenly, that is $40B per year in cloud spend. Anthropic's late-April run rate sits at roughly $40B in revenue, which means the company is committing to a compute bill that could consume most of a year's revenue over the next five years. That isn't a hedge - it's a strategic bet that demand will continue to outpace compute availability, and that locking capacity now is cheaper than scrambling for it later.

Anthropic CFO Krishna Rao framed it explicitly:

"This groundbreaking partnership with Google and Broadcom is a continuation of our disciplined approach to scaling infrastructure: we are building the capacity necessary to serve the exponential growth we have seen in our customer base while also enabling Claude to define the frontier of AI development."

The "exponential growth" claim isn't marketing copy this time. Run-rate revenue went from $9B at year-end 2025, to $19B by March 2026, to $30B by the time the April Broadcom deal closed, to an estimated $40B by late April. The company also crossed 1,000 enterprise customers spending $1 million or more annually in the same period - doubling from the 500 reported in February.

3.5 gigawatts

The raw capacity deal is a separate agreement, disclosed in a Broadcom SEC filing and first reported by TechCrunch in April. Anthropic is reserving 3.5 GW of compute capacity from Google and Broadcom, tripling the 1 GW deal it signed in October 2025. This capacity comes online starting in 2027, which means the company is making long-horizon infrastructure bets that extend well beyond its current model generation.

The hardware behind the 3.5 GW reservation is almost certainly Google's eighth-generation TPU family, which Google unveiled at Cloud Next on April 22. The TPU 8t (training) and TPU 8i (inference) are two specialized chips built for the agentic era.

A close-up of a semiconductor chip, representative of the AI accelerator hardware at the center of Anthropic's compute strategy Google announced its eighth-generation TPU 8t and 8i chips on April 22 at Cloud Next. Anthropic's 3.5 GW reservation is widely expected to run on this generation, which delivers 121 exaflops per superpod. Source: pexels.com

The Hardware: TPU 8t and TPU 8i

The TPU 8t is a training-focused chip. A single TPU 8t superpod scales to 9,600 chips, delivers 121 exaflops of FP4 compute, and holds two petabytes of shared high-bandwidth memory. Google says it delivers nearly 3x the compute performance of the previous Ironwood generation for the same price, and targets over 97% "goodput" - the fraction of compute time actually producing useful work rather than waiting on memory or network transfers.

The TPU 8i handles inference. It ships with 288 GB of high-bandwidth memory per chip and 384 MB of on-chip SRAM (triple the previous generation), plus a Collectives Acceleration Engine that cuts on-chip latency by 5x. Performance-per-dollar improves 80% versus Ironwood for inference workloads. At production scale - where Claude serves millions of requests across consumer and enterprise products - inference efficiency compounds into hundreds of millions of dollars in saved compute costs per year.

SpecTPU 8t (Training)TPU 8i (Inference)
Per-superpod chips9,600Not yet disclosed
Compute (FP4)121 ExaFlopsNot yet disclosed
HBM per chip2 PB (shared)288 GB
On-chip SRAMNot yet disclosed384 MB (3x prev gen)
Latency improvement-5x (vs Ironwood)
Performance/dollar2.8x vs Ironwood1.8x vs Ironwood
General availabilityLater 2026Later 2026

Both chips are expected to be generally available to cloud customers later in 2026 - just in time to inform the infrastructure that'll come online for Anthropic in 2027.

Anthropic's Three-Cloud Strategy

Despite the scale of the Google commitment, Anthropic doesn't run on a single cloud. It trains and deploys Claude across three hardware families simultaneously.

Google Cloud TPUs

The primary training partner for Claude's most recent model generations. Google has been Anthropic's anchor investor since 2023, and the relationship has deepened with each funding round. The new $200B commitment and 3.5 GW capacity reservation cement Google's position as Anthropic's compute spine.

AWS Trainium

Amazon remains the primary cloud and deployment partner. Claude is available on AWS Bedrock, and Anthropic has a separate multi-gigawatt capacity agreement with Amazon chips (nearly 1 GW, expected by year-end). Amazon was Anthropic's first major hyperscaler partner and continues to process the majority of production inference traffic.

NVIDIA GPUs

Anthropic uses NVIDIA hardware for workloads that benefit from the CUDA ecosystem - fine-tuning, research runs, and tasks where software tooling availability outweighs raw hardware efficiency. NVIDIA's H100 and H200 clusters remain the most widely available high-performance compute globally, which gives Anthropic flexibility.

CloudRoleHardwareStatus
Google CloudPrimary training + storageTPU 8t, TPU 8i$200B, 5-year commitment
AWSPrimary deployment + inferenceTrainium, NVIDIANearly 1 GW by year-end
AzureEnterprise distributionMixedAvailable on Azure Foundry
NVIDIA (colocation)Research, fine-tuning, flexH100, H200Ongoing

What This Means for Broadcom

The financial math for Broadcom is striking. Mizuho analysts estimate that Broadcom will recognize $21 billion in AI revenue from Anthropic in 2026, rising to $42 billion in 2027 when the 3.5 GW capacity comes online. Broadcom manufactures the custom silicon central to Google's TPU line - effectively, every dollar Anthropic spends on Google TPUs flows partly through Broadcom's fabrication agreements with TSMC.

A data center server corridor stretching into the distance, illustrating the scale of compute infrastructure that major AI labs now require Anthropic trains Claude across AWS Trainium, Google TPUs, and NVIDIA GPUs - a three-hardware strategy that requires managing massive, distributed data center infrastructure. Source: pexels.com

For context on the scale of the AI capex supercycle, Anthropic alone is creating more revenue for Broadcom next year than most mid-cap technology companies create in total. The AI accelerator market is concentrating around a small number of hyperscale buyers, and Anthropic is now clearly in that tier.

Where It Falls Short

Regulatory exposure

In March 2026, the US Defense Department labeled Anthropic a supply-chain risk - a designation that triggered a wave of enterprise concern before Anthropic won an injunction against the Trump administration's classification. A $200B cloud commitment concentrated in a single US-domiciled provider creates a single point of legal and regulatory exposure. If that designation resurfaces, or if export control regimes tighten around AI compute, the concentration boosts the risk.

Vendor lock-in at scale

Three-cloud training sounds like resilience. In practice, the $200B Google commitment will pull engineering investment disproportionately toward TPU-native tooling, optimization, and deployment patterns. JAX, XLA, and the TPU software stack aren't portable. The code written to squeeze performance out of TPU 8t superpods does not run cleanly on Trainium or H100s without significant re-engineering. The diversification is real, but the $200B tilts the gravity well toward Google in ways that'll compound over years.

The revenue assumption

The commitment assumes Anthropic's revenue arc continues. The company is projecting $70 billion in revenue by 2028 - a number that would make the $40B annual compute spend manageable. If enterprise demand for AI coding agents, which currently drives most of Anthropic's growth, plateaus or shifts to open-weight competitors, a $200B cloud bill becomes a structural problem rather than a growth investment. Anthropic's recent enterprise push into financial services and multi-agent workflows is partly a hedge against that concentration.


The $200B number is big enough to be a headline. The 3.5 GW number is big enough to reshape how Google's hyperscale infrastructure roadmap gets focused on over the next two years. Taken together, they describe a company that has decided the compute constraint is more dangerous than the cost constraint - and is willing to bet its balance sheet on being right.

Sources:

Sophie Zhang
About the author AI Infrastructure & Open Source Reporter

Sophie is a journalist and former systems engineer who covers AI infrastructure, open-source models, and the developer tooling ecosystem.