Inside Amazon's Trainium Lab - How It Beat NVIDIA

An exclusive TechCrunch tour of Amazon's Trainium chip lab reveals how AWS is training Claude for Anthropic and now holds an $138B commitment from OpenAI.

Inside Amazon's Trainium Lab - How It Beat NVIDIA

Amazon's custom AI chips are training Claude. They're also the foundation of a $138 billion commitment from OpenAI. TechCrunch got an exclusive tour of the Trainium lab in Austin this weekend, and the picture that emerged is of a company that quietly built a credible alternative to NVIDIA - not by matching raw compute numbers, but by engineering around them.

TL;DR

  • Anthropic is running over 1 million Trainium2 chips to train Claude models; Amazon has rolled out 1.4 million Trainium chips total across all generations
  • As part of its $50B investment in OpenAI, Amazon secured an $138 billion chip commitment from OpenAI over eight years and 2 gigawatts of compute capacity
  • Trainium2 delivers 667 TFLOP/s - less than NVIDIA's GB200 at 2,500 - but costs 30-40% less per unit of useful work for training workloads
  • Trainium3 (current generation) runs on 3nm process with 2.52 PFLOPS FP8 and is 50% cheaper for inference than H100 clusters

The Numbers Behind the Win

Amazon CEO Andy Jassy put it plainly in a statement accompanying the TechCrunch report: "The two largest AI labs are both significantly betting on Trainium." That's not marketing copy. Anthropic has deployed more than one million Trainium2 chips for Claude training. OpenAI has now committed, as part of the Amazon $50B investment deal, to spend $138 billion on Amazon chips over the next eight years.

The raw specs tell only part of the story. Trainium2, the previous-generation chip, sits well below NVIDIA's top hardware on peak compute:

ChipPeak ComputeMemoryMemory BandwidthPrice-Performance vs H100
NVIDIA GB2002,500 TFLOP/s192 GB HBM3e8 TB/sbaseline
Amazon Trainium32,520 TFLOP/s FP8144 GB HBM3e4.9 TB/s+50% lower inference cost
Amazon Trainium2667 TFLOP/s~96 GB HBM2,900 GB/s30-40% better vs H100 P5e

Trainium2 losing to GB200 on raw TFLOP/s sounds damning until you read the SemiAnalysis breakdown: the chip's memory bandwidth per dollar is the variable that matters for large-scale reinforcement learning, and Amazon wins that comparison. Reinforcement learning - the method powering modern reasoning models - is memory-bound, not compute-bound.

Amazon's Trainium chip lab in Austin showing server racks Inside the Amazon Trainium lab in Austin, where Anthropic's Claude models are trained. Amazon built this facility faster than any data center in its history. Source: techcrunch.com

Inside the Trainium Architecture

Trainium3 - What Amazon Is Shipping Now

The current generation runs on a 3-nanometer process and delivers 2.52 PFLOPS of FP8 compute per chip, a 4.4x improvement over Trainium2. Memory has grown to 144 GB HBM3e per chip with 4.9 TB/s of bandwidth. Amazon's UltraServer configuration packs 144 Trainium3 chips into a single rack unit with 20.7 TB of aggregate HBM3e and 706 TB/s of aggregate memory bandwidth.

Scale matters. Amazon says these UltraServers can be linked up to one million chips - a figure that would represent a ten-fold jump from the previous generation's maximum cluster size.

The networking story is also new. NeuronLinkv3 is an all-to-all scale-up fabric that mirrors what NVIDIA built with NVLink, placing four switch trays in the center of each rack. Inter-chip communication latency sits below 10 microseconds. For the distributed training jobs running across hundreds of thousands of chips, that number directly affects training throughput.

Where Anthropic Sits in This Picture

According to the SemiAnalysis report published alongside the TechCrunch tour, Amazon is building more than 1.3 gigawatts of IT capacity across three AWS campuses specifically for Anthropic's training needs. Anthropic has "heavy involvement in all Trainium design decisions" - a co-design relationship that SemiAnalysis compares to Google DeepMind's deep integration with TPUs.

That's not a vendor-customer relationship. It's a hardware partnership where one of the world's most safety-focused AI labs is shaping the silicon roadmap of the world's largest cloud provider.

Amazon Trainium AI chip product view Amazon Trainium - Amazon's custom AI training silicon, now rolled out across more than 1.4 million chips in AWS data centers. Source: aws.amazon.com

How OpenAI Fits In

The OpenAI dimension is newer and structurally different. OpenAI didn't co-design Trainium the way Anthropic did. The commitment comes as part of the larger $50 billion investment deal, and it's mostly a procurement agreement: $138 billion over eight years in exchange for 2 gigawatts of Trainium3 and Trainium4 compute capacity. OpenAI becomes, in effect, a large customer rather than a hardware partner.

Trainium4 - scheduled for late 2026 or early 2027 - promises a 6x improvement in FP4 performance, 3x in FP8, and a 4x jump in memory bandwidth over Trainium3. If those numbers hold, OpenAI will be training on hardware that didn't exist when the deal was signed.

The Microsoft lawsuit over the exclusivity terms of the OpenAI deal adds a complication to this picture. Microsoft has alleged that the AWS agreement violates terms of its own agreements with OpenAI. Whatever the outcome, Amazon has already secured commitments that no chip manufacturer outside NVIDIA has come close to.

What It Does Not Tell You

The TechCrunch tour wasn't an arms-length inspection. It was an invited tour, which means Amazon controlled what journalists saw and what context they received. No third-party benchmarks from this specific facility have been published. The 30-40% price-performance advantage over H100 P5e instances is Amazon's own figure, and the comparison is against H100, not GB200 - the chip NVIDIA is currently shipping.

NVIDIA's Vera Rubin platform, announced at GTC 2026, promises 50 PFLOPS of inference per GPU and claims 10x lower token cost than Blackwell. If that holds at scale, the Trainium3 cost advantage narrows sharply. Amazon's own Trainium4 roadmap is built in part as a response.

The deeper question is software. NVIDIA's CUDA ecosystem has decades of optimization work behind it. Amazon's Neuron SDK is improving - Anthropic's co-design work has pushed it forward - but the software stack still requires meaningful porting effort for teams that built on CUDA. The chips being cheaper doesn't automatically mean they're easier to use.


For now, the scorecard reads: Anthropic trains on Trainium, OpenAI has committed to doing the same. Amazon has two of the four major frontier AI labs running their most expensive workloads on its custom silicon, with a third - Apple - reportedly assessing Trainium for on-device inference work, according to the TechCrunch report. NVIDIA remains the dominant player by installed base and by software ecosystem, but Amazon has made the case that you can build a serious alternative. The Trainium2 deployment numbers from Anthropic are proof it runs at scale. The $138 billion from OpenAI is proof someone is willing to bet on it.

Sources: TechCrunch - Exclusive tour of Amazon's Trainium lab · SemiAnalysis - Amazon's AI Resurgence · Yahoo Finance - Anthropic 1M Trainium2 chips · About Amazon - AWS OpenAI strategic partnership

Inside Amazon's Trainium Lab - How It Beat NVIDIA
About the author Senior AI Editor & Investigative Journalist

Elena is a technology journalist with over eight years of experience covering artificial intelligence, machine learning, and the startup ecosystem.