Cerebras IPO 20x Oversubscribed Signals AI Chip Crunch

Cerebras Systems raised its IPO price to $150-$160 per share after investors placed over $10 billion in orders on a $3.5 billion offering - 20 times oversubscribed.

Cerebras IPO 20x Oversubscribed Signals AI Chip Crunch

Blackwell GPU allocation slots for enterprise customers outside the hyperscaler tier now stretch past year-end 2026. Wait lists for dedicated NVIDIA capacity run twelve months or longer for organizations below the top tier of enterprise contracts. Into that supply gap, Cerebras Systems has raised its Nasdaq IPO price range to $150-$160 per share - up from $115-$125 - after investor books filled to more than 20 times the number of shares available.

The company set pricing for May 13. At $160 per share on 30 million shares, the offering raises $4.8 billion, up from the original $3.5 billion. Banks underwriting the deal received more than $10 billion in orders before a single share priced. Revenue for 2025 came in at $510 million, up 76% from $290 million in 2024, anchored by a $20 billion compute contract with OpenAI covering 750 megawatts of inference capacity through 2028.

TL;DR

  • IPO price raised to $150-$160/share from $115-$125; shares increased to 30M from 28M
  • Books filled to more than 20x available shares; investors placed $10B+ in orders on a $3.5B offering
  • 2025 revenue: $510M (up 76%), anchored by a $20B OpenAI Master Relationship Agreement for 750MW through 2028
  • OpenAI also holds warrants on 33M+ shares and loaned Cerebras $1B at 6% to build out data center infrastructure
  • Pricing set for May 13, Nasdaq ticker CBRS; the number tells you how far AI compute demand has outrun supply

The Constraint Behind the Surge

NVIDIA controls AI chip revenue and will for the foreseeable future. What's in question is whether NVIDIA can supply enough compute to meet demand growing faster than its fab allocations can cover. The Cerebras book-build suggests the answer is no, not yet.

Cerebras makes something architecturally different from any NVIDIA product. The Wafer-Scale Engine 3 (WSE-3) occupies an entire 300mm silicon wafer: 4 trillion transistors, 900,000 compute cores, built on TSMC's 5nm process. Instead of stacking chiplets with HBM memory connected through packaging, Cerebras puts the memory directly on the die - removing the bandwidth bottleneck that limits conventional GPU performance on long-context inference workloads. The CS-3 system built around the WSE-3 delivers 125 petaflops of AI compute from a single rack unit.

For large model inference at scale, that distinction has real consequences. Moving weights from HBM to compute cores accounts for most of the latency at context lengths above 32K tokens. Cerebras's architecture skips that transfer. The chip compares to NVIDIA's B200 as follows: 58x larger, 19x more transistors, 250x more on-chip memory, 2,625x more memory bandwidth - all per Cerebras's own product page, which the company publishes publicly.

A technician in cleanroom gear holds the Cerebras WSE-3 wafer-scale chip The WSE-3 covers a full 300mm silicon wafer - the largest chip ever built by area. Source: spectrum.ieee.org

Who Is Supplying What

SupplierTechnologyAnchor Customer2025-2026 RevenueWhere the Bottleneck Lives
NVIDIABlackwell GPU clusters, NVLink fabricAzure, GCP, AWS$140B+ annualizedCoWoS-S packaging, TSMC N4P allocation
CerebrasWSE-3 wafer-scale chip, CS-3 systemsOpenAI ($20B, 750MW)$510M (2025)TSMC 300mm wafer line capacity
GroqLPU inference acceleratorsMeta, StripeUndisclosed (private)LPU fab production ramp
TaalasModel-weight ASIC chipsTBDPre-revenueManufacturing scale-up

NVIDIA's constraint sits in TSMC's CoWoS-S advanced packaging capacity, which runs at full use through 2026 serving NVIDIA and Apple orders. Cerebras uses a different fabrication method completely - the wafer is the chip, no chiplet packaging required - giving it a different supply chain dependency. But it still competes for TSMC wafer starts, and each WSE-3 consumes a full 300mm production wafer.

The OpenAI relationship is more intertwined than a standard compute contract. In January 2026, OpenAI signed a Master Relationship Agreement worth more than $20 billion over three years, with an option to expand by another 1.25 gigawatts through 2030 - potentially pushing total contract value above $50 billion if exercised. OpenAI also loaned Cerebras $1 billion at 6% annual interest, secured by warrants allowing OpenAI to buy more than 33 million shares. One of the company's largest customers also holds a meaningful equity stake and has a direct financial interest in Cerebras reaching this IPO.

The original Cerebras IPO roadshow in April targeted a $22-25 billion valuation. Seven weeks of investor meetings, and the final price range, suggest that investors read the demand signal differently than the initial filing.

Reuters, citing people with knowledge of the deal, reported May 10 that the IPO price range had been raised to $150-$160 per share after books filled to more than 20 times the number of shares on offer.

Who Gets Squeezed

Startups Waiting on Allocations

A startup running large-model inference at low latency in 2026 faces two realistic options: pay cloud providers' premium API pricing, or join a twelve-month wait list for dedicated NVIDIA capacity. H100 spot pricing on secondary markets runs 3-5x list price. Neither is acceptable when latency and cost targets are contractual obligations.

Cerebras's on-chip memory architecture handles exactly this class of problem. Long-context inference - where moving weights from HBM consumes most of the wall-clock time - runs faster on WSE-3 than on conventional GPU stacks. The performance advantage isn't a benchmark artifact; it's a physical consequence of eliminating the memory transfer step.

Inference Providers Locked Into Spot Markets

Real-time inference providers need predictable capacity at a predictable price. NVIDIA's enterprise allocation system doesn't offer that: wait lists shift, spot pricing swings by the quarter, and hyperscaler cloud pricing varies with demand. The Cerebras model - a fixed multi-year compute contract at a contractual price - is the structure inference providers want when building products on fixed gross margins.

NVIDIA committed $40 billion in equity deals to AI customers this year, keeping customers close when hardware allocations can't. Cerebras doesn't need that mechanism for OpenAI; the $1 billion loan and the warrants have already aligned interests more directly than any equity side deal.

The Cerebras CS-3 AI supercomputer system in a data center rack The CS-3 system, which houses the WSE-3 and powers the OpenAI compute infrastructure underpinning the IPO's anchor revenue. Source: cerebras.ai

Mid-Size Enterprises Priced Out

Mid-size enterprises without hyperscaler-level purchasing power watch AI compute as a rising cost line even as NVIDIA reports record earnings. The NVIDIA-Groq inference chip deal with OpenAI shows where NVIDIA's engineering investment is flowing - toward the customers it already has, not the queue below them.

What Breaks First

The 20x oversubscription creates a price anchor for the broader AI chip alternatives market. At $4.8 billion raised on $510 million of 2025 revenue, institutional investors have publicly uncovered what they'll pay for a credible non-NVIDIA compute source backed by a real anchor customer. That multiple will follow Groq to its eventual IPO filing and Taalas to its next funding round.

NVIDIA's Blackwell wait times extend if yields at TSMC slip or if hyperscaler demand continues to accelerate ahead of capacity forecasts. A second delay on Blackwell production in 2026 would sharpen demand for alternatives further, and Cerebras would be the only publicly traded company positioned to capture that directly.

TSMC capacity expansion into 2027 and 2028 - with new fabs in Arizona and Japan - is the factor that eventually relieves this constraint. When that capacity comes online on schedule, pricing pressure on GPU alternatives softens and the revenue multiple Cerebras is capturing today becomes harder to sustain at IPO-level prices.

Cerebras prices on May 13 under ticker CBRS. Trading begins May 14.


Sources:

Sophie Zhang
About the author AI Infrastructure & Open Source Reporter

Sophie is a journalist and former systems engineer who covers AI infrastructure, open-source models, and the developer tooling ecosystem.