AMD Instinct MI430X - Dual-Precision CDNA 5 Accelerator

The AMD Instinct MI430X is AMD's CDNA 5 HPC accelerator with 432GB HBM4, full FP64 support, and 19.6 TB/s bandwidth - designed for sovereign AI and scientific supercomputing alongside the MI455X AI GPU.

AMD Instinct MI430X - Dual-Precision CDNA 5 Accelerator

TL;DR

  • AMD's HPC-focused chip in the MI400 series: full FP64/FP32 support alongside FP8/FP4 for AI
  • 432GB HBM4 at 19.6 TB/s - 50% more memory than NVIDIA's Vera Rubin GPU, same bandwidth
  • Estimated ~211 TFLOPS FP64 vector performance (derived from Alice Recoque supercomputer specs)
  • Powers Discovery at Oak Ridge National Lab and Alice Recoque in France; ships H2 2026

Overview

AMD's MI400 series for 2026 splits into three distinct products: the MI455X for AI training and inference, the MI440X for enterprise AI, and the MI430X for high-performance computing and sovereign AI. The MI430X isn't mostly an AI accelerator in the deep learning sense. It's what AMD builds for the customers who still need to run physics simulations, climate models, and molecular dynamics in FP64 - workloads that GPUs have always been capable of but that AI chips increasingly deprioritize.

The distinction matters practically. The AMD MI455X and MI440X squeeze performance by focusing compute resources on FP8 and FP4 precision, where modern AI training and inference operate. The MI430X trades some of that low-precision density for dedicated high-precision compute paths. It runs FP64 and FP32 without software emulation, making it compatible with legacy HPC code that hasn't been and won't be ported to lower precision.

AMD announced the MI430X at CES 2026 with the full Helios rack architecture. It's bound for two high-profile deployments: Discovery at Oak Ridge National Laboratory in the United States and Alice Recoque in France. Both are major scientific computing facilities that need the HPC precision the MI455X doesn't offer, while still wanting the memory capacity and bandwidth to run large AI workloads side by side.

Key Specifications

SpecificationDetails
ManufacturerAMD
Product FamilyInstinct MI400 (CDNA 5)
Chip TypeGPU
Process NodeTSMC N2 (2nm, compute chiplets) + N3P (3nm, IO)
Memory432GB HBM4
Memory Bandwidth19.6 TB/s
FP64 Performance~211 TFLOPS (estimated, Chips and Cheese analysis)
FP8 PerformanceNot officially disclosed
FP4 PerformanceNot officially disclosed
FP64 SupportYes (native hardware)
FP32 SupportYes (native hardware)
TDPNot disclosed
InterconnectInfinity Fabric + UALink
TargetHPC, Sovereign AI, Mixed Precision
Release DateH2 2026
PricingNot disclosed

Performance Benchmarks

AMD hasn't published official FP64 numbers for the MI430X. The best available estimate comes from Chips and Cheese analysis of the Alice Recoque supercomputer: using the facility's announced 1 Exaflop HPL target and expected GPU count, the MI430X is estimated at around 211 TFLOPS FP64 vector performance. That's a major jump from the MI355X's 77 TFLOPS FP64, reflecting the architectural changes in CDNA 5 that increase the FP64 compute ratios.

MetricAMD MI430XAMD MI455XNVIDIA Vera Rubin (est.)AMD MI300X
FP64 TFLOPS~211 (est.)Not primary~45 (est.)163.4
HBM Capacity432GB HBM4432GB HBM4288GB HBM192GB HBM3
Memory Bandwidth19.6 TB/s19.6 TB/s22 TB/s5.3 TB/s
FP8 PerformanceNot focus20,000 TFLOPS~50,000 TFLOPS5,220 TFLOPS
Process NodeTSMC N2TSMC N2TSMC N2TSMC N5
TargetHPC + AIAI Training/InferenceAI Training/InferenceAI/HPC

The MI430X has more memory than NVIDIA's Vera Rubin GPU - 432GB versus 288GB - at similar bandwidth. For scientific applications where entire datasets need to fit in a single GPU's memory to avoid inter-node communication, this capacity advantage matters. For AI training, the MI455X is the better AMD option; for HPC, the MI430X is the only AMD option with full FP64 support in the current lineup.

AMD Instinct MI430X GPU system configuration for the Discovery supercomputer deployment AMD's MI430X GPU in the system configuration bound for Oak Ridge's Discovery supercomputer - pairing CDNA 5 compute with EPYC Venice CPUs on HPE Cray GX5000 nodes. Source: TweakTown / AMD

Key Capabilities

The defining feature of the MI430X is the chiplet design. AMD uses CDNA 5 compute chiplets on TSMC N2 - the same 2nm-class node as the MI455X - but builds the MI430X with a different chiplet mix. The MI455X and MI440X focus on AI tensor cores for FP4/FP8 operations. The MI430X allocates die area for full-width FP64 compute paths, which are expensive in silicon area but necessary for scientific accuracy.

This "dual-precision" design is what AMD means by flexibility: the same chip can run FP64 for an atmospheric modeling code in the morning and FP8 for a language model fine-tuning job in the afternoon. Neither mode requires software tricks or precision emulation. The flexibility comes at a cost - the MI430X won't match the MI455X on pure AI training throughput because some die area went to FP64 logic that AI workloads don't use.

Both Infinity Fabric for scale-up connectivity and UALink for rack-scale interconnects are supported, putting the MI430X in AMD's Helios rack architecture alongside the MI455X. This matters for institutions like Oak Ridge that need to run heterogeneous workloads across the same physical infrastructure without separate GPU pools for science and AI.

Software Stack

AMD's ROCm software stack supports the MI430X across PyTorch, TensorFlow, and JAX for AI workloads, and AMD supports standard HPC libraries including OpenMP offloading and HIP for CUDA-like programming. The MI430X should be compatible with the HPC code that MI300X users have already ported, though AMD hasn't published detailed migration guidance.

For HPC codes that rely on NVIDIA's CUDA ecosystem, the situation is unchanged from prior AMD generations: porting effort is required. Discovery at Oak Ridge will be one of the largest MI430X deployments and AMD's most visible test of whether CDNA 5 can sustain complex scientific workloads at scale.

Pricing and Availability

No pricing has been disclosed. The MI430X is slated for H2 2026 availability, which aligns with the Discovery supercomputer installation timeline. For context, MI300X systems from cloud providers run $6-10/hour for a 8-GPU node; MI430X cloud pricing, if offered, would likely be in a similar range.

The primary channel for the MI430X is institutional procurement through AMD's OEM partners (HPE Cray, Lenovo, Dell), not direct data center orders. Individual cloud instances haven't been announced for this chip, given its HPC positioning.

Strengths

  • Full native FP64 and FP32 support - critical for scientific computing codes
  • 432GB HBM4 capacity exceeds NVIDIA Vera Rubin GPU by 50% (288GB)
  • Same memory bandwidth as Vera Rubin (19.6 TB/s)
  • Dual-precision flexibility - runs both HPC and AI workloads on one chip
  • CDNA 5 chiplet design on TSMC N2 (same advanced node as MI455X)
  • Part of AMD Helios rack architecture with UALink scale-out

Weaknesses

  • FP8/FP4 AI performance won't match MI455X - die area is split with FP64 compute paths
  • No confirmed FP8 TFLOPS published; hard to benchmark against pure AI chips
  • H2 2026 availability - not shipping yet
  • Primary channel is institutional procurement, not general availability
  • ROCm ecosystem still lags CUDA in third-party framework support depth

Sources

✓ Last verified April 15, 2026

AMD Instinct MI430X - Dual-Precision CDNA 5 Accelerator
About the author AI Benchmarks & Tools Analyst

James is a software engineer turned tech writer who spent six years building backend systems at a fintech startup in Chicago before pivoting to full-time analysis of AI tools and infrastructure.