Hardware

Hailo-10H - Edge AI With On-Device LLMs

Hailo-10H - Edge AI With On-Device LLMs

Complete specs, benchmarks, and analysis of the Hailo-10H - a 2.5W edge AI accelerator with 40 TOPS INT4, on-module LPDDR4, and the ability to run LLMs and VLMs on a Raspberry Pi at 10 tokens per second.

NVIDIA Rubin CPX - Inference GPU With GDDR7

NVIDIA Rubin CPX - Inference GPU With GDDR7

Full specs, benchmarks, and analysis of the NVIDIA Rubin CPX - a purpose-built inference GPU with 128GB GDDR7, 30 PFLOPS NVFP4, and 3x faster attention versus Blackwell, targeting million-token context workloads.

Tenstorrent Blackhole p150a - RISC-V AI Card

Tenstorrent Blackhole p150a - RISC-V AI Card

Complete specs, benchmarks, and analysis of the Tenstorrent Blackhole p150a - the $1,399 PCIe AI accelerator with 120 Tensix cores, 768 RISC-V processors, 32GB GDDR6, and fully open-source software.

AMD Instinct MI300X

AMD Instinct MI300X

AMD Instinct MI300X specs, benchmarks, and real-world performance data. 192GB HBM3, 5,300 GB/s bandwidth, 2,610 TFLOPS FP8 on CDNA 3 chiplet architecture.

AMD Instinct MI350X

AMD Instinct MI350X

AMD Instinct MI350X specs and performance estimates. 288GB HBM3e, ~6,000 GB/s bandwidth, ~3,600 TFLOPS FP8 on CDNA 4 architecture at TSMC 3nm.

AWS Trainium2 - Amazon's Cloud Training Chip

AWS Trainium2 - Amazon's Cloud Training Chip

AWS Trainium2 is Amazon's second-generation custom AI training chip, powering EC2 Trn2 instances with 96GB HBM2e per chip and tight integration with the AWS Neuron SDK and SageMaker ecosystem.