
Cambricon MLU590 - China's Inference Accelerator
Full specs and analysis of the Cambricon MLU590 - 192GB HBM2e, ~2,400 GB/s bandwidth, TSMC 7nm, and what it means for AI inference outside the NVIDIA ecosystem.

Full specs and analysis of the Cambricon MLU590 - 192GB HBM2e, ~2,400 GB/s bandwidth, TSMC 7nm, and what it means for AI inference outside the NVIDIA ecosystem.

Huawei Ascend 910B specs, benchmarks, and real-world performance. 64GB HBM2e, ~1,200 GB/s bandwidth, ~600 TFLOPS FP16 - the chip that trained DeepSeek.

Huawei Ascend 910C specs, benchmarks, and performance analysis. 96GB HBM2e, ~1,800 GB/s bandwidth, ~800 TFLOPS FP16 - China's flagship AI chip under US sanctions.

DeepSeek has denied Nvidia and AMD pre-release access to its upcoming V4 model while granting Huawei and domestic Chinese chipmakers a multi-week optimization window, signaling a strategic pivot toward building a parallel AI software ecosystem on Chinese silicon.

Baidu's third consecutive quarter of revenue decline wiped $11 billion off its market cap, exposing the gap between its AI ambitions and the advertising business still funding them.

Comparing two Chinese AI models with MIT-family licenses - Moonshot AI's trillion-parameter Kimi K2.5 against Zhipu AI's ultra-efficient GLM-4.7-Flash that punches well above its weight on coding and agentic tasks.

A senior Trump administration official confirms DeepSeek trained its upcoming AI model on Nvidia's most advanced Blackwell chips at an Inner Mongolia data center, despite US export controls banning the hardware from reaching China.

Zhipu's GLM-4.7-Flash is a 30B-A3B MoE model that posts 59.2% on SWE-bench Verified and 79.5% on tau2-Bench while running on a single RTX 4090 - MIT licensed and free via the Z.AI API.

MiniMax M2.5 is a 230B MoE model (10B active) that scores 80.2% on SWE-Bench Verified while costing 1/10th to 1/20th of frontier competitors like Claude Opus 4.6 and GPT-5.2.

Claude Sonnet 4.6 identifies itself as DeepSeek when prompted in Chinese, just one day after Anthropic accused DeepSeek of industrial-scale distillation attacks. The cause is training data contamination, not an identity crisis - but the timing is spectacular.

Anthropic accuses three Chinese AI labs of industrial-scale distillation attacks using 24,000 fraudulent accounts and 16 million exchanges with Claude. MiniMax ran the largest operation at 13 million exchanges. None of the three companies have responded.

Alibaba releases Qwen 3.5, a 397-billion-parameter open-weight model that claims to outperform US frontier models at a fraction of the cost.