
Transformers.js v4 Ships WebGPU Runtime for Browser ML
HuggingFace's Transformers.js v4 rewrites its WebGPU runtime in C++, supports 200+ architectures, and delivers up to 4x faster inference in browsers and server-side JS runtimes.

HuggingFace's Transformers.js v4 rewrites its WebGPU runtime in C++, supports 200+ architectures, and delivers up to 4x faster inference in browsers and server-side JS runtimes.

Meta releases SAM 3.1 with Object Multiplex, processing all tracked objects in one shared pass for 7x faster inference at 128 objects and improvements on 6 of 7 VOS benchmarks.

Cohere releases its first audio model - a 2B-parameter open-source ASR system beating Whisper Large v3 by 27% on the HuggingFace Open ASR Leaderboard.

A new USCC report finds Chinese open-source models now dominate US AI startup stacks, with Qwen surpassing Llama in global downloads and Chinese models taking 41% of all Hugging Face downloads.

Italian-Legal-BERT is a 110M-parameter domain-adapted BERT model for Italian legal NLP, trained on 3.7GB of court decisions from Italy's National Jurisprudential Archive.

Researchers at Scuola Superiore Sant'Anna in Pisa built Italian-Legal-BERT, a 110M-parameter model trained on 3.7GB of Italian court decisions that outperforms general Italian BERT on legal NLP tasks.

A Hugging Face survey of 16 open-source reinforcement learning libraries finds the entire ecosystem has converged on async disaggregated training to fix a single brutal bottleneck: GPU idle time during long rollouts.

Hugging Face ships its largest LeRobot update yet: Unitree G1 humanoid support, Pi0-FAST VLA, Real-Time Chunking, 10x faster image training, and PEFT/LoRA fine-tuning for large robot policies.

Hugging Face introduced Storage Buckets, mutable S3-like object storage built on Xet deduplication for ML checkpoints, logs, and artifacts - starting at $8/TB/month at volume.

Alpamayo 1, a 10-billion-parameter vision-language-action model that explains its own driving decisions, has become the most downloaded robotics model on Hugging Face - less than two months after its CES 2026 debut.

Georgi Gerganov's ggml.ai joins Hugging Face, bringing the most important local inference project under the $13.5 billion AI platform's umbrella.

Georgi Gerganov and the ggml.ai team behind llama.cpp are joining Hugging Face. The deal unifies model hosting, model definition, and local inference under one open-source roof.