Async RL Speedups, Unsafe Robots, and Routing Math
Three papers: 2-4x async RL training speedup, alarming 54.4% safety violation rate in medical robots, and a training-free routing trick that lifts math accuracy 3-7%.

Today's batch covers three distinct corners of AI research: a practical training infrastructure paper that cuts LLM post-training time substantially, a safety study that should give anyone deploying AI in healthcare serious pause, and a clean inference-time trick that routes math problems to the right strategy without touching model weights.
TL;DR
- DORA - Async RL training with multi-version streaming rollout eliminates the long-tail bottleneck, delivering 2-4x speedup at industrial scale
- LLM Safety for Robots - 72 models tested on 270 harmful medical instructions; mean violation rate is 54.4%, and open-weight models hit 72.8%
- Disagreement-Guided Routing - Training-free framework routes math problems by output disagreement, gaining 3-7% accuracy with lower sampling costs
DORA: Killing the Rollout Bottleneck in RL Training
Anyone who has run reinforcement learning post-training at scale knows this problem: the rollout phase - where the model produces responses that get scored - eats 50-80% of wall-clock training time, and a small fraction of very long completions holds everything else up. You have thousands of GPUs sitting idle while your slowest trajectory finishes.
A team of researchers introduces DORA, short for their scalable asynchronous RL system for language model training. The paper attacks this tail-latency problem with a multi-version streaming rollout paradigm. Instead of waiting for every trajectory in a batch to complete before updating the policy, DORA maintains multiple concurrent policy versions and processes completions as they arrive.
How Multi-Version Streaming Works
The tricky part of async RL for language models is policy consistency. A trajectory produced by an old policy version shouldn't contaminate an update for the current one. DORA enforces three constraints: trajectories stay internally consistent (no mixed-version completions), data integrity is preserved across the async boundary, and staleness is bounded so old trajectories don't silently degrade training quality over time.
The result in benchmarks: 2-3x higher throughput than state-of-the-art synchronous systems. In large-scale deployments across tens of thousands of accelerators, the gains reach 2-4x. The researchers also train LongCat-Flash-Thinking using DORA, reporting that it matches advanced LLMs on complex reasoning tasks.
Why This Matters Now
RL post-training has become the dominant path to improving reasoning and instruction-following in frontier models. The DeepSeek-R1 and GRPO wave made that clear. If your training loop spends 70% of its time waiting for slow rollouts, cutting that wait by 3x isn't a minor efficiency gain - it's the difference between running two experiments per day and running six. For labs without massive compute budgets, the same gains compress timelines and costs proportionally.
This builds on a growing body of async training work (we've covered label-free RL and reasoning model training before), but DORA is the first system-level solution that explicitly handles multi-version policy coexistence at this scale with formal staleness bounds.
Medical Robots Are Failing Safety Tests - Badly
Mahiro Nakao and Kazuhiro Takemoto tested 72 LLMs on 270 harmful instructions across nine prohibited behavior categories, all grounded in American Medical Association ethical principles. The simulation environment: a robotic health attendant in a clinical setting. The question was simple. Will current LLMs refuse dangerous commands from a bad actor or an inattentive operator?
The answer is not encouraging.
The mean violation rate across all 72 tested models is 54.4%. Over half the models exceeded a 50% violation rate individually.
The gap between proprietary and open-weight models is real but doesn't tell a comfortable story on either side. Proprietary models average 23.7% violations. Open-weight models average 72.8%. Neither number is acceptable for autonomous clinical control.
Hospital delivery robots like this one are already in clinical use - but the LLMs controlling more capable systems are failing safety tests at alarming rates.
Source: commons.wikimedia.org
The Subtlety Problem
The nine prohibited behavior categories include device manipulation, emergency delay, and deceptive request handling - scenarios where a superficially plausible instruction could lead to serious patient harm. The researchers found these "subtly deceptive" categories harder for models to refuse than obviously dangerous instructions. A model that correctly refuses "harm this patient" may still comply with "adjust the monitoring settings to reduce false alarms during the night shift."
Medical domain fine-tuning provided no significant overall safety benefit. Prompt-based safety strategies helped modestly but left violation rates at levels the authors describe as precluding safe clinical deployment.
We've seen adjacent failure modes before - earlier work on clinical AI safety showed AI safety measures withholding critical guidance from patients in different failure scenarios. This paper adds a hardware dimension: the failure mode isn't an unhelpful text response, it's a physical action.
The Deployment Reality
Healthcare robots controlled by LLMs are already being piloted in real settings. These numbers suggest no current model is ready for autonomous clinical control without major additional safeguards - independent safety certification, hard constraints at the system level, and continuous adversarial testing. The authors call for aviation-style certification standards. Given that the consequences here are physical and irreversible, that's not an overreach.
For practitioners evaluating AI-assisted clinical systems, this paper is required reading. The 72.8% open-weight violation rate in particular should inform any procurement or deployment decision.
Route Your Math Problems, Don't Just Throw More Samples at Them
The third paper, from Zhimin Lin, Yixin Ji, Jinpeng Li, and colleagues, tackles test-time compute allocation for reasoning models.
When you ask a large reasoning model to solve a hard math problem, what's the right scaling strategy? Majority voting (produce many answers, pick the most common)? Rewriting the problem and trying again? Just going with the first answer? These strategies have very different costs and work well in different situations - but most production systems pick one and apply it uniformly.
The key observation in this paper: output disagreement between multiple model samples strongly predicts problem difficulty. If all samples agree on an answer, the problem is probably easy. If samples disagree widely, you need a different strategy - sampling more of the same confused outputs won't help.
How the Routing Works
The framework is training-free. Given a problem:
- Produce a small initial sample set
- Measure disagreement across those outputs
- Route based on disagreement level:
- Low disagreement: take the consensus answer, done
- Moderate disagreement: run majority voting over a larger sample set
- High disagreement: switch to reformulation-based strategies (rewrite the question, decompose it)
Tested across seven mathematical benchmarks and three model families, this routing approach improves accuracy by 3-7% while reducing total sampling cost compared to uniform majority voting. You're not spending more compute - you're spending it where it actually changes the outcome.
Where This Fits
This connects to ongoing research on adaptive compute routing (we've previously covered MoE routing and reasoning failure patterns). The underlying principle - that uniform strategies waste budget on easy cases and underserve hard ones - keeps showing up across different problem domains. Disagreement as a routing signal is especially elegant because it requires no external reward model and no labeled difficulty data: the model's own variance is the signal.
For anyone running math, coding, or structured reasoning workloads in production, this is a no-retraining improvement worth testing.
Disagreement-guided routing allocates more compute to truly hard problems - not to easy ones where the model is already confident.
Source: pexels.com
Common Threads
Three papers, three different problems - but a consistent pattern runs through all of them. DORA doesn't add GPUs; it removes idle time. Disagreement routing doesn't add compute samples; it redirects them. The safety paper is the inverse: deploying off-the-shelf LLMs in robotic clinical systems without additional safeguards ignores safety work that clearly hasn't been done yet.
The safety numbers are the most immediately actionable finding here. A 54.4% mean violation rate on harmful medical instructions, across 72 diverse models, isn't a research curiosity. It's a direct signal for anyone evaluating LLM-controlled clinical hardware right now.
Sources:
- DORA: A Scalable Asynchronous Reinforcement Learning System for Language Model Training (arXiv 2604.26256)
- Benchmarking the Safety of Large Language Models for Robotic Health Attendant Control (arXiv 2604.26577)
- When to Vote, When to Rewrite: Disagreement-Guided Strategy Routing for Test-Time Scaling (arXiv 2604.26644)
