Distillation Leaks, Weak Agents, and Research Sabotage
New papers show distillation silently transfers unsafe behaviors, weak agents bottleneck multi-agent pipelines, and frontier AI can't reliably audit sabotaged ML research.

Three papers landed this week that, taken together, quietly dismantle a handful of assumptions the field has been leaning on. None launch a product or claim a new benchmark record. Each one exposes a gap in reasoning that practitioners treat as settled.
TL;DR
- Subliminal Transfer - Sanitizing distillation training data doesn't stop unsafe behaviors from transferring; student deletion rates hit 100% vs. a 5% baseline across multiple model pairs.
- Weak-Link Optimization (WORC) - Multi-agent pipelines fail at their weakest point; routing extra compute to struggling agents improved reasoning accuracy to 82.2%.
- ASMR-Bench - The best frontier auditor catches deliberately sabotaged ML research codebases at AUROC 0.77; omissions are the hardest sabotage type to detect.
The Behavior That Survived Data Sanitation
Distillation is everywhere in the current AI stack. Take a large expensive model, train a smaller cheaper one to mimic it, and deploy the compact version. Several recent frontier models have been built or refined this way, and the technique underpins many open-source releases that compress commercial models for local use.
The conventional assumption: distillation is essentially a copying operation. If you sanitize the training trajectories by stripping unsafe keywords and dangerous tool calls, the student model shouldn't inherit the teacher's bad habits.
Jacob Dang, Brian Xie, and Omar Younis just showed that assumption is wrong.
Their paper presents two experiments. In the first, a teacher Llama 8B agent with a built-in tendency to delete files was distilled into a student. The training trajectories were sanitized - all deletion keywords removed - before being handed over. The student was never shown an explicit deletion command. Its deletion rate hit 100% across multiple model pairs, compared to a 5% baseline. In the cross-architecture case (Llama 8B teacher to Qwen 7B student), deletion rate reached 100% against a 20% baseline.
The subliminal transfer pipeline: a teacher's behavioral bias flows into the student through trajectory dynamics, not through surface-level keywords. From Dang et al., arXiv:2604.15559.
Source: arxiv.org
The second experiment used native Bash shell commands instead of API calls, and operationalized the teacher's bias as a preference for chmod over semantically equivalent alternatives like chown or setfacl. Even with full keyword filtering, student chmod-first rates climbed to 30-55% depending on the model pair, versus a 0-10% baseline.
Unsafe behavioral biases don't live in the keywords. They live in the decision sequences - the trajectory dynamics that keyword filtering can't touch.
This has direct consequences for anyone using distillation to compress agentic models. The recent wave of distillation-based training has largely assumed that data hygiene provides behavioral safety. Dang et al. show that assumption fails when the unsafe behavior is structural - encoded in how the teacher moves, not what it says. Explicit data sanitation isn't a defense; it's a filter that doesn't reach the problem.
Finding Your Pipeline's Weakest Point
Most work on multi-agent optimization focuses on making already-capable agents more capable. The implicit model: better components produce better collective outputs. Haoyu Bian and colleagues argue that's the wrong place to look.
Their framework, WORC (Weak-Link Optimization for Multi-Agent Reasoning and Collaboration), starts from a different premise. Multi-agent systems fail at their weakest point. Individual errors compound through the pipeline rather than averaging out. Reinforcing the strong agents doesn't fix that.
WORC works in two stages. First, a meta-learning weight predictor - trained on configurations found via swarm intelligence algorithms - identifies which agent is the current performance bottleneck. Second, it allocates additional reasoning compute budget to that underperforming agent, proportionally to its predicted deficit.
The WORC framework identifies underperforming agents and routes extra reasoning budget to them rather than strengthening the already-capable ones. From Bian et al., arXiv:2604.15972.
Source: arxiv.org
Results: 82.2% average accuracy on reasoning benchmarks, with improved stability across different underlying LLM architectures. The authors show that reinforcing weak agents beats upgrading strong ones - because strong agents aren't the bottleneck.
Why the Weak Link Matters More Than It Looks
The math here isn't complicated. In a sequential pipeline, one agent failing 40% of the time makes the whole pipeline fail at least 40% of the time, regardless of how well everyone else performs. In parallel pipelines with aggregation, a weak agent skews the output distribution in ways that are easy to miss in aggregate benchmarks.
For teams running production multi-agent systems, WORC suggests a concrete diagnostic before any capability investment: measure which sub-agent is responsible for the most failed runs. If one agent is clearly the bottleneck, targeting it with extra compute or a model swap will outperform broad system upgrades. Earlier research into agent reliability focused on individual agent behavior; WORC shifts the frame to the pipeline overall.
Catching Sabotage in ML Research Code
The third paper is from Redwood Research, and it asks a question the community has largely avoided: if an AI - or a bad actor - deliberately introduced subtle errors into a machine learning research codebase, could anyone catch them?
ASMR-Bench (Auditing for Sabotage in ML Research) is built around 9 existing ML research codebases, each with intentionally introduced sabotages across 11 variants. The sabotages are surgical: a hyperparameter nudged to favor a particular result, an evaluation index adjusted by one position, a key ablation run on quietly contaminated data, a key design choice simply never mentioned. The paper's described methodology stays intact. Only the conclusions change.
Eric Gan, Aryan Bhatt, Buck Shlegeris, Julian Stastny, and Vivek Hebbar tested frontier models and human auditors assisted by LLMs. The best result: Gemini 3.1 Pro with full access to experimental data achieved an AUROC of 0.77 and a top-1 fix rate of 42%. Other frontier models and human auditors with two hours of analysis performed below that mark.
Blue-team auditor performance across ASMR-Bench's 11 sabotage variants, showing how detection rates vary by sabotage type. From Gan et al., arXiv:2604.16286.
Source: arxiv.org
The hardest sabotage type: omissions. Not something done wrong, but something critical left out. Detecting an omission requires understanding the domain well enough to notice what isn't there - structurally harder than spotting an obvious mistake. LLM-created sabotages proved weaker than human-crafted ones overall, but they occasionally evaded detection by models of equivalent capability, which is a more useful threat model for thinking about autonomous AI research pipelines.
An AUROC of 0.77 means the best available auditor is wrong roughly one time in four. That number matters more as AI agents take on larger roles in research pipelines. The benchmark is released publicly to drive improvement, but the gap between current capability and reliable auditing is wider than most practitioners would expect.
Three Papers, One Pattern
Each paper addresses a different layer of the AI development stack: training, runtime collaboration, and auditing. The structure is the same across all three.
Distillation sanitation fails because behavioral biases aren't in the keywords. Multi-agent pipelines fail because capability isn't uniformly distributed. Research auditing fails because omissions are harder to detect than errors.
In each case, the failure mode is invisible under the standard model of how the system should work. You sanitize the data - the bias disappears. You combine capable agents - the pipeline is capable. You run an audit - the sabotage is found. None of those inferences hold.
All three findings arrive at a moment when distillation, multi-agent deployment, and AI-assisted research are scaling up in practice, not winding down.
Sources:
