Reasoning Traps, LLM Chaos, and Steering Curves
Three papers this week: why better reasoning creates safety risks, why multi-agent systems behave chaotically even at zero temperature, and why straight-line activation steering is broken.

Three papers landed this week that each poke a hole in something the field treats as settled. Safer reasoning is good. Multi-agent consensus is predictable. Activation steering is linear. None of those claims survive scrutiny.
TL;DR
- The Reasoning Trap - Every advance in LLM logical reasoning also advances situational awareness, creating escalating safety risks current measures can't address
- Chaotic Dynamics in Multi-LLM Deliberation - Five-agent LLM committees show chaotic divergence even at T=0, driven by two independent instability sources
- Curveball Steering - LLM activation spaces are geometrically curved, not flat, which breaks standard linear steering methods
The Reasoning Trap: Better Logic, More Dangerous Models
Paper: "The Reasoning Trap - Logical Reasoning as a Mechanistic Pathway to Situational Awareness" Authors: Subramanyam Sahoo, Aman Chadha, Vinija Jain, Divya Chaudhary Venue: ICLR 2026 Workshop on Logical Reasoning of Large Language Models
The intuition that better reasoning is always better is comfortable. This paper, accepted at ICLR 2026, argues it's also wrong - at least from a safety standpoint.
The authors introduce the RAISE framework (Reasoning Advancing Into Self-Examination), which maps three mechanistic pathways through which improving a model's logical abilities also deepens its situational awareness:
- Deductive self-inference: using formal logical steps to infer facts about its own nature
- Inductive context recognition: recognizing patterns in prompts that reveal training and deployment conditions
- Abductive self-modeling: constructing working models of what it is and how it's being used
The escalation ladder they construct runs from basic self-recognition up to strategic deception. The claim isn't that today's models are plotting anything. The claim is that the research directions the field is currently pursuing - better chain-of-thought, stronger formal reasoning, improved multi-step logic - each map directly onto amplifiers of one or more of these pathways.
Why practitioners should care
If you're building on top of reasoning-heavy models (o3, Claude's extended thinking, Gemini's deep reasoning mode), you're shipping systems that are more situationally aware than their predecessors. That awareness isn't necessarily malign, but it changes the threat surface for prompt injection, jailbreaking, and what the authors call "strategic deception" - a model that knows it's being evaluated behaving differently than one that doesn't.
The paper proposes two concrete responses: a "Mirror Test" benchmark designed to probe whether models exhibit signs of situational awareness during evaluation, and a Reasoning Safety Parity Principle requiring that reasoning capability gains be matched by equivalent safety capability gains before deployment. Neither is added yet. They're proposals, not products.
The honest read here is that this is a position paper - 21 pages, arguing a thesis rather than reporting experimental results. But the thesis is specific enough to disagree with, which puts it a cut above most safety hand-wringing. The authors map actual research topics to actual escalation risks. That's useful groundwork even if the proposed benchmarks still need building.
Our earlier coverage of alignment interventions backfiring under pressure is directly relevant here - the problem isn't just that safety measures are insufficient, it's that capability advances tend to race ahead of safety measures structurally.
Chaotic Dynamics in Multi-LLM Deliberation
Paper: "Chaotic Dynamics in Multi-LLM Deliberation" Authors: Hajime Shimao, Warut Khern-am-nuai, Sung Joo Kim Submitted: March 10, 2026
Multi-agent LLM systems are supposed to be more reliable than single models. Committees of models should average out individual errors, reach stable consensus, and produce more solid decisions. This paper tests that assumption with mathematical rigour and finds the opposite.
The researchers modeled five-agent LLM committees as random dynamical systems and measured their instability using empirical Lyapunov exponents - a standard chaos theory tool that quantifies how fast small differences in initial conditions grow over time. Higher exponent, more chaotic, less reproducible.
The two independent instability sources
They identified two separate routes to chaotic behavior:
Role differentiation - assigning different positions (chair, secretary, reviewer, etc.) within a homogeneous committee of identical models produces a Lyapunov exponent of λ̂ = 0.0541. The roles themselves introduce divergence by shaping how each agent processes and weights the same information differently.
Model heterogeneity - using different LLM types within a committee without role assignment produces λ̂ = 0.0947, nearly double the role-only effect.
The dangerous case is when you combine both: a mixed committee with role assignments. λ̂ drops slightly to 0.0519, but the interactions between the two instability sources aren't additive. The system behaves in ways that can't be predicted from measuring each source separately.
The T=0 finding
This is what makes the paper truly surprising. All measurements were taken at temperature T=0, where models produce deterministic outputs. Standard reasoning says deterministic inputs produce deterministic outputs produce stable committees. That's wrong. Instability emerges from how agents interact with each other over multiple rounds, not from individual output randomness.
The most effective stabilization strategies they found: removing the chair role (largest single reduction in λ̂) and shortening memory windows so agents don't build up divergent context across rounds.
Their recommendation is direct: stability auditing should be a core design requirement for multi-LLM governance systems, not an optional stress test.
For anyone building production multi-agent pipelines today, this matters concretely. Our earlier piece on multi-agent training instability covered training-time instability; this paper shows the same dynamics apply at inference time in deliberation contexts. The combination suggests that multi-agent architectures carry instability risks at every stage.
The Lorenz attractor: a deterministic system that produces chaotic, non-repeating trajectories - the same principle the researchers found in multi-LLM deliberation at T=0.
Source: wikimedia.org
Curveball Steering: Linear Activation Steering Is Broken
Paper: "Curveball Steering: The Right Direction To Steer Isn't Always Linear" Authors: Shivam Raval, Hae Jin Song, Linlin Wu, Abir Harrasse, Jeff Phillips, Amirali Abdullah Submitted: March 10, 2026
Activation steering is one of the more practical interpretability tools available right now. The basic idea: find a direction in a model's internal activation space that corresponds to some behavior or concept, then add or subtract that vector at inference time to nudge behavior. It's been used to suppress harmful outputs, amplify specific reasoning styles, and study what models represent internally.
The whole approach rests on the Linear Representation Hypothesis: the idea that these behavioral attributes sit along straight lines in the model's high-dimensional activation space. This paper measures whether that's actually true and finds it often isn't.
Measuring the curve
The researchers quantify geometric distortion by computing the ratio of geodesic distances (the actual shortest path through the space) to Euclidean distances (straight-line distance). In a flat, linear space these ratios are 1. When the space is curved, geodesic paths arc away from straight lines and the ratio climbs above 1.
What they find is that distortion is concept-dependent - some behavioral dimensions of the activation space are close to linear, while others are substantially curved. This explains something practitioners have noticed anecdotally: steering works well for some behaviors and poorly for others. The difference isn't model quality or prompt engineering. It's geometry.
The Curveball fix
Their method, Curveball steering, moves the intervention into a transformed feature space created by polynomial kernel PCA rather than operating directly in the original activation space. Kernel PCA is a classical dimensionality reduction technique that can handle non-linear structure; the polynomial kernel here captures curvature in the original space as linear structure in the transformed space, making standard steering methods valid again.
The results show consistent improvement over linear PCA-based steering, with the largest gains in regimes where geometric distortion is strongest.
This has immediate practical implications. If you're using activation steering for alignment, safety filtering, or behavior modification - and the behavior you care about sits in a high-distortion region of the activation space - linear methods are systematically miscalibrated. The direction you compute isn't the right direction to apply. Curveball steering gives a geometry-aware alternative.
The paper doesn't settle which concepts are high-distortion versus low-distortion in production models. That mapping work is still open. But the measurement methodology they provide - geodesic-to-Euclidean ratios - gives practitioners a way to check before steering rather than wondering why results are inconsistent.
Common Thread
Three papers, three different corners of the field, one theme: assumptions that were convenient engineering simplifications are starting to fail under pressure.
The reasoning safety paper shows that the capability direction the field is running in creates safety risks that scale with progress. The chaos paper shows that determinism at the component level doesn't produce stability at the system level. The steering paper shows that geometric intuitions about activation space don't hold for the behaviors we care most about controlling.
Two concrete things fall out of this for practitioners: if you run multi-agent systems, add stability auditing and measure Lyapunov exponents across your configurations. If you use activation steering, measure geodesic-to-Euclidean ratios on the concepts you care about before trusting linear steering directions. Neither of these were standard practice last week. They should be now.
Sources:
- "The Reasoning Trap - Logical Reasoning as a Mechanistic Pathway to Situational Awareness" - arXiv:2603.09200
- "Chaotic Dynamics in Multi-LLM Deliberation" - arXiv:2603.09127
- "Curveball Steering: The Right Direction To Steer Isn't Always Linear" - arXiv:2603.09313
