
Today in AI Research: Stable Agent Training, Compound AI Limits, and the Algorithm Trust Paradox
New papers tackle training collapse in agentic RL with a unified stabilization recipe, reveal when querying multiple models actually helps, and expose a paradox where LLMs claim to trust humans but bet on algorithms.