Anthropic: Better AI Output Means Worse Oversight
Anthropic's AI Fluency Index reveals that when Claude produces polished code and documents, users question its reasoning 5.6 times less often.

The better AI gets at its job, the worse humans get at checking its work.
That's the central finding of Anthropic's AI Fluency Index, a research report published on February 23 that analyzed 9,830 anonymized Claude conversations from January 2026. The study tracked 11 observable user behaviors - things like questioning reasoning, fact-checking, and identifying missing context - and found a pattern that should worry anyone building with or rolling out AI systems.
When Claude produces polished artifacts like code, documents, or interactive tools, users become markedly more directive in their initial instructions but dramatically less critical in evaluating the results. They specify exactly what they want, then stop checking whether they got it right.
TL;DR
- Anthropic studied 9,830 Claude conversations and identified 11 observable "fluency" behaviors
- Users who receive polished artifacts (code, documents) question AI reasoning 5.6x less often
- 85.7% of users iterate on responses, but only 30% explicitly instruct Claude on how to collaborate
- Fact-checking drops 3.7 percentage points when the output looks professional
The Numbers Behind the Paradox
The study introduces what the researchers call the "Artifact Effect." Artifacts - code blocks, formatted documents, interactive tools - appeared in 12.3% of conversations analyzed. In those sessions, users shifted their behavior in a specific and troubling way.
| Behavior | With Artifacts | Without Artifacts | Change |
|---|---|---|---|
| Clarifying goals | Higher | Lower | +14.7pp |
| Specifying format | Higher | Lower | +14.5pp |
| Providing examples | Higher | Lower | +13.4pp |
| Iteration | Higher | Lower | +9.7pp |
| Identifying missing context | Lower | Higher | -5.2pp |
| Fact-checking | Lower | Higher | -3.7pp |
| Questioning reasoning | Lower | Higher | -3.1pp |
The top four rows show users getting better at asking for what they want. The bottom three show them getting worse at verifying they received it.
Anthropic's study tracked 11 observable user behaviors across nearly 10,000 Claude conversations to quantify the scrutiny gap.
The Iteration Signal
Iteration turned out to be the single strongest predictor of AI fluency across the board. Of the 9,830 conversations studied, 85.7% showed iterative behavior - users building on previous exchanges rather than accepting the first response and moving on. Conversations with iteration averaged 2.67 additional fluency behaviors, roughly double the 1.33 average for non-iterative sessions.
Users who iterated were 5.6 times more likely to question Claude's reasoning and four times more likely to catch missing context. The problem is that iteration tends to increase in artifact conversations (users polish their prompts), but critical evaluation simultaneously drops.
The 30% Gap
Only 30% of conversations included explicit instructions about how the user wanted Claude to behave - prompts like "push back if my assumptions are wrong" or "tell me what you're uncertain about." The research suggests that this kind of prompt engineering pays dividends for output quality, yet most users skip it entirely.
Inside the 4D-AI Fluency Framework
Anthropic built the study on a framework developed in collaboration with Professors Rick Dakan and Joseph Feller that identifies 24 competency behaviors across four dimensions. Only 11 of those behaviors are directly observable in Claude conversations. The remaining 13 - including being honest about AI's role in work and considering consequences of sharing AI-produced output - happen outside the chat window and couldn't be measured.
"As AI models become increasingly capable of producing polished-looking outputs, the ability to critically evaluate those outputs will become more valuable rather than less."
The researchers used Claude Sonnet 4 for classification and Claude Haiku 3.5 for language detection, applying 11 separate binary classifiers to each conversation. Results remained consistent across seven days of the week (most behaviors varied only 1 to 5 percentage points) and across six languages: English, French, Spanish, Chinese, Japanese, and German. Saturday showed slightly lower iteration rates (81.4% versus the weekday peak of 87.9%), but the overall patterns held.
Why It Matters for Code
When AI-generated code compiles cleanly, developers are measurably less likely to review it for logical errors.
The effects are sharpest for code generation. When Claude produces a clean, syntactically correct function, users are measurably less likely to verify that the function does what they intended. This connects directly to the security findings in recent audits of AI coding tools that found dozens of vulnerabilities in code that developers launched without scrutiny.
A separate Anthropic study published earlier this year found that Claude Code sessions have roughly doubled in autonomy as users build trust through repeated positive interactions. Taken together, the picture is clear: users are handing over more control to AI while simultaneously checking its work less.
The Extended Conversation Problem
The report also flags a technical nuance often overlooked. Extended conversations degrade AI output quality as irrelevant context builds up in the chat window. This means the very sessions where users iterate most - and theoretically engage most fluently - may also be the ones where Claude performs worst. Fresh conversations, the researchers suggest, may produce better results than long, detailed threads.
What It Does Not Tell You
The study has significant limitations that the researchers acknowledge.
The sample is biased toward early adopters who are already comfortable with AI. The behaviors of a general population - many of whom still struggle with basic prompt construction - would likely look very different. This is a snapshot of the best-case scenario for AI literacy, and it is still showing a scrutiny gap.
The analysis is also platform-specific. Users behave differently on Claude.ai than they might on ChatGPT, Gemini, or open-source models running locally. The Artifact Effect might be stronger or weaker depending on how different interfaces present code and documents.
Most importantly, the 13 unobservable behaviors - the ones involving ethical judgment, transparency about AI use, and consideration of downstream consequences - are what the researchers themselves describe as "arguably some of the most consequential dimensions of AI fluency." The study measures how people talk to AI. It can't measure what they do with the output.
And it is purely correlational. The report identifies patterns but can't prove that polished outputs cause reduced scrutiny. It's possible that users who are less inclined to verify simply gravitate toward artifact-producing tasks in the first place.
The Artifact Effect is not a bug in AI design. It's a feature of human psychology: we trust things that look finished. The better AI systems get at producing clean code, formatted reports, and polished documents, the harder humans will need to work to maintain the habit of questioning them. Anthropic's data suggests that most users are not doing that work yet - and as the broader AI safety conversation shifts from hypothetical risks to everyday workflow, the scrutiny gap may matter more than any benchmark score.
Sources:
- Anthropic Education Report: The AI Fluency Index
- The Decoder: Anthropic's AI Fluency Index finds polished AI output makes users less likely to check for errors
- Blockchain News: Anthropic Study Reveals Users Skip Critical Checks on AI-Generated Code
- Futurism: Anthropic Introduces an AI Fluency Index to Measure How Well People Use AI
- PYMNTS: Classrooms Turn AI Into a Lesson in Collaboration
