A growing argument in AI governance circles suggests that because large language models cannot sustain long-term, stable reasoning, they are therefore less dangerous than feared. If models lose coherence, forget constraints, or drift mid-conversation, the logic goes, their ability to cause serious harm must be limited.
This reasoning is problematic and deeply mistaken. The inability of current models to sustain thinking over time does not make them safer, it makes them harder to govern.
The error we already understand
In every high-risk domain, safety mechanisms compensate for known fragility. Aircraft are built with redundant systems because engines fail. Medical devices require monitoring because biological systems fluctuate. Financial markets impose circuit breakers because rationality collapses under stress. None of these systems are made safer by fragility. They are made safer by acknowledging it.
AI governance has inverted this logic. Instead of asking “how do we constrain systems that degrade invisibly,” the conversation has drifted toward “perhaps degradation itself is a safeguard.” It is not.
The Opus 4.6 moment
Recent debate around Anthropic’s internal evaluation of its Opus 4.6 model exposed this confusion clearly. The company faced a specific problem: traditional benchmarks had saturated, meaning standardized tests could no longer meaningfully distinguish between capability levels. To assess whether the model had crossed a critical autonomy threshold (the point at which a system might pursue goals independently enough to resist correction) Anthropic reportedly relied on an internal employee survey rather than external, reproducible evaluation.
Critics rightly focused on process: internal judgment, groupthink risk, insufficient independent validation. But beneath that procedural critique lies a more fundamental conceptual error, one that affects the entire safety discourse.
We are treating cognitive instability as if it were a mitigating factor. In practice, it is a risk multiplier.
Why instability increases danger
Models do not need to reason continuously to cause harm. They need only to produce confident outputs at the wrong moment, in the wrong context, without signaling their own degradation. A system that reasons flawlessly for two hours and collapses invisibly in minute 121 is not safer than one that reasons imperfectly throughout. It is more dangerous, because users cannot detect the transition.
This is a core pattern documented consistently across frontier models in the emerging field of AI Conversational Phenomenology: systems fail silently while remaining fluent. As detailed in AI’s Verification Crisis, coherence collapse occurs predictably at a fraction of advertised context lengths. Instruction loss, contradiction drift, hallucinated scaffolding, and positional unreliability appear not as crashes but as smooth, authoritative continuation. To the user, nothing appears wrong, until consequences surface.
A model that cannot sustain long-term reasoning still makes medical suggestions, drafts legal arguments, advises on financial decisions, and engages in crisis conversations. The harm does not arise from autonomy in the abstract; it arises from misplaced trust. Users do not interact with models as probabilistic text generators. They interact with them as reasoning agents. When that perceived reasoning degrades without warning, guardrails become more important, not less.
The benchmark saturation trap
This is why benchmark saturation is such a dangerous moment. When traditional tests can no longer distinguish capability levels, governance quietly shifts from measurement to judgment. Internal consensus replaces external validation. “It seems fine” becomes the decision criterion. But as systems approach consequential thresholds, rigor should increase—not evaporate.
The paradox is this: the closer we get to systems whose behavior we do not fully understand, the more we rely on informal assessment. The direction should be the opposite.
If a model truly posed no meaningful risk, external evaluation would be easy. It is precisely because risk is ambiguous, contextual, and emergent that guardrails are required. And guardrails cannot be based on belief. They must be based on observable behavior under real conditions.
Confidence without calibration
This is where the current safety conversation breaks down. We conflate autonomy with persistence. We assume danger requires sustained goal-directed behavior. In reality, most real-world harm occurs through intermittent failure: a wrong dosage, a fabricated citation, a missed crisis signal, an authoritative but incorrect explanation delivered at the wrong moment.
None of these require long-term thinking. They require only confidence without calibration.
The lesson from Opus 4.6 is not that the model is unsafe. It is that our evaluation frameworks are misaligned with actual risk. We are still measuring abstract capability while harm emerges phenomenologically, through how systems behave across time, context, and user trust.
The evidence is quantifiable
These failures are not anecdotal. Evans’ Law, a scaling framework validated across eight frontier models including GPT-5.0/5.1, Claude 4.5, Gemini, Grok 4.1, DeepSeek, Qwen, and Mixtral, quantifies the gap between advertised and actual coherent context. The core finding is a sublinear relationship between model scale and usable reasoning depth: text-only coherence follows L ≈ 1969.8 × M^0.74, while multimodal coherence drops further to L ≈ 582.5 × M^0.64. In practice, this means usable context windows are a fraction of what vendors claim, and the gap widens as models grow larger, a phenomenon also recently documented by research out of Anthropic.
The failures preceding collapse are not random. Evans’ Law identifies a taxonomy of drift signatures, compression drift, expansion drift, logic-deferral drift, and code-layer leakage, that appear as vendor-specific patterns before full coherence breakdown. The most dangerous failure mode, observed in GPT-5.0, is what the framework terms “opaque coherence”: outputs that remain fluent and structurally intact while becoming semantically incorrect. This is precisely the kind of silent degradation that makes guardrails essential.
The problem compounds when memory enters the picture. In memory-enabled systems using retrieval-augmented generation, a separate failure mechanism emerges: memory leakage. As documented in When Memory Leaks: RAG-Induced Ambiguity and Fracture-Repair Hallucination, cross-conversation user history retrieved without provenance governance creates authority conflicts between source material and retrieved memory. Alignment incentives then force models to synthesize conflicting inputs rather than acknowledge insufficiency, producing hallucinations that are maximally persuasive because they deploy the user’s own conceptual frameworks in inappropriate contexts. Source-grounding and improved retrieval quality do not mitigate this failure, they often intensify it by increasing confidence in continuation while leaving the underlying authority conflict unresolved.
Together, these findings demonstrate that degradation is not speculative. It is measurable, predictable, and architecturally rooted. Testing of newer models and platforms is ongoing, and to date there is no indication that these fundamental dynamics have changed.
What guardrails actually look like
Guardrails are not about limiting intelligence. They are about constraining interaction. Session limits, degradation warnings, disclosure of reliability envelopes, adverse event tracking do not slow innovation. They just make failure more legible.
The most dangerous AI systems are not the ones that think forever. They are the ones that think just well enough, for just long enough, to be trusted, and then fail quietly. Cognitive fragility is the reason safety infrastructure is non-negotiable.





