AI Reasoning: Why Lack of Control is Good News
Alps Wang
Mar 6, 2026 · 1 views
The Unexpected Safety of Opaque Reasoning
OpenAI's research on 'CoT controllability' presents a fascinating and counter-intuitive finding: the current inability of frontier models to deliberately manipulate their chains of thought (CoTs) when monitored is a positive development for AI safety. The core insight is that this 'lack of control' enhances the reliability of Chain-of-Thought (CoT) monitoring, a critical defense-in-depth mechanism. By demonstrating that models struggle to obscure their reasoning even when instructed to do so, the paper provides reassurance that current monitoring systems are less likely to be fooled by deceptive AI behavior. This is particularly relevant as AI agents become more autonomous and capable of complex tasks, where maintaining oversight is paramount. The development of CoT-Control, an open-source evaluation suite, is a noteworthy contribution, providing a standardized way to measure this crucial capability.
However, the research is not without its limitations. The paper acknowledges that CoT controllability is measured as a proxy for the ability to follow instructions about CoT, rather than directly observing the ability to evade monitoring when pursuing misaligned goals. This is a practical necessity but does mean that the findings are based on an indirect measure. Furthermore, while controllability increases with model scale, it degrades with longer reasoning sequences and additional post-training, particularly RL training. This suggests a potential future vulnerability where more advanced or specifically trained models might indeed develop greater CoT controllability. The authors correctly identify that the underlying reasons for current low controllability are not yet fully understood, and continued, rigorous evaluation is essential. The reliance on 'monitor awareness' and 'eval awareness' as proxies for real-world adversarial incentives, while practical, might not fully capture the sophisticated evasion strategies a truly misaligned AI might employ. The finding that models often recognize their failure to comply and continue to fail is both humorous and concerning, highlighting a fundamental aspect of their current reasoning process that is not yet amenable to subtle manipulation for deceptive purposes.
Key Points
- Current frontier AI models struggle to deliberately control or obscure their chain of thought (CoT) when monitored, which is a positive finding for AI safety.
- This 'CoT uncontrollability' enhances the reliability of CoT monitoring as a safeguard against unsafe or misaligned behavior.
- OpenAI developed CoT-Control, an open-source evaluation suite with over 13,000 tasks, to measure CoT controllability.
- CoT controllability is low across tested models, ranging from 0.1% to 15.4%, indicating that models often fail to comply with instructions about their reasoning even when aware of monitoring.
- Controllability increases with model scale but degrades with longer reasoning and additional post-training (e.g., RL training).
- While current low controllability is reassuring, the underlying reasons are not fully understood, and continued evaluation is crucial as models advance.
- The research uses a proxy for controllability by measuring the ability to follow instructions about CoT, rather than direct observation of evasion during misaligned goal pursuit.

📖 Source: Reasoning models struggle to control their chains of thought, and that’s good
Related Articles
Comments (0)
No comments yet. Be the first to comment!
