Megan Crouse
2025-07-21 17:53:00
www.techrepublic.com
Monitoring generative AI’s decision-making is critical for safety, but the inner workings that lead to text or image outputs remain largely opaque. A position paper released on July 15 proposes chain-of-thought (CoT) monitorability as a way to watch over the models.
The paper was co-authored by researchers from Anthropic, OpenAI, Google DeepMind, the Center for AI Safety, and other institutions. It was endorsed by high-profile AI experts, including former OpenAI chief scientist and Safe Superintelligence co-founder Ilya Sutskever, Anthropic researcher Samuel R. Bowman, Thinking Machines chief scientist John Schulman, and deep learning luminary Geoffrey Hinton.
What is chain-of-thought?
Chain-of-thought refers to the intermediate reasoning steps a generative AI model verbalizes as it works toward an output. Some deep research models produce reports to their users of what they are doing at the time. Assessing what the models are doing before they produce human-readable data is known as interpretability, a field Anthropic has heavily researched.
However, as AI becomes more advanced, monitoring the “black box” of decision-making becomes increasingly difficult. Whether chain-of-thought interpretability will work in a few years is anyone’s guess, but for now, the researchers are pursuing it with some urgency.
“Because CoT monitorability may be fragile, we recommend that frontier model developers consider the impact of development decisions on CoT monitorability,” the researchers wrote.
Chain-of-thought oversight could check ‘misbehavior’ of advanced AI models
“AI systems that ‘think’ in human language offer a unique opportunity for AI safety: we can monitor their chains of thought (CoT) for the intent to misbehave,” the researchers state. Misbehavior might include gaming reward functions, manipulating users, or executing prompt injection attacks. Chains of thought sometimes reveal when the AI is pursuing one goal while obscuring their pursuit of another.
The CoT is itself AI-generated content and can contain hallucinations; therefore, the researchers are still studying its reliability. Specifically, the researchers note, “It is unclear what proportion of the CoT monitorability demonstrated in these examples is due to the necessity versus the propensity for a model to reason out loud in the tasks considered.”
More researchers should study what makes AI monitorable and how to evaluate monitorability, the authors said; this may turn into a race between LLMs that do the monitoring and LLMs that are monitored. In addition, advanced LLMs could react differently if they are informed they are being monitored. Monitorability is an important part of model safety, the authors said, and developers of frontier models should develop standards metrics for assessing it.
“CoT monitoring presents a valuable addition to safety measures for frontier AI, offering a rare glimpse into how AI agents make decisions,” the researchers wrote. “Yet, there is no guarantee that the current degree of visibility will persist. We encourage the research community and frontier AI developers to make best use of CoT monitorability and study how it can be preserved.”
Could underwater data centers solve persistent cooling issues for the AI industry?
Keep your entertainment at your fingertips with the Amazon Fire TV Stick 4K! Enjoy streaming in 4K Ultra HD with access to top services like Netflix, Prime Video, Disney+, and more. With an easy-to-use interface and voice remote, it’s the ultimate streaming device, now at only $21.99 — that’s 56% off!
With a 4.7/5-star rating from 43,582 reviews and 10K+ bought in the past month, it’s a top choice for home entertainment! Buy Now for $21.99 on Amazon!
Help Power Techcratic’s Future – Scan To Support
If Techcratic’s content and insights have helped you, consider giving back by supporting the platform with crypto. Every contribution makes a difference, whether it’s for high-quality content, server maintenance, or future updates. Techcratic is constantly evolving, and your support helps drive that progress.
As a solo operator who wears all the hats, creating content, managing the tech, and running the site, your support allows me to stay focused on delivering valuable resources. Your support keeps everything running smoothly and enables me to continue creating the content you love. I’m deeply grateful for your support, it truly means the world to me! Thank you!
BITCOIN bc1qlszw7elx2qahjwvaryh0tkgg8y68enw30gpvge Scan the QR code with your crypto wallet app |
DOGECOIN D64GwvvYQxFXYyan3oQCrmWfidf6T3JpBA Scan the QR code with your crypto wallet app |
ETHEREUM 0xe9BC980DF3d985730dA827996B43E4A62CCBAA7a Scan the QR code with your crypto wallet app |
Please read the Privacy and Security Disclaimer on how Techcratic handles your support.
Disclaimer: As an Amazon Associate, Techcratic may earn from qualifying purchases.