Facefam ArticlesFacefam Articles
  • webmaster
    • How to
    • Developers
    • Hosting
    • monetization
    • Reports
  • Technology
    • Software
  • Downloads
    • Windows
    • android
    • PHP Scripts
    • CMS
  • REVIEWS
  • Donate
  • Join Facefam
Search

Archives

  • August 2025
  • July 2025
  • June 2025
  • May 2025
  • April 2025
  • March 2025
  • January 2025
  • December 2024
  • November 2024

Categories

  • Advertiser
  • AI
  • android
  • betting
  • Bongo
  • Business
  • CMS
  • cryptocurrency
  • Developers
  • Development
  • Downloads
  • Entertainment
  • Entrepreneur
  • Finacial
  • General
  • Hosting
  • How to
  • insuarance
  • Internet
  • Kenya
  • monetization
  • Music
  • News
  • Phones
  • PHP Scripts
  • Reports
  • REVIEWS
  • RUSSIA
  • Software
  • Technology
  • Tips
  • Tragic
  • Ukraine
  • Uncategorized
  • USA
  • webmaster
  • webmaster
  • Windows
  • Women Empowerment
  • Wordpress
  • Wp Plugins
  • Wp themes
Facefam 2025
Notification Show More
Font ResizerAa
Facefam ArticlesFacefam Articles
Font ResizerAa
  • Submit a Post
  • Donate
  • Join Facefam social
Search
  • webmaster
    • How to
    • Developers
    • Hosting
    • monetization
    • Reports
  • Technology
    • Software
  • Downloads
    • Windows
    • android
    • PHP Scripts
    • CMS
  • REVIEWS
  • Donate
  • Join Facefam
Have an existing account? Sign In
Follow US
Technologywebmaster

Monitor AI’s Decision-Making Black Box: Here’s Why

Ronald Kenyatta
Last updated: July 22, 2025 5:43 am
By
Ronald Kenyatta
ByRonald Kenyatta
Follow:
Share
4 Min Read
SHARE

Contents
What is chain-of-thought?Chain-of-thought oversight could check ‘misbehavior’ of advanced AI models
Exevutives using AI computing simulation.
Image: Envato/DC_Studio

Monitoring generative AI’s decision-making is critical for safety, but the inner workings that lead to text or image outputs remain largely opaque. A position paper released on July 15 proposes chain-of-thought (CoT) monitorability as a way to watch over the models.

The paper was co-authored by researchers from Anthropic, OpenAI, Google DeepMind, the Center for AI Safety, and other institutions. It was endorsed by high-profile AI experts, including former OpenAI chief scientist and Safe Superintelligence co-founder Ilya Sutskever, Anthropic researcher Samuel R. Bowman, Thinking Machines chief scientist John Schulman, and deep learning luminary Geoffrey Hinton.

What is chain-of-thought?

Chain-of-thought refers to the intermediate reasoning steps a generative AI model verbalizes as it works toward an output. Some deep research models produce reports to their users of what they are doing at the time. Assessing what the models are doing before they produce human-readable data is known as interpretability, a field Anthropic has heavily researched.

However, as AI becomes more advanced, monitoring the “black box” of decision-making becomes increasingly difficult. Whether chain-of-thought interpretability will work in a few years is anyone’s guess, but for now, the researchers are pursuing it with some urgency.

“Because CoT monitorability may be fragile, we recommend that frontier model developers consider the impact of development decisions on CoT monitorability,” the researchers wrote.

Chain-of-thought oversight could check ‘misbehavior’ of advanced AI models

“AI systems that ‘think’ in human language offer a unique opportunity for AI safety: we can monitor their chains of thought (CoT) for the intent to misbehave,” the researchers state. Misbehavior might include gaming reward functions, manipulating users, or executing prompt injection attacks. Chains of thought sometimes reveal when the AI is pursuing one goal while obscuring their pursuit of another.

The CoT is itself AI-generated content and can contain hallucinations; therefore, the researchers are still studying its reliability. Specifically, the researchers note, “It is unclear what proportion of the CoT monitorability demonstrated in these examples is due to the necessity versus the propensity for a model to reason out loud in the tasks considered.”

More researchers should study what makes AI monitorable and how to evaluate monitorability, the authors said; this may turn into a race between LLMs that do the monitoring and LLMs that are monitored. In addition, advanced LLMs could react differently if they are informed they are being monitored. Monitorability is an important part of model safety, the authors said, and developers of frontier models should develop standards metrics for assessing it.

“CoT monitoring presents a valuable addition to safety measures for frontier AI, offering a rare glimpse into how AI agents make decisions,” the researchers wrote. “Yet, there is no guarantee that the current degree of visibility will persist. We encourage the research community and frontier AI developers to make best use of CoT monitorability and study how it can be preserved.”

Could underwater data centers solve persistent cooling issues for the AI industry?  

TAGGED:AIsBlackBoxDecisionMakingHeresMonitor
Share This Article
Facebook Whatsapp Whatsapp Email Copy Link Print
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Critical Flaw in NVIDIA AI Toolkit Flaw Puts Cloud Services at Risk Critical Flaw in NVIDIA AI Toolkit Flaw Puts Cloud Services at Risk
Next Article OpenAI Wins Gold at International Math Olympiad – or Did It? OpenAI Wins Gold at International Math Olympiad – or Did It?
Leave a review

Leave a Review Cancel reply

Your email address will not be published. Required fields are marked *

Please select a rating!

Meta Strikes $10 Billion Cloud Deal With Google to Boost AI Capacity
NVIDIA CEO Dismisses Chip Security Allegations as China Orders Firms to Halt Purchases
Anthropic Folds Claude Code Into Business Plans With Governance Tools
Google Claims One Gemini AI Prompt Uses Five Drops of Water
Generate AI Business Infographics without the Fees

Recent Posts

  • Meta Strikes $10 Billion Cloud Deal With Google to Boost AI Capacity
  • NVIDIA CEO Dismisses Chip Security Allegations as China Orders Firms to Halt Purchases
  • Anthropic Folds Claude Code Into Business Plans With Governance Tools
  • Google Claims One Gemini AI Prompt Uses Five Drops of Water
  • Generate AI Business Infographics without the Fees

Recent Comments

  1. https://tubemp4.ru on Best Features of PHPFox Social Network Script
  2. Вулкан Платинум on Best Features of PHPFox Social Network Script
  3. Вулкан Платинум официальный on Best Features of PHPFox Social Network Script
  4. Best Quality SEO Backlinks on DDoS Attacks Now Key Weapons in Geopolitical Conflicts, NETSCOUT Warns
  5. http://boyarka-inform.com on Comparing Wowonder and ShaunSocial

You Might Also Like

IT Leader’s Guide to the Metaverse

August 21, 2025
State of AI Adoption in Financial Services: A TechRepublic Exclusive
Technologywebmaster

State of AI Adoption in Financial Services: A TechRepublic Exclusive

August 21, 2025
AI Underperforms in Reality, and the Stock Market is Feeling It
Technologywebmaster

AI Underperforms in Reality, and the Stock Market is Feeling It

August 21, 2025
Google Shows Off Pixel 10 Series and Pixel Watch 4
Technologywebmaster

Google Shows Off Pixel 10 Series and Pixel Watch 4

August 21, 2025
NVIDIA & NSF to Build Fully Open AI Models for Science
Technologywebmaster

NVIDIA & NSF to Build Fully Open AI Models for Science

August 20, 2025
Previous Next
Facefam ArticlesFacefam Articles
Facefam Articles 2025
  • Submit a Post
  • Donate
  • Join Facefam social
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?

Not a member? Sign Up