back
Get SIGNAL/NOISE in your inbox daily
Recent work from Anthropic and others claims that LLMs’ chains of thoughts can be “unfaithful”. These papers make an important point: you can’t take everything in the CoT at face value. As a result, people often use these results to conclude the CoT is useless for analyzing and monitoring AIs. Here, instead of asking whether the CoT always contains all information relevant to a model’s decision-making in all problems, we ask if it contains enough information to allow developers to monitor models in practice. Our experiments suggest that it might.
Recent Stories
Jan 19, 2026
Stop ignoring AI risks in finance, MPs tell BoE and FCA
Treasury committee urges regulators and Treasury to take more ‘proactive’ approach
Jan 19, 2026OpenAI CFO Friar: 2026 is year for ‘practical adoption’ of AI
OpenAI CFO Sarah Friar said the company is focused on "practical adoption" in 2026, especially in health, science, and enterprise.
Jan 19, 2026OpenAI’s 2026 ‘focus’ is ‘practical adoption’
As the company spends a huge amount of money on infrastructure, OpenAI is working to close the gap on what AI can do and how people actually use it.