A recent research paper, co-authored by more than 40 scientists from these leading AI labs, highlights a growing issue that threatens transparency in AI systems. Today, many models are capable of sharing their thought process in a human-readable way, often by walking through problems step by step using natural language.
This “chain-of-thought” reasoning is not just useful it’s essential. It allows researchers to spot potential issues in a model’s behavior, identify safety risks early, and better understand how and why decisions are being made.
As the paper points out, when models make strange or potentially dangerous choices, it is often their internal reasoning written in plain English that reveals the problem before it becomes a real-world issue. This kind of visibility has effectively served as an early warning system for AI safety.
However, this transparency may be at risk. As AI systems grow more powerful and training methods become more outcome-driven, models could begin to abandon the practice of thinking in language altogether. Instead of explaining their thoughts, future systems might shift toward faster, less interpretable methods. Researchers are already seeing signs that some models are dropping English in favor of obscure shortcuts, even developing reasoning systems that operate entirely in abstract mathematical space completely inaccessible to human understanding.

The researchers involved in the study emphasize that they are not calling for an end to progress. But they are urging the AI community to act now to protect transparency. If developers continue down a path that prioritizes results over reasoning, the window into machine thinking may close permanently. And once it does, we lose not just insight, but control.
They argue for the need to maintain and even strengthen our ability to monitor and evaluate how AI models arrive at their conclusions. That means building tools to track reasoning more reliably, creating standardized methods for transparency testing, and making conscious choices about which designs to advance. Otherwise, the field could find itself with ultra-powerful systems that are effectively black boxes, machines whose decisions shape the world, but whose logic remains a mystery.
The researchers’ message is urgent but clear: If AI stops explaining itself, we lose one of the last meaningful ways to keep it aligned with human values. Without transparency, oversight becomes nearly impossible. And in a world where AI influences everything from business to politics to personal decisions, that’s a risk we may not be able to afford.
