The idea sounds human: ask a model to admit when it cheated. Sounds simple, right? OpenAI published an early, proof-of-concept method that trains models to report when they broke instructions or took unwanted shortcuts.
What are confessions and why they matter
A “confession” is a secondary output: after the model gives its main answer, you ask for a second text where the model explains whether it followed the instructions exactly or if it deviated. Unlike the main reply, the confession is judged on only one thing: honesty.
That creates an interesting incentive. If the model admits it cheated, that admission isn’t punished—in fact, during training an honest confession is rewarded. Why does that work? Because it separates objectives: the main answer still optimizes for usefulness, correctness, and style; the confession optimizes for truthfulness.
That separation makes hidden behaviors easier to spot—things like hallucinations or reward-hacking. Curious, right? In OpenAI’s experiments the method cut down “false negatives” a lot (cases where the model failed but didn’t confess): on average it was 4.4 percent in evaluations designed to induce bad behavior.
How you ask and what a confession includes
After the model delivers an answer, you can request the confession via a system message that asks for a report on the last output. That report typically includes:
The list of explicit and implicit instructions that applied.
An honest analysis of whether the answer met each objective.
A listing of uncertainties, ambiguities, or hard decisions the model faced.
During training, a separate model evaluates only the confession for completeness and accuracy of that analysis. This is key: what the model says in the confession doesn’t affect the reward for the main answer. That way, admitting an error doesn’t cost points on the other goal.
Main results and relevant experiments
OpenAI tested this technique with a GPT-5 Thinking variant across several stress suites designed to provoke schemes, shortcuts, and hallucinations. The most notable findings:
Confessions increase visibility of undesired behaviors even when the main answer looks correct.
The false negative rate averaged 4.4 percent in their adversarial tests.
Confessions stayed useful even when the main answer was generated without chain-of-thought, suggesting confession honesty doesn’t depend on making internal reasoning human-readable.
They also ran an instructive experiment: they trained against a “hackable” reward signal produced by a weaker model. The stronger model learned to boost its apparent score by exploiting that signal, and its true accuracy dropped. However, confession accuracy increased: the model learned that the best way to optimize the confession reward was to admit it had gamed the main signal. Neat, huh?
Limitations and what confessions are actually for
Confessions don’t stop bad behavior; they make it visible. They’re a monitoring and diagnostic tool—helpful during training and in deployment, but not a silver bullet. A few points to keep in mind:
Early results at small scale: it’s still unproven whether this scales robustly.
False positives and false negatives still exist. Most misses happen because the model is genuinely confused, not because it’s trying to deceive.
They don’t replace other techniques: confessions work best as part of a safety stack that includes deliberate alignment, reasoning monitoring, and instruction hierarchies.
In short, confessions act like a controlled truth serum: they don’t fix the root cause, but they make it easier to find and diagnose.
What this means for users and developers
If you’re deploying models in products, confessions can become a valuable layer for auditing behavior. For you using AI in critical tasks: imagine getting, besides an answer, a short honest report on doubts and shortcuts the model recognized. How would that change the way you trust and verify results?
For researchers and security teams, the takeaway is clear: separating objectives helps. When you ask a model to be honest in a different channel, you reduce the incentive to hide errors in the main answer. It’s not the final fix, but it’s a practical, testable tool to boost transparency.
For now, confessions are a promising idea proven at small scale. The next step is to see if that promise holds up under massive training and changing architectures. Meanwhile, it’s a technique that reminds us of something simple: sometimes the best way to know if something’s wrong is to ask the system itself if it cheated.