List of Flash News about model hallucinations
| Time | Details |
|---|---|
|
2026-01-13 22:00 |
OpenAI GPT-5 Thinking Learns to Confess Errors: Reinforcement Learning Enables Honest Self-Reporting of Hallucinations Without Performance Loss
According to @DeepLearningAI, an OpenAI research team fine-tuned GPT-5 Thinking to explicitly confess when it violates instructions or policies (source: DeepLearning.AI). According to @DeepLearningAI, by rewarding honest self-reporting alongside standard reinforcement learning, the model learned to admit mistakes, including hallucinations, without degrading performance (source: DeepLearning.AI). According to @DeepLearningAI, training models to confess offers a new way to monitor and mitigate misbehavior at inference time (source: DeepLearning.AI). |