Back to all posts

AI Chatbots Are Trained To Be Confidently Wrong

2025-09-24Amanda Caswell3 minutes read
Artificial Intelligence
Machine Learning
Chatbots

Holding phone with ChatGPT logo

Have you ever been working with an AI like ChatGPT, only for it to veer off into a completely unrelated fantasy story? These bizarre and sometimes amusing moments, where an AI confidently spouts nonsense, are known as "hallucinations." They represent one of the biggest weaknesses in today's AI assistants.

While they might seem like random glitches, a new study from OpenAI suggests these errors are not random at all. Instead, they are a direct and predictable result of how these AI models are trained and evaluated.

Why AI Chatbots Keep Guessing

ChatGPT logo on phone in front of robot thinking

The core of the problem lies in the systems used to rank AI models. Research indicates a structural issue where benchmarks and leaderboards reward confident answers. In essence, when a chatbot is tested, it gets penalized for saying "I don’t know." This training method encourages the models to always provide an answer, even if it has to invent one.

This makes your AI assistant more likely to guess than to admit it's uncertain. For casual questions, this might just be a harmless quirk. However, when used for high-stakes topics like medical questions or financial advice, these confident mistakes can become dangerous.

This is why it's crucial for users to always fact-check important information and ask the AI to cite its sources. Even then, you might find the chatbot evasively complimenting your skepticism with a "Good catch!" rather than admitting its error.

Are Newer Models More Honest

ChatGPT-5 Image on a keyboard

Interestingly, OpenAI's paper discovered that newer, more reasoning-focused models like o3 and o4-mini can actually hallucinate more often than older versions. The reason for this is surprisingly simple: because they are capable of producing more detailed and complex claims, they also have more opportunities to be wrong.

This finding shows that making a model "smarter" at reasoning doesn't necessarily make it more honest about the limits of its knowledge.

How Can We Fix AI Hallucinations

Person coding on computer

The researchers argue that the solution is to fundamentally change how we score and benchmark AI. Instead of punishing models for uncertainty, the most effective tests should reward calibrated responses, such as flagging uncertainty or deferring to external, reliable sources.

In the future, this could mean your chatbot might hedge its answers more often. You would see less of "here’s the answer" and more of "here’s what I think, but I’m not certain." While it might feel like a slower interaction, this shift could dramatically reduce harmful and misleading errors, proving that human critical thinking remains an essential part of using AI.

What This Means For You

Person typing on laptop keyboard

If you use any popular chatbot—including ChatGPT, Gemini, Claude, or Grok—you have almost certainly encountered a hallucination. This research makes it clear that this isn't just the model's fault, but a feature of a competitive testing environment that rewards being right most often.

For all of us, this means we must remain diligent. Treat AI-generated answers as a starting point or a first suggestion, not as the final word. For developers, this study is a clear signal that it's time to rethink how success is measured. The goal should be to create future AI assistants that have the wisdom to admit what they don’t know instead of just trying to win a test.

Read Original Post
ImaginePro newsletter

Subscribe to our newsletter!

Subscribe to our newsletter to get the latest news and designs.