OpenAI's latest AI models are generating alarming levels of misinformation

Agencies Ghacks
May 7, 2025
Updated • May 7, 2025
Internet
|
6

OpenAI's latest AI models seem to have a big problem. A report reveals that the GPT o3 and o4-mini are producing misinformation at an alarming rate.

AI-generated misinformation, aka hallucinations, are common among most artificial intelligence services. The New York Times has published an investigation conducted by OpenAI that discovered its own models are generating more fake content than others. This in turn has raised serious concerns about their reliability.

GPT o3 and o4-mini have been designed to mimic human reasoning and logic. When these were put to the test in benchmarks involving public figures, nearly one-third of GPT o3's results were found to be hallucinations. In comparison, GPT o1 had less than half of that error rate in tests that were conducted last year. GPT o4-mini fared even worse, as it hallucinated on 48% of its tasks. When these models tackled general knowledge questions, hallucinations soared to 51% for GPT o3, and a staggering 79% for o4-mini.

OpenAI says that the hallucinating problem is not because the reasoning models are worse, but because they could simply be more verbose and adventurous in their answers, and are speculating possibilities rather than repeating predictable facts. Developers initially aimed for these systems to think critically and reason through complex queries; however, this ambitious approach appears to have led to an increase in creativity at the expense of factuality.

This could pose a big problem for OpenAI's ChatGPT, as rival services like Google Gemini, Anthropic Claude, have been designed to provide information more accurately. Unlike simpler models focused on high-confidence predictions, GPT o3 and o4-mini often speculate, blurring the line between possible scenarios and outright fabrications. This raises red flags for users in high-stakes environments, from legal professionals to educators and healthcare providers, where reliance on AI could lead to significant missteps.

The more useful AI becomes, the greater the potential for critical errors. While AI models may outperform humans in certain tasks, the risk of inaccuracies diminishes AI's overall credibility. Until these hallucination issues are effectively addressed, users are advised to approach AI-generated information with caution and skepticism.

Source: Tech Radar

Advertisement

Tutorials & Tips


Previous Post: «
Next Post: «

Comments

  1. Anonymous said on May 8, 2025 at 9:34 am
    Reply

    Okay, who crawled the Yahoo Answers archives? Should have stuck with Reddit – nobody ever makes things up there…

  2. Richard Hack said on May 8, 2025 at 3:24 am
    Reply

    So, in other words, they actually act like humans for once. LOL Most of the humans I see on the Internet are hallucinating “facts” at least 90 percent of the time.

    1. boris said on May 8, 2025 at 3:35 am
      Reply

      The problem is that when people ask questions from Chatbots or Google Snippets, people believe they’re asking a reputable source, not some rando from the internet.

  3. boris said on May 7, 2025 at 10:21 pm
    Reply

    So OpenAI admitted that they have problems. Actually pretty refreshing. I still believe that Google Search snippets generate mostly fake responses because they train their model on Reddit answers.

  4. KNTRO said on May 7, 2025 at 1:39 pm
    Reply

    “The New York Times has published an investigation conducted by OpenAI has revealed that points out OpenAI’s models are generating more fake content than others.”

    It seems like gHacks articles uses these very same AI models! ?

    1. Tom Hawack said on May 7, 2025 at 6:31 pm
      Reply

      It seems? Doesn’t seem to me, at least not on the basis of fake content. Accuracy here is a tradition.
      But maybe has your very comment been carried out by an AI given speculation here confines to hallucination, my dear friend :)

      Concerning AI, let’s face it : even with a narrow 10% of hallucinations it wouldn’t deserve the qualification of intelligence, IMO. Assertions must be bullet-proof in terms of factual, otherwise speculations must be notified as such. Like humans being when they state “I know” rather than “I think, I believe” … at least when they do differentiate their imagination from their knowledge which, nowadays, is on a negative climb, so to say (lol).

Leave a Reply

Check the box to consent to your data being stored in line with the guidelines set out in our privacy policy

We love comments and welcome thoughtful and civilized discussion. Rudeness and personal attacks will not be tolerated. Please stay on-topic.
Please note that your comment may not appear immediately after you post it.