News: 1758117792

  ARM Give a man a fire and he's warm for a day, but set fire to him and he's warm for the rest of his life (Terry Pratchett, Jingo)

OpenAI says models are programmed to make stuff up instead of admitting ignorance

(2025/09/17)


AI models often produce false outputs, or "hallucinations." Now OpenAI has admitted they may result from fundamental mistakes it makes when training its models.

The admission came in a [1]paper [PDF] published in early September, titled "Why Language Models Hallucinate," and penned by three OpenAI researchers and Santosh Vempala, a distinguished professor of computer science at Georgia Institute of Technology. It concludes that "the majority of mainstream evaluations reward hallucinatory behavior."

Language models are primarily evaluated using exams that penalize uncertainty

The fundamental problem is that AI models are trained to reward guesswork, rather than the correct answer. Guessing might produce a superficially suitable answer. Telling users your AI can't find an answer is less satisfying.

As a test case, the team tried to get an OpenAI bot to report the birthday of one of the paper's authors, OpenAI research scientist Adam Tauman Kalai. It produced three incorrect results because the trainers taught the engine to return an answer, rather than admit ignorance.

"Over thousands of test questions, the guessing model ends up looking better on scoreboards than a careful model that admits uncertainty," OpenAI admitted in a blog post accompanying the release.

[2]

The authors explained that the pretraining stage of AI model building embeds this unhelpful behavior because the info trainers feed into models contains many examples of certain data – such as correct spellings of words. If a few misspellings make it into the corpus used to train a model, AIs still have many examples of correct spellings and can learn how to produce accurate results.

[3]

[4]

But when the corpus used to train a model does not contain a learnable pattern of data, such as in the birthday example, the AI takes a shot – and often misses.

"The hallucination rate, after pretraining, should be at least the fraction of training facts that appear once," the paper states.

[5]

"For instance, if 20 percent of birthday facts appear exactly once in the pretraining data, then one expects base models to hallucinate on at least 20 percent of birthday facts."

Techniques used in the post-training stage of model development exacerbate the situation.

"Many language-model benchmarks mirror standardized human exams, using binary metrics such as accuracy or pass-rate," the paper states.

[6]

"Optimizing models for these benchmarks may therefore foster hallucinations. Humans learn the value of expressing uncertainty outside of school, in the school of hard knocks. On the other hand, language models are primarily evaluated using exams that penalize uncertainty."

[7]OpenAI's GPT-5 looks less like AI evolution and more like cost cutting

[8]'Suddenly deprecating old models' users depended on a 'mistake,' admits OpenAI's Altman

[9]Open the pod bay door, GPT-4o

[10]White House bans 'woke' AI, but LLMs don't know the truth

Ultimately, it's about stating something, even if it's wrong. The authors liken it to a multiple-choice questionnaire where even if you pick vaguely plausible answers at random, you are likely to score better than if you pick no answers at all.

"We argue that the majority of mainstream evaluations reward hallucinatory behavior," they conclude. "Simple modifications of mainstream evaluations can realign incentives, rewarding appropriate expressions of uncertainty rather than penalizing them. This can remove barriers to the suppression of hallucinations, and open the door to future work on nuanced language models."

In theory, AI model makers could eliminate hallucinations by using a dataset that contains no errors. But the paper admits such a scenario isn't remotely possible, particularly since the huge volumes of data used in training likely contain mistakes.

The more palatable answer, OpenAI suggests, is to adapt models so they more often respond with "I don't know," even if that deters users. The outfit [11]claims to have adapted its training regime to account for this with ChatGPT-5, but in this hack's experience, users of the new model will still find it produces some absolute howlers.

We've asked the authors for clarification and will add more data as it comes in – verified by a human. ®

Get our [12]Tech Resources



[1] https://arxiv.org/pdf/2509.04664

[2] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=2&c=2aMrbGJM9t-ZF6drTGXbYQAAAAsY&t=ct%3Dns%26unitnum%3D2%26raptor%3Dcondor%26pos%3Dtop%26test%3D0

[3] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aMrbGJM9t-ZF6drTGXbYQAAAAsY&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0

[4] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aMrbGJM9t-ZF6drTGXbYQAAAAsY&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0

[5] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aMrbGJM9t-ZF6drTGXbYQAAAAsY&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0

[6] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aMrbGJM9t-ZF6drTGXbYQAAAAsY&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0

[7] https://www.theregister.com/2025/08/13/gpt_5_cost_cutting/

[8] https://www.theregister.com/2025/08/11/openai_tweaks_gpt5_after_user/

[9] https://www.theregister.com/2025/08/20/gpt4o_pod_bay_door/

[10] https://www.theregister.com/2025/07/24/white_house_wants_no_woke_ai/

[11] https://www.theregister.com/2025/08/07/openai_gpt_5/

[12] https://whitepapers.theregister.com/



OpenAI says models are programmed to make stuff up instead of admitting ignorance

ITMA

And there we have it!

A.I. going for (senior) management roles.

"Even a wrong answer is right some of the time"

Pascal Monett

I note that you didn't go for for the time-honored "even a broken watch is right twice a day".

Obviously. The kids with their smartphones grafted to their hand don't know what a watch is . . .

Re: "Even a wrong answer is right some of the time"

werdsmith

Of course they do. It's a wearable that counts your physical activity and monitors your health.

Is this not just how they work?

Alfred

Everything they output is made up. There is no truth and lies, no reality or fiction in there. Just prompts and the high-scoring outputs that are created in response to them. Trying to apply the idea of truth and lie to what they do just makes no sense.

Re: Is this not just how they work?

joewilliamsebs

Exactly; they're language models, not truth machines.

They are very good at producing words that have the shape of an answer - confident language that appears to model logical thought process, but has no actual thought behind it.

I have not taken a lot of money from investors by building one of these so could be wrong, but to me we appear to be going about the whole process backwards.

Re: Is this not just how they work?

LionelB

In other words, LLMs do exactly what they were designed to do (and on the whole they do it rather well): produce plausibly human-like textual responses to textual input queries.

If you think that an LLM is "hallucinating", "making something up" or "giving you the wrong answer", the problem is not the LLM; it is your expectations that are at fault (the clue is in the inappropriately anthropomorphic language).

Re: Is this not just how they work?

ThatOne

> it is your expectations that are at fault

But people have been trained by almost half a century of all-knowing movie robots who always have supernatural knowledge. Have you ever seen a Sci Fi movie robot (or supercomputer) answering "Sorry, I really don't have the faintest idea"?...

Given it's hard to retrain people, I guess it's easier to retrain AIs. It must be simple to add a "beware of dangerous guessing" mechanism, where due to context the AI will rather refrain from hallucinating an authoritative-sounding answer and instead suggest to ask somebody bound to know (like a doctor or a lawyer).

tony72

Guessing might produce a superficially suitable answer. Telling users your AI can't find an answer is less satisfying.

This only mirrors the direction search companies went long before LLMs. Google went from proudly proclaiming it was an AND search, and that every result matched all your keywords back in the day, to the situation now where you have to fight to stop it from ignoring your most important keywords and returning utterly irrelevant results. Apparently lots of bad results is better than few accurate results. Same old same old.

ThatOne

I think it's a different issue here: Google makes money from promoting paying customers, not from handing out relevant search results for free.

In other words, it's a themed ad dispenser, not a search engine. Your search keywords might at best just influence the choice of promoted links you'll get served.

I think therefore I glam

Baird34

I'm not sure if I read all of this article, can't be certain, so I'll just say that I have. It looks better that way.

ChrisElvidge

Ultimately, it's about stating something, even if it's wrong. The authors liken it to a multiple-choice questionnaire where even if you pick vaguely plausible answers at random, you are likely to score better than if you pick no answers at all.

Bad MCQ marking, then. Should get +1 for correct and -(1/(n-1)) for incorrect, where n is the number of alternatives.

Doctor Syntax

"I can't find anything to match" is the correct answer if it that's the situation. Anything else in those circumstances would be incorrect.

ThatOne

That's if you're serious.

But if you're just out to please the crowds you'll deliver a spiel, and some little untruths are part and parcel of that. It's about the performance, not the information. AIs are politicians, they need to woo the voters so they vote with their wallets.

I have countless transcripts

JimmyPage

Where CharGPT has simply Made Shit Up.

parameters that don't exist. File paths that don't exist. Commands that are outdated or simply wrong.

No point telling it that "it has made a mistake" because all you will get is a "sorry" and then a different style of turd. And then, when you've had a few screenfuls, the original wrong answer will cycle back in again.

Ask it to advise on setting a web gui password on pihole v6 in docker. That'll keep you busy for hours.

Re: I have countless transcripts

Paul Herber

I know it was a typo but, how about

CharladyGPT

TaxiDriverGPT

HairdresserGPT

BartenderGPT

PoliticianGPT

All good for ill-informed advice!

GlenP

I've got a test query which I use, relating to our company and sample sizes.

ChatGPT now just says it doesn't have the info (correct answer).

Google has spouted something that is vaguely relevant but factually incorrect and doesn't actually answer the query; when I dive deeper it suggests contacting the company but also states that we are linked to an academic paper where experiments were done in 15ml test tubes - we supply by the pallet or truck load!

It's like my old football coach said....

Androgynous Cow Herd

"You miss 100% of the shots you don't take"

Which is why he coached football rather than teaching math.

Now I lay me back to sleep.
The speaker's dull; the subject's deep.
If he should stop before I wake,
Give me a nudge for goodness' sake.
-- Anonymous