Skip to main content


ELI5

What is up with the LLM hallucinations?

I am assuming this is a fancy name for the byproducts of fuzzy logic capabilities... or is there more at work here?

Any good primers would be appreciated.

in reply to The Gibson πŸ……

you mean like, AI that intentionally does it? like daydream?

or when AI accidentally does it?

or when people hack into it?

in reply to {Insert Pasta Pun}

@risottobias when the system accidentally (I assume, that phrasing troubles me because it indicates the presence of will) generates a similar, but not accurate answer.
in reply to The Gibson πŸ……

AI is built on a corpus of confident sounding responses.

it is trained to sound confident, even if there's no basis in reality.

the sentences that it tunes to are meant to seem realistic.

nothing about it is unit tested to whether it really works.

it writes fiction.

in reply to {Insert Pasta Pun}

@risottobias hmmm

Except it doesn't sometimes. When asked to do a specific function and then reverse that function, the answers don't always match, for example.

This is not a cute hallucination, but sort of befuddlingly useless.

in reply to The Gibson πŸ……

@The_Gibson :hackers_town: @Risotto Bias I want to be clear that it *always* writes fiction.

ChatGPT can't run a function nor can it know what a function is. It just runs down a path in a tokenizer, it doesn't really run code or whatever .

in reply to The Gibson πŸ……

yeah more like a terrible name for fuzziness. The LLMs don’t know what words mean, or what true is. They just know how words go together, and as a result can write a ton of fiction interspersed with fact. The way they write doesn’t help, they’ve been tutored in writing in a confident, knowledgeable style... which given they can just write fiction is a bit of a poor choice.
in reply to The Gibson πŸ……

yeah exactly. The machine is spitting out not-quite-random stochastically linked words. The magic isn’t in the machine, it’s in our heads.
in reply to The Gibson πŸ……

LLMs simply don’t know what is factual and what isn’t, they merely statistically infer what is a highly probable thing to say next based on what it has received as input. Any correct output is mere coincidence or β€œlucky” in the sense that training data created very highly probable output for a given topic. Sometimes it simply statistically takes a wrong turn in large or small ways and mixes that into the reply. It is entirely made up but statistically plausible from a β€œingested the entire internet into probabilities” perspective…
in reply to The Gibson πŸ……

it is literally probabilistically generating the next word based on its neural net weightings. those weightings cannot be right every time for every context because, math, but also the LLMs are trained on the sum of written English accessible from the Internet.

similar to how the image detection systems are only 94% sure that's a face in the image. but tweak the input a bit and now it's an apple (adversarial machine learning). these systems are fragile and basically piles of numbers in matrices, without knowledge or logic encoded in them.

Unknown parent

in reply to The Gibson πŸ……

@theruran Joseph wrote an entire book trying to explain this effect and how computers aren’t actually intelligent after he invented Eliza and the entire computer science field immediately fell into the belief that it was, in fact, conscious and knowingly answering their questions.
in reply to The Gibson πŸ……

@vortex_egg it falls apart quickly when you prompt it do basic math. It's like it dug up some answers on Quora and Reddit and mixed them together.

I've found it's more helpful to prompt me, as a feedback mechanism to check my own knowledge. because its answers are frequently slightly wrong

in reply to The Gibson πŸ……

I also find it confounding. I asked an LLM for the Metroid cheat code. It spit back β€œJustin Bailey.” Correct. Then I asked for some SimCity 2000 cheat codes and it spit back garbage. Like pure garbage made up fantasy cheat codes. Why? Surely both appear in the training corpus, probably many times over. Why the correct result for one but not the other?
in reply to The Gibson πŸ……

building a response from an LLM involves following a sort of "pathway" through data. Like a very complex version of predictive text that takes a lot more variables into consideration when deciding how to interpret input and generate output.

You can think of a hallucinations as the algorithm sort of "taking a wrong turn" and going down a pathway that has factual errors or incorrect connections or what have you.

Yes, it's a lot more complex than that, but keeping it ELI5…

in reply to theruran 🌐🏴

@theruran This is also why it consistently fabricates citations and books that don’t exist and assigns them to authors who didn’t write them. The best way to think about it is that it tries to statistically answer your prompt in the way that [a superposition of a random person on reddit|quora|stormfront] would be likely to respond to your prompt, one token at a time. So the responses sound β€œlike” a reasonable response when they comply with your expectation bias, and they sound β€œlike” a hallucination when they comply with your expectation bias but you also have the curse of knowledge to know it isn’t accurate.
in reply to Scott, Drowning in Information

@vortex_egg @theruran in short, LLMs are bullshit generators. With tuning, the bullshit can often have utility; sometimes it can even seem like it's not bullshit at all. But it's still actually just bullshit

Bullshitting this convincingly has some useful and practical applications, and it's an interesting advance in NLP and a couple other parts of AI research. But it's not even close to "thinking"

in reply to The Gibson πŸ……

LLMs don't answer questions, they generate text that looks like what an answer to a question that looks like your question looks like. It's got enough text from the internet in it that something that looks like an answer is often the right answer because, well, your not as creative as you think and someone's probably had the same conversation you're having before and the LLM eavesdropped on it. Hallucinations are just when text that looks like an answer happens to not be right. It's not really a special case, it's just the other side of the same coin.
in reply to montag

so let me get this straight... You're telling me, that some marketing guy wormed his way into the engineering team, and decided to anthropomorphize a pile of data by pretending it has a mind and a will?!?

Who would do such a thing? I find it irresponsible.

This entry was edited (1 year ago)

I am Jack's Lost 404 reshared this.

in reply to The Gibson πŸ……

@montag I think you would be misreading the situation by denying that engineers are the ones driving this. If you go into the spaces of the people who are the biggest true believers that this stuff is *actually already sentient*, they are living and breathing linear algebra and python.
in reply to Scott, Drowning in Information

@vortex_egg @montag
Not sure about this... not a single one of my engineering colleagues thinks this or is a fan of AI in general.
It is seen as a new set of tools that might be useful for very specific applications instead.
Unknown parent

calcifer
I am a fire demon, or such a convincing simulation of one that the difference is unimportant
in reply to Scott, Drowning in Information

@vortex_egg so, it's like (authors of) a TV show explaining something; it may sound plausible because they use terms from that field, but when you actually have knowledge of that field, you realise it quickly falls apart?
@theruran @TheGibson
in reply to The Gibson πŸ……

I understand like 5% of this, but here’s probably the best primer I’ve come across: writings.stephenwolfram.com/20…
in reply to The Gibson πŸ……

I like this answer:
social.coop/@DrewKadel/1101540…

@DrewKadel


My daughter, who has had a degree in computer science for 25 years, posted this observation about ChatGPT on Facebook. It's the best description I've seen:
This entry was edited (1 year ago)
in reply to Scott, Drowning in Information

the fake citations suck. I'm like, oh great, I wish that paper existed but it doesn't. it's just giving me something based on my input.
⇧