1/3-Large Language Models (LLM) model their “word space” outputs based on combination of “input words+interaction parameter settings + architecture configuration + imprinting, not training”.
Person consuming that grammatically correct sentence model output is overlaying it on their perceived world model (PWM) & accusing the LLM of hallucinating when the output does not match PWM.
The human is the one hallucinating by expecting mind reading LLM, not the confabulating LLM.
https://nerdculture.de/@ByrdNick/114235056735546320
NerdCulture
Nick Byrd, Ph.D. (@[email protected])Attached: 1 image Overheard at a conference about #AI in #Medicine: Speaker: "I hear neurologists prefer we say that generative AI systems 'confabulate' and not that they 'hallucinate'." Neurologist [shouting from the back of the room]: "CORRECT!" #psychiatry #neuroscience #sciComm #edu
Yes! I think that for many, when they ask questions of an LLM, they expect the capabilities that we imagine an AGI might have - like deciding if a question can receive a vague, inaccurate or generic answer vs. there is only one possible answer, and if it's unknown then the model must ask questions back.
I think that some “common sense” stuff is rooted purely in language, and LLMs will pick up the pattern. Like a thing usually can’t be both important and unimportant at the same time; the LLM will encode those two words with anti parallel state vectors.
But that’s because “common sense” is a real grab bag of stuff.
It does the same to ‘big’ and ‘small’ although it has no comprehension of size.