2/3 - “Your door is red” is a grammatically correct statement, but may not accurately match intended door object.
Asking,”What colour is my door?”, that’s 1 possible answer in infinite array of colour choice.
Without grounding the question in a reference, like detailed information about building in question & a point in time, as same door could’ve had various colours over time, LLM will grammatically construct the reply & pick a colour with highest probability in it’s imprinting data I.e. red.
Yes! I think that for many, when they ask questions of an LLM, they expect the capabilities that we imagine an AGI might have - like deciding if a question can receive a vague, inaccurate or generic answer vs. there is only one possible answer, and if it's unknown then the model must ask questions back.
I think that some “common sense” stuff is rooted purely in language, and LLMs will pick up the pattern. Like a thing usually can’t be both important and unimportant at the same time; the LLM will encode those two words with anti parallel state vectors.
But that’s because “common sense” is a real grab bag of stuff.
It does the same to ‘big’ and ‘small’ although it has no comprehension of size.