

1·
10 days agoThe LLM doesn’t know how confident it is, and even if you ask it, it’s just going to pick whichever number came up most often when people asked each other how confident they were in the model’s training data. That’s still an unsolved problem beyond basic tasks like getting an LLM to find sentences in some prose that aren’t supported by any of the sentences in some other prose.
That’ll tell you how likely it is that an answer’s gramatically correct and how likely it is to resemble text in the training data. If the LLM starts reciting a well known piece of literature that shows up several times in the training data and finishes it correctly, that’ll give a really good score by this metric, but no matter how much the LLM output resembles the script of Bee Movie, Bee Movie isn’t a true story or necessarily relevant to the question a user asked.