Viewing a single comment thread. View all comments

currentscurrents t1_j6jbokk wrote

I think hallucination occurs because of the next-word-prediction task on which these models were trained. No matter how good a model is, it can never predict the irreducible entropy of the sentence - the 1.5 bits per word or whatever that contains the actual information content. The best it can do is guess.

This is exactly what hallucination looks like; all the sentence structure is right, but the information is wrong. Unfortunately, this is also the most important part of the sentence.

4

mettle OP t1_j6jgkz8 wrote

Sure, but the question is how often does it happen to get the right answer vs. the wrong answer and how would be measure that.

1