Fixing Hallucinations Would Destroy ChatGPT, Expert Finds
In a paper published earlier this month, OpenAI researchers said they'd found the reason why even the most powerful AI models still suffer from rampant "hallucinations," in which products like ChatGPT confidently make assertions that are factually false. They found that the way we evaluate the output of large language models, like the ones driving ChatGPT, means they're "optimized to be good test-takers" and that "guessing when uncertain improves test performance." In simple terms, the creator