The Fight to Stop AI Hallucinations

by Doug McCord
August 07, 2024
Combating AI Hallucinations

If you used AI much at all, you’ve been lied to. Told that glue is tasty on pizzas, that daily rocks can be good for you, or that US presidents from centuries past have recently graduated from the University of Wisconsin.  

This is nothing new, of course, but for a technology with so many applications, it’s proving disturbingly hard to get rid of.  

When computer and cognitive scientist Douglas Hofstadter worked with OpenAI’s GPT-3 back in 2022, he reported numerous ways to get it to answer with garbage, with one famous example being: 

Q: When was the Golden Gate Bridge transported for the second time across Egypt? 

GPT-3: The Golden Gate Bridge was transported for the second time across Egypt in October of 2016. 

Still it continues apace. In the legal field, where adoption’s high (3/4 of lawyers report planning to use genAI), hallucinations continue to wreak havoc, with a recent study from Stanford and Yale (published April 2024) reporting a disturbingly high rate of fabrication with GPT-4 and other public models (“at least 58% of the time”).  

And yet in surveys by Tidio, for example, 72% of users report trusting AI to provide “reliable and truthful information,” even as 77% of them admit to being deceived by hallucinations.  

The general view of AI hallucination risks seems to be that they’re a growing pain, but one we will soon move past. Even as researchers are not so sure.   

One thing is certain, with so much at stake, billions will be spent to improve generative AI’s factual reliability. 

In this week’s edition of The PTP Report, we focus on preventing AI hallucinations: why they persist, the best performers, and current research reducing their impact. 

Hallucinations: Why AIs Just Can’t Stop Lying 

Any discussion of “hallucination” has to begin with a definition of the term. We all agree it includes cases where an AI returns information that is factually untrue