Sign In

Communications of the ACM

ACM News

Chatbots May 'Hallucinate' More Often Than Many Realize

View as: Print Mobile App Share:
Amr Awadallah, the chief executive of Vectara, warns that its chatbot software doesnt always tell the truth.

“Hallucinations” may not be a problem for people tinkering with chatbots on their personal computers, but it is a serious issue for anyone using this technology with court documents, medical information, or sensitive business data.

Credit: Cayce Clifford/The New York Times

When the San Francisco start-up OpenAI unveiled its ChatGPT online chatbot late last year, millions were wowed by the humanlike way it answered questions, wrote poetry and discussed almost any topic. But most people were slow to realize that this new kind of chatbot often makes things up.

When Google introduced a similar chatbot several weeks later, it spewed nonsense about the James Webb telescope. The next day, Microsoft's new Bing chatbot offered up all sorts of bogus information about the Gap, Mexican nightlife and the singer Billie Eilish. Then, in March, ChatGPT cited a half dozen fake court cases while writing a 10-page legal brief that a lawyer submitted to a federal judge in Manhattan.

Now a new start-up called Vectara, founded by former Google employees, is trying to figure out how often chatbots veer from the truth. The company's research estimates that even in situations designed to prevent it from happening, chatbots invent information at least 3%of the time — and as high as 27%.

From The New York Times
View Full Article – May Require Paid Registration



No entries found