害羞草研究所

Skip to content

Untruthful tech: Chatbots prone to making things up

Not everyone thinks AI害羞草研究所檚 hallucination problem is fixable

Spend enough time with ChatGPT and other artificial intelligence chatbots and it doesn害羞草研究所檛 take long for them to .

Described as hallucination, confabulation or just plain making things up, it害羞草研究所檚 now a problem for every business, organization and high school student trying to get a generative AI system to compose documents and get work done. Some are using it on tasks with the potential for high-stakes consequences, from psychotherapy to researching and .

害羞草研究所淚 don害羞草研究所檛 think that there害羞草研究所檚 any model today that doesn害羞草研究所檛 suffer from some hallucination,害羞草研究所 said Daniela Amodei, co-founder and president of Anthropic, maker of the chatbot Claude 2.

害羞草研究所淭hey害羞草研究所檙e really just sort of designed to predict the next word,害羞草研究所 Amodei said. 害羞草研究所淎nd so there will be some rate at which the model does that inaccurately.害羞草研究所

Anthropic, ChatGPT-maker OpenAI and other major developers of AI systems known as large language models say they害羞草研究所檙e working to make them more truthful.

How long that will take 害羞草研究所 and whether they will ever be good enough to, say, safely dole out medical advice 害羞草研究所 remains to be seen.

害羞草研究所淭his isn害羞草研究所檛 fixable,害羞草研究所 said Emily Bender, a linguistics professor and director of the University of Washington害羞草研究所檚 Computational Linguistics Laboratory. 害羞草研究所淚t害羞草研究所檚 inherent in the mismatch between the technology and the proposed use cases.害羞草研究所

A lot is riding on the reliability of generative . The McKinsey Global Institute projects it will add the equivalent of $2.6 trillion to $4.4 trillion to the global economy. Chatbots are only one part of that frenzy, which also includes technology that can generate new images, video, music and computer code. Nearly all of the tools include some language component.

Google is already product to news organizations, for which accuracy is paramount. The Associated Press is also exploring use of the technology as part of , which is paying to use part of AP害羞草研究所檚 text archive to improve its AI systems.

In partnership with India害羞草研究所檚 hotel management institutes, computer scientist Ganesh Bagler has been working for years to get AI systems, including a precursor, to invent recipes for South Asian cuisines, such as novel versions of rice-based biryani. A single 害羞草研究所渉allucinated害羞草研究所 ingredient could be the difference between a tasty and inedible meal.

When , visited India in June, the professor at the Indraprastha Institute of Information Technology Delhi had some pointed questions.

害羞草研究所淚 guess hallucinations in ChatGPT are still acceptable, but when a recipe comes out hallucinating, it becomes a serious problem,害羞草研究所 Bagler said, standing up in a crowded campus auditorium to address Altman on the New Delhi stop of the U.S. tech executive害羞草研究所檚 .

害羞草研究所淲hat害羞草研究所檚 your take on it?害羞草研究所 Bagler eventually asked.

Altman expressed optimism, if not an outright commitment.

害羞草研究所淚 think we will get the hallucination problem to a much, much better place,害羞草研究所 Altman said. 害羞草研究所淚 think it will take us a year and a half, two years. Something like that. But at that point we won害羞草研究所檛 still talk about these. There害羞草研究所檚 a balance between creativity and perfect accuracy, and the model will need to learn when you want one or the other.害羞草研究所

But for some experts who have studied the technology, such as University of Washington linguist Bender, those improvements won害羞草研究所檛 be enough.

Bender describes a language model as a system for 害羞草研究所渕odeling the likelihood of different strings of word forms,害羞草研究所 given some written data it害羞草研究所檚 been trained upon.

It害羞草研究所檚 how spell checkers are able to detect when you害羞草研究所檝e typed the wrong word. It also helps power automatic translation and transcription services, 害羞草研究所渟moothing the output to look more like typical text in the target language,害羞草研究所 Bender said. Many people rely on a version of this technology whenever they use the 害羞草研究所渁utocomplete害羞草研究所 feature when composing text messages or emails.

The latest crop of chatbots such as ChatGPT, Claude 2 or try to take that to the next level, by generating entire new passages of text, but Bender said they害羞草研究所檙e still just repeatedly selecting the most plausible next word in a string.

When used to generate text, language models 害羞草研究所渁re designed to make things up. That害羞草研究所檚 all they do,害羞草研究所 Bender said. They are good at mimicking forms of writing, such as legal contracts, or sonnets.

害羞草研究所淏ut since they only ever make things up, when the text they have extruded happens to be interpretable as something we deem correct, that is by chance,害羞草研究所 Bender said. 害羞草研究所淓ven if they can be tuned to be right more of the time, they will still have failure modes 害羞草研究所 and likely the failures will be in the cases where it害羞草研究所檚 harder for a person reading the text to notice, because they are more obscure.害羞草研究所

Those errors are not a huge problem for the marketing firms that have been turning to Jasper AI for help writing pitches, said the company害羞草研究所檚 president, Shane Orlick.

害羞草研究所淗allucinations are actually an added bonus,害羞草研究所 Orlick said. 害羞草研究所淲e have customers all the time that tell us how it came up with ideas 害羞草研究所 how Jasper created takes on stories or angles that they would have never thought of themselves.害羞草研究所

The Texas-based startup works with partners like OpenAI, Anthropic, Google or Facebook parent Meta to offer its customers a smorgasbord of AI language models tailored to their needs. For someone concerned about accuracy, it might offer up Anthropic害羞草研究所檚 model, while someone concerned with the security of their proprietary source data might get a different model, Orlick said.

Orlick said he knows hallucinations won害羞草研究所檛 be easily fixed. He害羞草研究所檚 counting on companies like Google, which he says must have a 害羞草研究所渞eally high standard of factual content害羞草研究所 for its search engine, to and resources into solutions.

害羞草研究所淚 think they have to fix this problem,害羞草研究所 Orlick said. 害羞草研究所淭hey害羞草研究所檝e got to address this. So I don害羞草研究所檛 know if it害羞草研究所檚 ever going to be perfect, but it害羞草研究所檒l probably just continue to get better and better over time.害羞草研究所

Techno-optimists, including Microsoft co-founder Bill Gates, have been forecasting a rosy outlook.

害羞草研究所淚害羞草研究所檓 optimistic that, over time, AI models can be taught to distinguish fact from fiction,害羞草研究所 Gates said in a July blog post detailing his thoughts on AI害羞草研究所檚 societal risks.

He cited a 2022 paper from OpenAI as an example of 害羞草研究所減romising work on this front.害羞草研究所

But even Altman, as he markets the products for a variety of uses, doesn害羞草研究所檛 count on the models to be truthful when he害羞草研究所檚 looking for information for himself.

害羞草研究所淚 probably trust the answers that come out of ChatGPT the least of anybody on Earth,害羞草研究所 Altman told the crowd at Bagler害羞草研究所檚 university, to laughter.

READ ALSO:

READ ALSO:





(or

害羞草研究所

) document.head.appendChild(flippScript); window.flippxp = window.flippxp || {run: []}; window.flippxp.run.push(function() { window.flippxp.registerSlot("#flipp-ux-slot-ssdaw212", "Black Press Media Standard", 1281409, [312035]); }); }