Before the advent of Generative AI, hallucinations were the stuff of fever dreams or chemical substances. More recently, the term has come to refer to the fanciful responses that language models occasionally produce. Alternatively funny and alarming, AI hallucinations raise concerns about the potential risks enterprises can face from leveraging Generative AI. With this article we explain what AI hallucinations are and why they occur, and how enterprises can avoid them.
AI hallucinations are AI-generated statements that do not exist in the model’s training data. As a result, hallucinations can include falsehoods that the Generative AI presents as if they were true. AI hallucinations are particularly dangerous for the way they can blend invented information with facts, dates, and real people. The result can be inaccurate and misleading text that is nonetheless compelling and convincing.
For example, when The New York Times (NYT) asked a number of AI chatbots when the paper first reported on AI, each answered with a different hallucination that merged fact with invention. In addition, each cited purported NYT articles that did not exist. The reporters spotted the errors easily because they knew the real answer.
But, what if your business relied on a Generative AI solution to answer questions automatically without human oversight? What if you were relying on it to identify patterns or generate texts that human workers would not be able to fact check? What if using GenAI to create ad or email copy? In these cases, AI hallucinations could pose serious reputational and legal risks to your business.
Every enterprise function is actively exploring how to leverage Generative AI. McKinsey estimates the biggest economic and productivity impacts from the technology will be for marketing, sales, software development, and customer operations. Yet back office functions like HR, finance, legal, procurement, and others are also actively exploring Generative AI applications. The major risks from AI hallucinations span the different functions where the technology is being used, including:
If AI hallucinations include information that is false, misleading, or in any way inconsistent with the way you speak to your customer, it can erode your brand identity. By extension, they can damage the connection customers have with you.
Many factors go into cultivating a brand image capable of cultivating loyalty. The look and feel of your products and locations, how you deliver your services, and the tone and approach you take with your marketing all contribute to the feeling you want customers to have from working with you. Consistency is key. If your physical stores, branches, or outlets offer one experience and your digital channels another, it can unsettle your customers and make them look elsewhere.
Without strong controls and even human oversight, AI-generated messages could spread inaccurate information about your products or services. They could also make commitments that the brand cannot fulfill without significant expense. That’s one way AI hallucinations create risk for your brand.
Consider AI chatbots commonly used in chatbot marketing, one of the common applications of Generative AI. If those chatbots generate hallucinations at scale promising refunds or bonus products, it could cost a brand either significant loyalty (if it does not honor the promise) or significant revenue (if it does).
Another popular use case for AI is to automate common decisions to free up human oversight for more complex work. Effective automation depends on accurate information and consistent execution, however. If an algorithm prompts a Generative AI tool to summarize a series of customer interactions and the AI introduces hallucinations, any decisions made based on the summary would be misinformed.
The level of risk this presents varies greatly. Getting the decision about what product to offer a given customer in an email marketing campaign may be relatively low risk, given the average customer engagement rate. In contrast, a decision to erroneously close a financial account could lead to serious reputational damage. The same is true for hallucinations in systems used by your finance department or HR.
Law firms have shown high interest in Generative AI, as have the legal departments of enterprises. AI is also making the job of compliance officers easier. It does this by allowing them to scan and generate documents and monitor regulatory changes. With all this promise for improving legal and regulatory compliance, however, come risks. Accuracy is paramount for legal and compliance professionals. AI hallucinations threaten to introduce inaccuracies that are hard to identify and correct, given that they can be buried inside long pages of legalese. If these inaccuracies appear in financial statements of public companies, it could result in legal actions against the firm and its executive officers.
AI experts are not sure why Generative AI tends to hallucinate, though it is a widespread problem. Research on the most popular large language models (LLMs) found that all of the tested systems hallucinated. In a recent study, researchers prompted GPT-4 (OpenAI), Palmchat (Google), Claude 2 (Anthropic) and systems from Meta to summarize text. All of the AI-generated responses included hallucinations, though with varying levels of frequency. The lowest was 3 percent of the GPT-4-generated text up to 27 percent of the Palmchat-generated text. Researchers suspect that the frequency of hallucinations may be higher if the AI is asked to do something more complex and spontaneous than the simple text summarization task used in this study.
There are a number of technological and process approaches that can help reduce hallucinations. They include:
At Persado, we have been working for more than ten years to enhance our customers’ relationships using language. Our enterprise Generative AI is built on a proprietary language model that includes more than 10 years of marketing messages from Fortune 500 companies. Our focus on generating the marketing copy that is most likely to motivate customer action gives our text generation and machine learning technology a focused goal of improving your campaign results by optimizing for conversions. The Persado content intelligence team—composed of human writers, linguists, and data analysts—also works with the model to teach it what “good” language looks like and steer it away from inauthentic, off-brand, or inaccurate options.
In short, our team leverages a specific pool of high-quality data, coupled with a focused task and human oversight, to generate language that increases customer engagement above human-generated controls 96% of the time. The result is better results and higher customer engagement with no AI hallucination risk.
Learn more about Persado Motivation AI and explore getting your risk-free trial today.
The post Risks From AI Hallucinations and How to Avoid Them appeared first on Persado.
Jasper Research Lab’s new shadow generation research and model enable brands to create more photorealistic…
We’re announcing new updates to Gemini 2.0 Flash, plus introducing Gemini 2.0 Flash-Lite and Gemini…
Interactive digital agents (IDAs) leverage APIs of stateful digital environments to perform tasks in response…
This post is co-written with Martin Holste from Trellix. Security teams are dealing with an…
As AI continues to unlock new opportunities for business growth and societal benefits, we’re working…
An internal email obtained by WIRED shows that NOAA workers received orders to pause “ALL…