Stories of wider successful implementations of AI tools are published nearly every day. With ChatGPT, Midjourney and other models now available to the broad public, a greater number of people are starting to rely on AI in their daily lives.
While it is evident that machine learning algorithms are able to solve more challenging requirements, they are not yet perfect. Frequent hallucinations of artificial intelligence make them not the most reliable substitute for humans. And while for an ordinary user an AI error is just a glitch to laugh at, for business processes, such unpredictability can lead to consequences - from loss of client trust to lawsuits.
Some countries have begun drafting regulations around AI models to provide a framework around usage and applicability. Let’s figure out why and how neural networks start to hallucinate and how this can be minimised.
Though sometimes it is not possible to identify the cause of an AI error, often, hallucinations result from how generative systems create text. When responding to a user's query, the AI suggests a likely set of words based on an array of previous data. The likelihood that some words follow others is not a very reliable way of making sure the final sentence is accurate. AI can piece together terms that may sound plausible but are not necessarily accurate—and to a human eye may look like complete nonsense. An example where I struggled with is asking ChatGPT for examples of countries having matching and non-matching settlement markets. While it was able to provide ‘Continuous Net Settlement’ (CNS) as an example of a matching settlement system, I was interested in the country the system was in (United States in this case) and the prompt provided the wrong output in this case.
At times however, detecting an AI hallucination can be more tricky. While some errors are obvious, others are more subtle and may go unnoticed, especially when the output is processed automatically or handled by a person with a limited expertise in the field. Undetected AI issues can lead to unforeseen and unwanted consequences. This is especially true in areas where it is critical to have accurate and reliable information. In addition, typically the more specialised a prompt, the accuracy of the AI model may vary due to the lack of supporting collateral it may refer to. The CNS example above is again a great example; I was unable to find a list of countries through a Google search and hoped ChatGPT could provide a consolidated list but faced a similar hurdle with the latter.
These are the common types of issues that occur due to AI hallucinations:
Unreliable analytics: If AI generates inaccurate data, it can lead to unreliable analytical results. Organisations may make decisions based on incorrect information, and the outcomes can be costly. Sometimes the data can be out of date; a great example is ChatGPT’s free version, which only carries data until 2022 and therefore numbers gleaned from it may be unreliable.
Ethical and legal concerns: Due to hallucinations, AI models can reveal sensitive information or generate offensive content, leading to legal issues. Walled gardens can mitigate some risks with sensitivity.
Misinformation: Generating false information can cause various problems to companies and end users such as breaking trust, harming or negatively affecting the public opinion.
AI hallucinations are a complex problem, and their causes are not fully clear to users and developers alike. Here are a few key factors that may cause or contribute to such hallucinations:
It’s important to remember that LLMs work like a “black box”—not even data scientists can fully follow the generation process and predict the output. This is why it’s not possible to 100% safeguard your business from AI hallucinations. At the moment, companies that use AI models need to focus on preventing, detecting and minimising AI hallucinations. These are some tips for maintaining the “hygiene” of the ML models:
To avoid these uncertainties and confusion from the start, it's important to plan the development of AI models by emphasising their interpretability and explainability. This means documenting your model building processes, maintaining transparency with key stakeholders, and choosing an architecture that makes it easy to interpret and explain model performance despite growing data volumes and user requirements. This will also assist in regulatory requirements, as the field gets scrutinised by governments.
If you don’t use the AI model to handle sensitive information, you can try to apply search augmented generation to reduce the risk of hallucinations. Instead of relying only on existing training data and context from the user, the AI will search for relevant information online. However, this technology hasn’t shown very reliable results yet. The output of the unfiltered search can sometimes be just as untrue as an AI model hallucination.
Are hallucinations always bad?
The relationship between hallucinations and creativity in AI systems seems similar to the processes in the human imagination. Humans often come up with creative ideas by letting their minds wander beyond reality.
The AI models that generate the most innovative and original results also tend to sometimes create content that is not based on real facts.Some experts believe that getting rid of hallucinations completely could harm creative content creation.
However, it is important to understand that such types of output often lack a factual basis and logical thought, making them unsuitable for fact-based tasks.