Large Language Models (LLMs) have revolutionized the field of natural language processing, enabling sophisticated text generation, comprehension, and interaction capabilities. However, an open letter titled "Pause Giant AI Experiments" recently sparked global attention, calling for a temporary halt on the development of AI systems more powerful than GPT-4. This underscores growing concerns about the rapid development of advanced AI systems and their potential risks.
Among these risks, one particularly pressing issue is the phenomenon of ‘Hallucinations’ in LLMs. Hallucinations occur when these AI systems generate inaccurate, misleading, or completely fabricated information, despite presenting it with high confidence. Let us explore the various forms of Hallucinations, and the consequences false AI-generated information can have across industries.
In 2023, the Cambridge Dictionary named "Hallucinate" its Word of the Year, emphasizing the term's evolving significance in the context of Artificial Intelligence. This recognition reflects the growing awareness of AI's capabilities and its limitations, particularly the challenges posed by Hallucinations.
Unlike traditional errors that stem from incorrect data or algorithmic flaws, Hallucinations in LLMs resemble creative misinterpretations of the input. Essentially, the LLM produces output that is entirely fictional or unrelated to the provided context. These hallucinations can take various forms, such as:
According to “A Survey on Hallucination in Large Language Models” research paper, there are three types of LLM Hallucinations.
For input-conflicting hallucination, the LLM makes a mistake in the person name (Hill⇒Lucas) during summarizing.
For the context-conflicting hallucination, the LLM discusses Silver in the early stage, who was later referred to as Stern, resulting in a contradiction.
For the fact-conflicting hallucination, LLMs said the mother of Afonso II was Queen Urraca of Castile, while the correct answer is Dulce Berenguer of Barcelona.
Hallucinations of LLMs can have serious consequences, such as the spread of misinformation, data breaches and security concerns for real-world applications. For example, a hallucinated report generated from patient information in the medical field can pose a serious risk to the patient. Such hallucinations ultimately also affect the general trust of users in this technology, which is why it is important to address this problem quickly.
A real-life example is the case of a professor at Texas A&M University who failed his entire student body after ChatGPT falsely claimed their papers were written by AI. This resulted in many students being denied their degrees, which not only jeopardised their academic careers, but also undermined confidence in the reliability of such technologies.
ChatGPT also made a false accusation of sexual harassment against George Washington University Law Professor Jonathan Turley. The AI model invented a non-existent Washington Post article and falsely accused Turley of harassing a female student during a class trip. Such incidents show how dangerous and misleading AI-generated content can be, and emphasise the need to establish stricter vetting mechanisms and ethical guidelines for the use of AI.
Hallucinations in large language models can sometimes be viewed as a beneficial feature, especially when creativity and diversity are desired. These scenarios illustrate how hallucinations can be utilized effectively:
Hallucinations enable language models to generate unique and original content. For instance, if you ask a model like ChatGPT to craft a fantasy story, you would want it to produce an entirely new plot with original characters, settings, and storylines, rather than replicating existing stories. This creativity stems from the model's ability to "hallucinate" by not strictly relying on its training data but instead generating imaginative, novel outputs.
Hallucinations can foster diversity when exploring ideas. For example, during brainstorming, you may want the model to deviate from existing concepts in its training data and offer fresh perspectives. This ability to derive possibilities beyond known ideas allows users to explore innovative solutions and alternatives.
Many language models include a "temperature" setting, which controls the randomness of the model's output. Higher temperature values result in more varied and creative responses, introducing more hallucinations, while lower values make the output more deterministic and grounded in the training data. By adjusting the temperature through APIs, users can fine-tune the balance between creativity and accuracy based on their specific needs.
As LLM technology evolves, collaboration among researchers, developers, and policymakers will be critical to ensure its responsible development and deployment. Emphasizing the reduction of hallucinations and enhancing the benefits of LLMs will be key to unlocking their full potential while managing the associated risks.
https://futureoflife.org/open-letter/pause-giant-ai-experiments
https://arxiv.org/pdf/2309.01219.pdf
https://www.rollingstone.com/culture/culture-features/texas-am-chatgpt-ai-professor-flunks-students-false-claims-1234736601
https://www.indiatoday.in/technology/news/story/chatgpt-falsely-accuses-us-law-professor-of-sexually-harassing-a-student-2357597-2023-04-09