This chapter offers an expansive view of the hallucination phenomenon in large language models (LLMs), aiming to provide readers with a well-rounded understanding of both its complexity and implications. Hallucinations—instances where LLMs generate plausible yet factually incorrect or fabricated responses—pose a significant challenge in harnessing the true potential of these systems. Given the multitude of manifestations and underlying causes, detailing this phenomenon is inherently complex. Therefore, the chapter begins by exploring various definitions of hallucinations, enriched with concrete examples that illustrate the range of behaviors observed in these models. Following this foundational discussion, the analysis delves into the potential causes behind hallucinations. It examines factors such as the quality of training data, limitations in model architecture, and the inherent uncertainties of learning from finite datasets. By understanding these underlying issues, the chapter sheds light on why hallucinations occur and how they can undermine the reliability of LLM outputs. In addition to exploring causes, the chapter reviews a spectrum of mitigation strategies currently in development. It highlights methods designed to minimize the occurrence of hallucinations, thereby improving the overall trustworthiness of LLMs. A detailed look at state-of-the-art detection techniques further enriches this discussion, offering insights into how these tools can identify and flag erroneous outputs. This comprehensive overview not only captures the current landscape of research but also equips readers with the necessary knowledge to evaluate and implement responsible AI practices. Ultimately, this chapter seeks to balance the recognition of LLMs’ creative and transformative potential with a realistic appraisal of their limitations. By presenting a complete overview of the definitions, causes, and mitigation methods associated with hallucinations, it encourages a thoughtful and informed approach to using LLMs responsibly. The insights provided here aim to empower practitioners and decision-makers with the understanding needed to assess and improve the trustworthiness of these powerful tools in an era of rapid technological advancement.
Trustworthiness of large language models: hallucinations
Brunello, Nicolò
2026-01-01
Abstract
This chapter offers an expansive view of the hallucination phenomenon in large language models (LLMs), aiming to provide readers with a well-rounded understanding of both its complexity and implications. Hallucinations—instances where LLMs generate plausible yet factually incorrect or fabricated responses—pose a significant challenge in harnessing the true potential of these systems. Given the multitude of manifestations and underlying causes, detailing this phenomenon is inherently complex. Therefore, the chapter begins by exploring various definitions of hallucinations, enriched with concrete examples that illustrate the range of behaviors observed in these models. Following this foundational discussion, the analysis delves into the potential causes behind hallucinations. It examines factors such as the quality of training data, limitations in model architecture, and the inherent uncertainties of learning from finite datasets. By understanding these underlying issues, the chapter sheds light on why hallucinations occur and how they can undermine the reliability of LLM outputs. In addition to exploring causes, the chapter reviews a spectrum of mitigation strategies currently in development. It highlights methods designed to minimize the occurrence of hallucinations, thereby improving the overall trustworthiness of LLMs. A detailed look at state-of-the-art detection techniques further enriches this discussion, offering insights into how these tools can identify and flag erroneous outputs. This comprehensive overview not only captures the current landscape of research but also equips readers with the necessary knowledge to evaluate and implement responsible AI practices. Ultimately, this chapter seeks to balance the recognition of LLMs’ creative and transformative potential with a realistic appraisal of their limitations. By presenting a complete overview of the definitions, causes, and mitigation methods associated with hallucinations, it encourages a thoughtful and informed approach to using LLMs responsibly. The insights provided here aim to empower practitioners and decision-makers with the understanding needed to assess and improve the trustworthiness of these powerful tools in an era of rapid technological advancement.| File | Dimensione | Formato | |
|---|---|---|---|
|
hallucination_chapter_overleaf.pdf
accesso aperto
Descrizione: Versione full text del capitolo pubblicato
:
Pre-Print (o Pre-Refereeing)
Dimensione
198.55 kB
Formato
Adobe PDF
|
198.55 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


