This chapter offers an expansive view of the hallucination phenomenon in large language models (LLMs), aiming to provide readers with a well-rounded understanding of both its complexity and implications. Hallucinations—instances where LLMs generate plausible yet factually incorrect or fabricated responses—pose a significant challenge in harnessing the true potential of these systems. Given the multitude of manifestations and underlying causes, detailing this phenomenon is inherently complex. Therefore, the chapter begins by exploring various definitions of hallucinations, enriched with concrete examples that illustrate the range of behaviors observed in these models. Following this foundational discussion, the analysis delves into the potential causes behind hallucinations. It examines factors such as the quality of training data, limitations in model architecture, and the inherent uncertainties of learning from finite datasets. By understanding these underlying issues, the chapter sheds light on why hallucinations occur and how they can undermine the reliability of LLM outputs. In addition to exploring causes, the chapter reviews a spectrum of mitigation strategies currently in development. It highlights methods designed to minimize the occurrence of hallucinations, thereby improving the overall trustworthiness of LLMs. A detailed look at state-of-the-art detection techniques further enriches this discussion, offering insights into how these tools can identify and flag erroneous outputs. This comprehensive overview not only captures the current landscape of research but also equips readers with the necessary knowledge to evaluate and implement responsible AI practices. Ultimately, this chapter seeks to balance the recognition of LLMs’ creative and transformative potential with a realistic appraisal of their limitations. By presenting a complete overview of the definitions, causes, and mitigation methods associated with hallucinations, it encourages a thoughtful and informed approach to using LLMs responsibly. The insights provided here aim to empower practitioners and decision-makers with the understanding needed to assess and improve the trustworthiness of these powerful tools in an era of rapid technological advancement.

Trustworthiness of large language models: hallucinations

Brunello, Nicolò
2026-01-01

Abstract

This chapter offers an expansive view of the hallucination phenomenon in large language models (LLMs), aiming to provide readers with a well-rounded understanding of both its complexity and implications. Hallucinations—instances where LLMs generate plausible yet factually incorrect or fabricated responses—pose a significant challenge in harnessing the true potential of these systems. Given the multitude of manifestations and underlying causes, detailing this phenomenon is inherently complex. Therefore, the chapter begins by exploring various definitions of hallucinations, enriched with concrete examples that illustrate the range of behaviors observed in these models. Following this foundational discussion, the analysis delves into the potential causes behind hallucinations. It examines factors such as the quality of training data, limitations in model architecture, and the inherent uncertainties of learning from finite datasets. By understanding these underlying issues, the chapter sheds light on why hallucinations occur and how they can undermine the reliability of LLM outputs. In addition to exploring causes, the chapter reviews a spectrum of mitigation strategies currently in development. It highlights methods designed to minimize the occurrence of hallucinations, thereby improving the overall trustworthiness of LLMs. A detailed look at state-of-the-art detection techniques further enriches this discussion, offering insights into how these tools can identify and flag erroneous outputs. This comprehensive overview not only captures the current landscape of research but also equips readers with the necessary knowledge to evaluate and implement responsible AI practices. Ultimately, this chapter seeks to balance the recognition of LLMs’ creative and transformative potential with a realistic appraisal of their limitations. By presenting a complete overview of the definitions, causes, and mitigation methods associated with hallucinations, it encourages a thoughtful and informed approach to using LLMs responsibly. The insights provided here aim to empower practitioners and decision-makers with the understanding needed to assess and improve the trustworthiness of these powerful tools in an era of rapid technological advancement.
2026
Challenges and Applications of Generative Large Language Models
9780443335921
factuality
faithfulness
Hallucinations
large language models
trustworthiness
File in questo prodotto:
File Dimensione Formato  
hallucination_chapter_overleaf.pdf

accesso aperto

Descrizione: Versione full text del capitolo pubblicato
: Pre-Print (o Pre-Refereeing)
Dimensione 198.55 kB
Formato Adobe PDF
198.55 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11311/1311546
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact