Large Language Models (LLMs) show impressive performance in many natural language processing (NLP) tasks, including code generation and question answering. However, they suffer from various limitations, such as the generation of hallucinated content and reliance on outdated internal knowledge. To address these challenges, Retrieval-Augmented Large Language Models (RA-LLMs) have emerged as a promising solution, enhancing response accuracy by dynamically incorporating external knowledge through retrieval-augmented generation (RAG). Despite the advantages of RA-LLMs, the implementation of effective retrieval-augmented pipelines remains a complex task, since various techniques exist for document chunking, text similarity evaluation, and model selection, each influencing the overall system performance. In this work, we propose a general methodology for designing a RA-LLM pipeline, outlining key approaches for each phase and evaluating the effectiveness of different configurations. We also perform an experimental evaluation of the pipeline’s performance using real-world data, analyzing the impact of different techniques at each stage. The findings of this study offer insights into optimizing RA-LLM architectures for enhanced information retrieval and response generation.

Question Answering through Retrieval-Augmented Large Language Models: an Experimental Evaluation / Diamantini, Claudia; Mircoli, Alex; Pagnotta, Alessia; Potena, Domenico; Recchioni, Maria Cristina; Storti, Emanuele. - 4152:(2025). ( ITADATA 2025 Italian Conference on Big Data and Data Science 2025 Turin 9-11 September 2025).

Question Answering through Retrieval-Augmented Large Language Models: an Experimental Evaluation

Claudia Diamantini;Alex Mircoli;Alessia Pagnotta;Domenico Potena;Maria Cristina Recchioni;Emanuele Storti
2025-01-01

Abstract

Large Language Models (LLMs) show impressive performance in many natural language processing (NLP) tasks, including code generation and question answering. However, they suffer from various limitations, such as the generation of hallucinated content and reliance on outdated internal knowledge. To address these challenges, Retrieval-Augmented Large Language Models (RA-LLMs) have emerged as a promising solution, enhancing response accuracy by dynamically incorporating external knowledge through retrieval-augmented generation (RAG). Despite the advantages of RA-LLMs, the implementation of effective retrieval-augmented pipelines remains a complex task, since various techniques exist for document chunking, text similarity evaluation, and model selection, each influencing the overall system performance. In this work, we propose a general methodology for designing a RA-LLM pipeline, outlining key approaches for each phase and evaluating the effectiveness of different configurations. We also perform an experimental evaluation of the pipeline’s performance using real-world data, analyzing the impact of different techniques at each stage. The findings of this study offer insights into optimizing RA-LLM architectures for enhanced information retrieval and response generation.
2025
File in questo prodotto:
File Dimensione Formato  
Diamantini_Question-Answering-through-Retrieval_2025.pdf

accesso aperto

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza d'uso: Creative commons
Dimensione 232.73 kB
Formato Adobe PDF
232.73 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11566/352073
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact