Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/183825
Title: LLM hallucination study
Authors: Potdar, Prateek Anish
Keywords: Computer and Information Science
Issue Date: 2025
Publisher: Nanyang Technological University
Source: Potdar, P. A. (2025). LLM hallucination study. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/183825
Project: CCDS24-0767
Abstract: Large Language Models (LLMs) exhibit impressive generative capabilities but often produce hallucinations—outputs that are factually incorrect, misleading, or entirely fabricated. These hallucinations pose significant challenges in high-stakes applications such as medical diagnosis, legal reasoning, financial analysis, and scientific research, where factual accuracy is critical. As LLMs become increasingly integrated into real-world systems, mitigating hallucinations is essential to ensuring reliable, trustworthy, and ethically sound AI deployments. Without effective strategies to reduce hallucinations, AI-generated content risks contributing to misinformation, undermining user trust, and limiting the adoption of LLMs in professional domains. My report investigates techniques to reduce hallucinations through systematic experimentation on Meta’s LLaMa model, a state-of-the-art open-source LLM. Specifically, I explored the impact of key generative parameters, including temperature scaling, top-k sampling, and retrieval- augmented generation, on factuality and coherence. These parameters play a crucial role in balancing response creativity and accuracy, directly influencing the probability of hallucinated content. By carefully tuning these hyperparameters and integrating external knowledge retrieval, I aimed to assess how different configurations affect the reliability of LLaMa’s generated responses. I systematically evaluated the effectiveness of these mitigation techniques using factuality scoring and response coherence analysis. Factuality was assessed by measuring the alignment of generated responses with authoritative sources, while coherence analysis examined the logical consistency and contextual appropriateness of outputs. Results from these experiments provide quantitative insights into the trade-offs between factual reliability, creativity, and response variability, offering practical guidelines for optimizing LLMs across different use cases.
URI: https://hdl.handle.net/10356/183825
Schools: College of Computing and Data Science 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
FYP - LLM Hallucination Study - Prateek Potdar.pdf
  Restricted Access
CCDS24-0767: LLM Hallucination Study743.41 kBAdobe PDFView/Open

Page view(s)

40
Updated on May 7, 2025

Download(s)

4
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.