Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/183825
Title: | LLM hallucination study | Authors: | Potdar, Prateek Anish | Keywords: | Computer and Information Science | Issue Date: | 2025 | Publisher: | Nanyang Technological University | Source: | Potdar, P. A. (2025). LLM hallucination study. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/183825 | Project: | CCDS24-0767 | Abstract: | Large Language Models (LLMs) exhibit impressive generative capabilities but often produce hallucinations—outputs that are factually incorrect, misleading, or entirely fabricated. These hallucinations pose significant challenges in high-stakes applications such as medical diagnosis, legal reasoning, financial analysis, and scientific research, where factual accuracy is critical. As LLMs become increasingly integrated into real-world systems, mitigating hallucinations is essential to ensuring reliable, trustworthy, and ethically sound AI deployments. Without effective strategies to reduce hallucinations, AI-generated content risks contributing to misinformation, undermining user trust, and limiting the adoption of LLMs in professional domains. My report investigates techniques to reduce hallucinations through systematic experimentation on Meta’s LLaMa model, a state-of-the-art open-source LLM. Specifically, I explored the impact of key generative parameters, including temperature scaling, top-k sampling, and retrieval- augmented generation, on factuality and coherence. These parameters play a crucial role in balancing response creativity and accuracy, directly influencing the probability of hallucinated content. By carefully tuning these hyperparameters and integrating external knowledge retrieval, I aimed to assess how different configurations affect the reliability of LLaMa’s generated responses. I systematically evaluated the effectiveness of these mitigation techniques using factuality scoring and response coherence analysis. Factuality was assessed by measuring the alignment of generated responses with authoritative sources, while coherence analysis examined the logical consistency and contextual appropriateness of outputs. Results from these experiments provide quantitative insights into the trade-offs between factual reliability, creativity, and response variability, offering practical guidelines for optimizing LLMs across different use cases. | URI: | https://hdl.handle.net/10356/183825 | Schools: | College of Computing and Data Science | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | CCDS Student Reports (FYP/IA/PA/PI) |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
FYP - LLM Hallucination Study - Prateek Potdar.pdf Restricted Access | CCDS24-0767: LLM Hallucination Study | 743.41 kB | Adobe PDF | View/Open |
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.