Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/184069
Title: X-ray report generation
Authors: Ang, Kai Jun
Keywords: Computer and Information Science
Issue Date: 2025
Publisher: Nanyang Technological University
Source: Ang, K. J. (2025). X-ray report generation. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/184069
Project: CCDS24-0522
Abstract: Medical imaging and radiology reports are indispensable parts of the modern-day medicine, allowing doctors to diagnose and treat patients effectively. Nevertheless, the growing shortage of radiologists results in healthcare bottlenecks and increased workload for radiologists. Automated radiology report generation using machine learning can potentially overcome pain points in healthcare systems and improve the quality of healthcare services. Previous works on Vision Language Models have shown moderate success in radiology report generation, highlighting the potential for more improvement. In this work, efforts were made to improve the performance of LLaVA-Med and LLaVA for the task of free-text radiology report generation on single Chest X-ray (CXR) images by finetuning the Vision Encoder, Multi-Layer Perceptron (MLP) and the Large-Language Model (LLM) components of the model. Experimental results indicate that most models showed better performance compared to the pretrained models. The best model attained a CheXpert F1 score of 36.9%, a significant improvement from the baseline score of 5.3%.
URI: https://hdl.handle.net/10356/184069
Schools: College of Computing and Data Science 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
Ang_Kai_Jun_Amended_Final Report.pdf
  Restricted Access
1.71 MBAdobe PDFView/Open

Page view(s)

24
Updated on May 7, 2025

Download(s)

1
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.