Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/184157
Title: Addressing overfitting in graph convolutional network using data augmentation
Authors: Gupta, Atharv
Keywords: Computer and Information Science
Issue Date: 2025
Publisher: Nanyang Technological University
Source: Gupta, A. (2025). Addressing overfitting in graph convolutional network using data augmentation. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/184157
Project: CCDS24-0456
Abstract: This project investigates the application of Graph Convolutional Networks (GCNs) to functional brain networks for subject classification, with a focus on addressing model overfitting through hyperparameter optimization and data augmentation. Experiments were conducted on five publicly available neuroimaging datasets: TaoWu, Neurocon, PPMI, ABIDE, and ADNI using fully connected graphs derived from fMRI connectivity matrices. The study proceeded in two phases. The first phase utilized unweighted graphs to establish a baseline and evaluate traditional regularization techniques, including dropout and DropEdge. The second phase incorporated weighted graphs to better capture connectivity strength, and introduced two edge-aware augmentation strategies: Gaussian edge weight perturbation and edge weight scaling. Hyperparameter optimization consistently improved test accuracy and reduced performance variability across datasets. However, dropout and DropEdge failed to mitigate overfitting, and edge scaling yielded diminishing results across all datasets. Gaussian perturbation led to modest improvements in larger datasets but negatively affected smaller ones. Performance remained below that of domain-informed, semisupervised models from prior work, underscoring the limitations of imaging-only supervised learning. Across both stages, hyperparameter tuning was the most reliable source of performance gains. Moreover, persistent overfitting and plateauing validation performance pointed to limitations in the augmentation strategies and training dynamics. The findings highlight the importance of tuning, the sensitivity of GCNs to augmentation design, and the need for domain-aware modelling. Recommendations are made for future work, including the integration of multimodal data, augmentation parameter tuning, non-stochastic augmentation techniques, and improving early stopping strategies.
URI: https://hdl.handle.net/10356/184157
Schools: College of Computing and Data Science 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
FYP_Final.pdf
  Restricted Access
2.12 MBAdobe PDFView/Open

Page view(s)

16
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.