Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/183953
Title: Optimising the optimisation: performance analysis of memory management techniques for AI/ML training on non-NUMA hardware
Authors: Budi Syahiddin
Keywords: Computer and Information Science
Issue Date: 2025
Publisher: Nanyang Technological University
Source: Budi Syahiddin (2025). Optimising the optimisation: performance analysis of memory management techniques for AI/ML training on non-NUMA hardware. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/183953
Abstract: This research explores the potential of out-of-band optimisation techniques to enhance the efficiency of AI/ML training on non-NUMA hardware. By evaluating the impact of alternative memory allocators and transparent huge pages (THP) on training time across diverse AI/ML workloads, we aim to provide practical insights into memory management strategies. Our study focuses on the effectiveness of these techniques in improving training efficiency without modifying model architectures or algorithms. Through benchmarking different allocator configurations and THP settings, we aim to quantify their contributions to training performance, with the goal of informing future AI/ML development efforts. Our results shows that these low-level memory management strategies can squeeze out additional performance improvements, making them a valuable complement to traditional in-band optimisation methods.
URI: https://hdl.handle.net/10356/183953
Schools: College of Computing and Data Science 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
budi_syahiddin_fyp.pdf
  Restricted Access
4.3 MBAdobe PDFView/Open

Page view(s)

21
Updated on May 7, 2025

Download(s)

2
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.