Please use this identifier to cite or link to this item:
Title: Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
Authors: Kow, Li Ren
Keywords: DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Issue Date: 2018
Abstract: Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, it involves a lot of floating point matrix calculation and it will be time consuming training on a computer processing unit (CPU). Even graphic processing unit (GPU) can do better in floating point calculation but it still takes long time to complete the training if the dataset is large and models are deep. Hence, multiple GPU card could be used in parallel to accelerate the entire training process. It is important to understand how fast it can be with different kind of deep learning framework which include (Mxnet, Pytorch and Caffe2) and the key software and hardware factor in this parallel training process on a single node or multi node configuration.
Schools: School of Electrical and Electronic Engineering 
Organisations: NVIDIA
Rights: Nanyang Technological University
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:EEE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
FYP Report (Kow Li Ren).pdf
  Restricted Access
3.42 MBAdobe PDFView/Open

Page view(s)

Updated on Sep 23, 2023

Download(s) 50

Updated on Sep 23, 2023

Google ScholarTM


Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.