Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/74869
Title: | Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures | Authors: | Kow, Li Ren | Keywords: | DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence | Issue Date: | 2018 | Abstract: | Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, it involves a lot of floating point matrix calculation and it will be time consuming training on a computer processing unit (CPU). Even graphic processing unit (GPU) can do better in floating point calculation but it still takes long time to complete the training if the dataset is large and models are deep. Hence, multiple GPU card could be used in parallel to accelerate the entire training process. It is important to understand how fast it can be with different kind of deep learning framework which include (Mxnet, Pytorch and Caffe2) and the key software and hardware factor in this parallel training process on a single node or multi node configuration. | URI: | http://hdl.handle.net/10356/74869 | Schools: | School of Electrical and Electronic Engineering | Organisations: | NVIDIA | Rights: | Nanyang Technological University | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | EEE Student Reports (FYP/IA/PA/PI) |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
FYP Report (Kow Li Ren).pdf Restricted Access | 3.42 MB | Adobe PDF | View/Open |
Page view(s)
189
Updated on Sep 23, 2023
Download(s) 50
17
Updated on Sep 23, 2023
Google ScholarTM
Check
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.