Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/149869
Full metadata record
DC FieldValueLanguage
dc.contributor.authorPeng, Jiaweien_US
dc.date.accessioned2021-06-09T09:30:57Z-
dc.date.available2021-06-09T09:30:57Z-
dc.date.issued2021-
dc.identifier.citationPeng, J. (2021). Auxiliary network design for local learning in deep neural networks. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/149869en_US
dc.identifier.urihttps://hdl.handle.net/10356/149869-
dc.description.abstractThe training of deep neural networks utilizes the backpropagation algorithm which consists of the forward pass, backward pass and parameters update. The output of a certain layer is produced based on the output of its lower layers in a sequential manner, and the gradients can only flow back layer by layer. This forces the majority of the network to be idle during the training process and hence leads to inefficiency. This is recognized as forward, backward and update lockings. To break the lockings, various methods of decoupled learning have been investigated. Currently, these methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. To remove these limitations, in this Final Year Project, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. The proposed method splits a neural network into several modules and trains them synchronously on different workers. In particular, re-computation is adopted to solve the memory explosion problem. A weight prediction scheme is proposed to deal with the weight delay caused by re-computation. To execute weight prediction, several weight predictors are proposed. A batch compensation scheme is also explored which allows the proposed DTRP to run faster. Experiments are conducted on various Convolutional Neural Networks to perform image classification task, which shows comparable or better results against the state-of-art methods and the backpropagation. The experiments also reveal that the memory explosion problem is effectively solved, and a significant acceleration is achieved. Moreover, experiments show that the proposed DTRP can be applied to train very wide networks as well as extremely deep networks.en_US
dc.language.isoenen_US
dc.publisherNanyang Technological Universityen_US
dc.relationA3135-201en_US
dc.subjectEngineering::Electrical and electronic engineeringen_US
dc.titleAuxiliary network design for local learning in deep neural networksen_US
dc.typeFinal Year Project (FYP)en_US
dc.contributor.supervisorLin Zhipingen_US
dc.contributor.schoolSchool of Electrical and Electronic Engineeringen_US
dc.description.degreeBachelor of Engineering (Electrical and Electronic Engineering)en_US
dc.contributor.supervisoremailEZPLin@ntu.edu.sgen_US
item.grantfulltextrestricted-
item.fulltextWith Fulltext-
Appears in Collections:EEE Student Reports (FYP/IA/PA/PI)
Files in This Item:
File Description SizeFormat 
FYP_Report_Peng Jiawei_final_v2.pdf
  Restricted Access
1.62 MBAdobe PDFView/Open

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.