Please use this identifier to cite or link to this item:
Title: Cost of federated learning
Authors: Tan, Cheen Hao
Keywords: Engineering::Computer science and engineering
Issue Date: 2022
Publisher: Nanyang Technological University
Source: Tan, C. H. (2022). Cost of federated learning. Final Year Project (FYP), Nanyang Technological University, Singapore.
Project: SCSE21-0580
Abstract: Federated Learning (FL) is term coined by Google in 2016 at a time when misuse of personal data was gaining global traction. The main goal was to build a machine learning (ML) model based on datasets that are distributed across the world, all whilst retraining personal privacy. FL enables edge devices such as mobile phone or any Internet of Things (IoT) devices to collaboratively learn a shared prediction model while keeping all its training data intact and privately held. Traditional machine learning is usually centrally trained with a full distinct view of the entire dataset. But this will usually cause privacy issue when personal data are being transferred to the central server for ML training. Thus, FL attempts to solve this by enabling edge devices such as mobile devices to collaboratively learn a shared prediction model with each other while keeping all its local data, private. This will however cause network problem in the edge device if not optimized properly when each client tries to independently computes an update to the current model based on its data and sends the update back to the server for computation. The sheer amount of data just to send the model back and forth to the server would be too huge for a low computing resource such as an IoT device. Hence, communication costs is of utmost importance when edge devices are being studied in this framework. No matter the amount of issues brought about from the emergence of FL framework, its core technology still outweighs many of shortcoming. One – Data privacy of every client is preserved throughout all the training. Two – It is way more efficient at generalization such as getting raw data from edge devices from all over world as opposed to just training on one client, repeatedly. Undeniably, FL would be the future of ML paradigm. In this report, discussion on FL research with respect to edge devices will be made. By means of comparison and propose ways that would help in the reduction of communication costs by a substantial amount through compression, quantization, with and without privacy preserving techniques. Implementation will be performed and demonstrated solely on Google’s TensorFlow.
Schools: School of Computer Science and Engineering 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:SCSE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
  Restricted Access

Page view(s) 50

Updated on Feb 24, 2024

Download(s) 50

Updated on Feb 24, 2024

Google ScholarTM


Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.