Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/179060
Title: GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
Authors: Liu, Zelei
Chen, Yuanyuan
Yu, Han
Liu, Yang
Cui, Lizhen
Keywords: Computer and Information Science
Issue Date: 2022
Source: Liu, Z., Chen, Y., Yu, H., Liu, Y. & Cui, L. (2022). GTG-shapley: efficient and accurate participant contribution evaluation in federated learning. ACM Transactions On Intelligent Systems and Technology, 13(4), 60-. https://dx.doi.org/10.1145/3501811
Project: AISG2-RP-2020-019
A20G8b0102)
NWJ-2020-008
NSC-2019-011
Journal: ACM Transactions on Intelligent Systems and Technology
Abstract: Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy. To sustain the long-term operation of an FL ecosystem, it is important to attract high-quality data owners with appropriate incentive schemes. As an important building block of such incentive schemes, it is essential to fairly evaluate participants’ contribution to the performance of the final FL model without exposing their private data. Shapley Value (SV)–based techniques have been widely adopted to provide a fair evaluation of FL participant contributions. However, existing approaches incur significant computation costs, making them difficult to apply in practice. In this article, we propose the Guided Truncation Gradient Shapley (GTG-Shapley) approach to address this challenge. It reconstructs FL models from gradient updates for SV calculation instead of repeatedly training with different combinations of FL participants. In addition, we design a guided Monte Carlo sampling approach combined with within-round and between-round truncation to further reduce the number of model reconstructions and evaluations required. This is accomplished through extensive experiments under diverse realistic data distribution settings. The results demonstrate that GTG-Shapley can closely approximate actual Shapley values while significantly increasing computational efficiency compared with the state-of-the-art, especially under non-i.i.d. settings.
URI: https://hdl.handle.net/10356/179060
ISSN: 2157-6904
DOI: 10.1145/3501811
Schools: College of Computing and Data Science 
School of Computer Science and Engineering 
Rights: © 2022 Association for Computing Machinery. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1145/3501811.
Fulltext Permission: open
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Journal Articles

Files in This Item:
File Description SizeFormat 
2109.02053v1.pdf517.73 kBAdobe PDFView/Open

SCOPUSTM   
Citations 5

88
Updated on Jan 10, 2025

Page view(s)

98
Updated on Mar 15, 2025

Download(s)

22
Updated on Mar 15, 2025

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.