Greedy Shapley Client Selection for Communication-Efficient Federated Learning
CoRR(2023)
摘要
The standard client selection algorithms for Federated Learning (FL) are
often unbiased and involve uniform random sampling of clients. This has been
proven sub-optimal for fast convergence under practical settings characterized
by significant heterogeneity in data distribution, computing, and communication
resources across clients. For applications having timing constraints due to
limited communication opportunities with the parameter server (PS), the client
selection strategy is critical to complete model training within the fixed
budget of communication rounds. To address this, we develop a biased client
selection strategy, GreedyFed, that identifies and greedily selects the most
contributing clients in each communication round. This method builds on a fast
approximation algorithm for the Shapley Value at the PS, making the computation
tractable for real-world applications with many clients. Compared to various
client selection strategies on several real-world datasets, GreedyFed
demonstrates fast and stable convergence with high accuracy under timing
constraints and when imposing a higher degree of heterogeneity in data
distribution, systems constraints, and privacy requirements.
更多查看译文
关键词
client selection,data heterogeneity,federated learning,Shapley value,timing constraints
AI 理解论文
溯源树
样例
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要