Differential Privacy Meets Federated Learning Under Communication Constraints

IEEE Internet of Things Journal(2022)

引用 17|浏览19
暂无评分
摘要
The performance of federated learning systems is bottlenecked by communication costs and training variance. The communication overhead problem is usually addressed by three communication-reduction techniques, namely, model compression, partial device participation, and periodic aggregation, at the cost of increased training variance. Different from traditional distributed learning systems, federated learning suffers from data heterogeneity (since the devices sample their data from possibly different distributions), which induces additional variance among devices during training. Various variance-reduced training algorithms have been introduced to combat the effects of data heterogeneity, while they usually cost additional communication resources to deliver necessary control information. Additionally, data privacy remains a critical issue in FL and, thus, there have been attempts at bringing Differential Privacy to this framework as a mediator between utility and privacy requirements. This article investigates the tradeoffs between communication costs and training variance under a resource-constrained federated system theoretically and experimentally, and studies how communication reduction techniques interplay in a differentially private setting. The results provide important insights into designing practical privacy-aware federated learning systems.
更多
查看译文
关键词
Artificial intelligence,communication constraints,differential privacy,federated learning,training variance
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要