计算机科学
聚类分析
利用
人工智能
卷积神经网络
先验与后验
人口
机器学习
人气
信息隐私
协议(科学)
深度学习
特征学习
数据挖掘
计算机安全
社会心理学
替代医学
医学
人口学
社会学
病理
哲学
心理学
认识论
互联网隐私
作者
Felix Sattler,Klaus‐Robert Müller,Wojciech Samek
标识
DOI:10.1109/tnnls.2020.3015958
摘要
Federated learning (FL) is currently the most widely adopted framework for collaborative training of (deep) machine learning models under privacy constraints. Albeit its popularity, it has been observed that FL yields suboptimal results if the local clients' data distributions diverge. To address this issue, we present clustered FL (CFL), a novel federated multitask learning (FMTL) framework, which exploits geometric properties of the FL loss surface to group the client population into clusters with jointly trainable data distributions. In contrast to existing FMTL approaches, CFL does not require any modifications to the FL communication protocol to be made, is applicable to general nonconvex objectives (in particular, deep neural networks), does not require the number of clusters to be known a priori, and comes with strong mathematical guarantees on the clustering quality. CFL is flexible enough to handle client populations that vary over time and can be implemented in a privacy-preserving way. As clustering is only performed after FL has converged to a stationary point, CFL can be viewed as a postprocessing method that will always achieve greater or equal performance than conventional FL by allowing clients to arrive at more specialized models. We verify our theoretical analysis in experiments with deep convolutional and recurrent neural networks on commonly used FL data sets.
科研通智能强力驱动
Strongly Powered by AbleSci AI