规范化(社会学)
计算机科学
特征(语言学)
原始数据
联合学习
人工神经网络
深层神经网络
趋同(经济学)
人工智能
独立同分布随机变量
编码(集合论)
深度学习
GSM演进的增强数据速率
机器学习
数据挖掘
数学
社会学
哲学
经济
集合(抽象数据类型)
程序设计语言
随机变量
统计
经济增长
语言学
人类学
作者
Xiaoxiao Li,Meirui Jiang,Xiaofei Zhang,Michael Kamp,Qi Dou
出处
期刊:Cornell University - arXiv
日期:2021-01-01
被引量:19
标识
DOI:10.48550/arxiv.2102.07623
摘要
The emerging paradigm of federated learning (FL) strives to enable collaborative training of deep models on the network edge without centrally aggregating raw data and hence improving data privacy. In most cases, the assumption of independent and identically distributed samples across local clients does not hold for federated learning setups. Under this setting, neural network training performance may vary significantly according to the data distribution and even hurt training convergence. Most of the previous work has focused on a difference in the distribution of labels or client shifts. Unlike those settings, we address an important problem of FL, e.g., different scanners/sensors in medical imaging, different scenery distribution in autonomous driving (highway vs. city), where local clients store examples with different distributions compared to other clients, which we denote as feature shift non-iid. In this work, we propose an effective method that uses local batch normalization to alleviate the feature shift before averaging models. The resulting scheme, called FedBN, outperforms both classical FedAvg, as well as the state-of-the-art for non-iid data (FedProx) on our extensive experiments. These empirical results are supported by a convergence analysis that shows in a simplified setting that FedBN has a faster convergence rate than FedAvg. Code is available at https://github.com/med-air/FedBN.
科研通智能强力驱动
Strongly Powered by AbleSci AI