Robust Federated Learning: The Case of Affine Distribution Shifts

作者: Ramtin Pedarsani , Ali Jadbabaie , Farzan Farnia , Amirhossein Reisizadeh

DOI:

关键词:

摘要: Federated learning is a distributed paradigm that aims at training models using samples across multiple users in network while keeping the on users' devices with aim of efficiency and protecting privacy. In such settings, data often statistically heterogeneous manifests various distribution shifts users, which degrades performance learnt model. The primary goal this paper to develop robust federated algorithm achieves satisfactory against samples. To achieve goal, we first consider structured affine shift captures device-dependent heterogeneity settings. This perturbation model applicable problems as image classification where images undergo imperfections, e.g. different intensity, contrast, brightness. address propose Learning framework Robust Affine (FLRA) provably Wasserstein observed solve FLRA's minimax problem, fast efficient optimization method provide convergence guarantees via gradient Descent Ascent (GDA) method. We further prove generalization error bounds for classifier show proper from empirical true underlying distribution. perform several numerical experiments empirically support FLRA. an indeed suffices significantly decrease new test user, our proposed significant gain comparison standard adversarial methods.

参考文章(41)
Diederik P. Kingma, Jimmy Ba, Adam: A Method for Stochastic Optimization arXiv: Learning. ,(2014)
Christian Szegedy, Ian J. Goodfellow, Jonathon Shlens, Explaining and Harnessing Adversarial Examples arXiv: Machine Learning. ,(2014)
David A. McAllester, PAC-Bayesian model averaging conference on learning theory. pp. 164- 170 ,(1999) , 10.1145/307400.307435
M Sai Praneeth, Xudong Peng, Alice Li, Shahrzad Hosseini Vajargah, Going deeper with convolutions computer vision and pattern recognition. pp. 1- 9 ,(2015) , 10.1109/CVPR.2015.7298594
Wolfram Wiesemann, Daniel Kuhn, Melvyn Sim, Distributionally Robust Convex Optimization Operations Research. ,vol. 62, pp. 1358- 1376 ,(2014) , 10.1287/OPRE.2014.1314
Ilya Sutskever, Geoffrey E. Hinton, Alex Krizhevsky, ImageNet Classification with Deep Convolutional Neural Networks neural information processing systems. ,vol. 25, pp. 1097- 1105 ,(2012)
Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun, Deep Residual Learning for Image Recognition computer vision and pattern recognition. pp. 770- 778 ,(2016) , 10.1109/CVPR.2016.90
Ilya Sutskever, Ian J. Goodfellow, Gregory S. Corrado, Michael Isard, Matthieu Devin, Vincent Vanhoucke, Martin Wicke, Manjunath Kudlur, Rajat Monga, Vijay Vasudevan, Geoffrey Irving, Yangqing Jia, Fernanda B. Viégas, Kunal Talwar, Martin Wattenberg, Ashish Agarwal, Martín Abadi, Yuan Yu, Rafal Józefowicz, Craig Citro, Sherry Moore, Paul Barham, Benoit Steiner, Pete Warden, Josh Levenberg, Derek Gordon Murray, Paul A. Tucker, Jonathon Shlens, Jeffrey Dean, Xiaoqiang Zheng, Chris Olah, Andy Davis, Dan Mané, Mike Schuster, Sanjay Ghemawat, Andrew Harp, Oriol Vinyals, Eugene Brevdo, Zhifeng Chen, Lukasz Kaiser, TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems arXiv: Distributed, Parallel, and Cluster Computing. ,(2015)
Jakub Konečný, Ananda Theertha Suresh, Dave Bacon, Felix X. Yu, Peter Richtarik, H. Brendan McMahan, Federated Learning: Strategies for Improving Communication Efficiency arXiv: Learning. ,(2016)
Blaise Aguera y Arcas, Daniel Ramage, H. Brendan McMahan, Seth Hampson, Eider Moore, Communication-Efficient Learning of Deep Networks from Decentralized Data international conference on artificial intelligence and statistics. pp. 1273- 1282 ,(2017)