CoCoA: A General Framework for Communication-Efficient Distributed Optimization

作者: Virginia Smith , Simone Forte , Chenxin Ma , Martin Takáč , Michael I Jordan

DOI:

关键词:

摘要: The scale of modern datasets necessitates the development efficient distributed optimization methods for machine learning. We present a general-purpose framework computing environments, CoCoA, that has an communication scheme and is applicable to wide variety problems in learning signal processing. extend cover general non-strongly-convex regularizers, including L1-regularized like lasso, sparse logistic regression, elastic net regularization, show how earlier work can be derived as special case. provide convergence guarantees class convex regularized loss minimization objectives, leveraging novel approach handling regularizers non-smooth functions. resulting markedly improved performance over state-of-the-art methods, we illustrate with extensive set experiments on real datasets.

参考文章(12)
S. Sathiya Keerthi, S. Sundararajan, Dhruv Mahajan, A distributed block coordinate descent method for training $l_1$ regularized linear classifiers arXiv: Learning. ,(2014)
Carlos Guestrin, Tyler Johnson, Blitz: A Principled Meta-Algorithm for Scaling Sparse Optimization international conference on machine learning. pp. 1171- 1179 ,(2015)
Martin Takáč, Peter Richtárik, Distributed coordinate descent method for learning with big data Journal of Machine Learning Research. ,vol. 17, pp. 2657- 2681 ,(2016)
Zhaosong Lu, Lin Xiao, On the complexity analysis of randomized block-coordinate descent methods Mathematical Programming. ,vol. 152, pp. 615- 642 ,(2015) , 10.1007/S10107-014-0800-2
Po-Wei Wang, Chih-Jen Lin, Iteration complexity of feasible descent methods for convex optimization Journal of Machine Learning Research. ,vol. 15, pp. 1523- 1548 ,(2014)
Shenghuo Zhu, Rong Jin, Yuanqing Lin, Tianbao Yang, Analysis of Distributed Stochastic Dual Coordinate Ascent arXiv: Distributed, Parallel, and Cluster Computing. ,(2013)
Yu. Nesterov, Smooth minimization of non-smooth functions Mathematical Programming. ,vol. 103, pp. 127- 152 ,(2005) , 10.1007/S10107-004-0552-5
Guo-Xun Yuan, Chih-Jen Lin, Chia-Hua Ho, An improved GLMNET for L1-regularized logistic regression Journal of Machine Learning Research. ,vol. 13, pp. 1999- 2030 ,(2012)
Stephen Boyd, Lieven Vandenberghe, Convex Optimization Cambridge University Press. ,(2004)
Martin Takáč, Jakub Konečný, Chenxin Ma, Virginia Smith, Martin Jaggi, Peter Richtárik, Michael I. Jordan, Distributed Optimization with Arbitrary Local Solvers arXiv: Learning. ,(2015)