Asynch-SGBDT: Asynchronous Parallel Stochastic Gradient Boosting Decision Tree based on Parameters Server

作者: Zhang Yunquan , Cheng Daning , Li Shigang , Xia Fen

DOI:

关键词:

摘要: In AI research and industry, machine learning is the most widely used tool. One of important algorithms Gradient Boosting Decision Tree, i.e. GBDT whose training process needs considerable computational resources time. To shorten time, many works tried to apply on Parameter Server. However, those are synchronous parallel which fail make full use this paper, we examine possibility using asynchronous methods train model name algorithm as asynch-SGBDT (asynchronous stochastic gradient boosting decision tree). Our theoretical experimental results indicate that scalability influenced by sample diversity datasets, sampling rate, step length setting tree. Experimental also show reaches a linear speedup in manner when datasets trees meet high requirements.

参考文章(22)
Jeffrey Dean, Greg Corrado, Rajat Monga, Kai Chen, Matthieu Devin, Mark Mao, Marc'aurelio Ranzato, Andrew Senior, Paul Tucker, Ke Yang, Quoc Le, Andrew Ng, None, Large Scale Distributed Deep Networks neural information processing systems. ,vol. 25, pp. 1223- 1231 ,(2012)
Tianjun Xiao, Tianqi Chen, Chiyuan Zhang, Zheng Zhang, Yutian Li, Min Lin, Minjie Wang, Naiyan Wang, Mu Li, Bing Xu, MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous Distributed Systems arXiv: Distributed, Parallel, and Cluster Computing. ,(2015)
Martín Abadi, TensorFlow: learning functions at scale international conference on functional programming. ,vol. 51, pp. 1- 1 ,(2016) , 10.1145/2951913.2976746
Jie Jiang, Jiawei Jiang, Bin Cui, Ce Zhang, TencentBoost: A Gradient Boosting Tree System with Parameter Server 2017 IEEE 33rd International Conference on Data Engineering (ICDE). pp. 281- 284 ,(2017) , 10.1109/ICDE.2017.87
Tie-Yan Liu, Taifeng Wang, Thomas Finley, Weidong Ma, Qi Meng, Qiwei Ye, Guolin Ke, Wei Chen, LightGBM: a highly efficient gradient boosting decision tree neural information processing systems. ,vol. 30, pp. 3149- 3157 ,(2017)
Alexander Smola, John Langford, Martin Zinkevich, Slow Learners are Fast arXiv: Optimization and Control. ,(2009)
Benjamin Recht, Feng Niu, Christopher Re, Stephen J. Wright, HOGWILD!: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent arXiv: Optimization and Control. ,(2011)
R. H. Byrd, S. L. Hansen, Jorge Nocedal, Y. Singer, A Stochastic Quasi-Newton Method for Large-Scale Optimization Siam Journal on Optimization. ,vol. 26, pp. 1008- 1031 ,(2016) , 10.1137/140954362
Carlos Guestrin, Tianqi Chen, XGBoost: A Scalable Tree Boosting System arXiv: Learning. ,(2016) , 10.1145/2939672.2939785
Jie Zhou, Tong Zhang, Tong Zhang, Peng Sun, A Convergence Rate Analysis for LogitBoost, MART and Their Variant international conference on machine learning. pp. 1251- 1259 ,(2014)