Multi-core structural SVM training

作者: Kai-Wei Chang , Vivek Srikumar , Dan Roth

DOI: 10.1007/978-3-642-40991-2_26

关键词:

摘要: Many problems in natural language processing and computer vision can be framed as structured prediction problems. Structural support vector machines (SVM) is a popular approach for training predictors, where learning an optimization problem. Most structural SVM solvers alternate between model update phase inference (which predicts structures all examples). As become more complex, becomes bottleneck thus slows down considerably. In this paper, we propose new algorithm SVMs called DEMIDCD that extends the dual coordinate descent by decoupling phases into different threads. We take advantage of multicore hardware to parallelize with minimal synchronization phases.We prove our not only converges but also fully utilizes available processors speed up learning, validate on two real-world NLP problems: part-of-speech tagging relation extraction. both cases, show achieves competitive performance. For example, it relative duality gap 1% POS problem 192 seconds using 16 threads, while standard implementation multi-threaded same number threads requires than 600 reach solution quality.

参考文章(25)
Vivek Srikumar, Dan Roth, Ming-wei Chang, Dan Goldwasser, Structured Output Learning with Indirect Supervision international conference on machine learning. pp. 199- 206 ,(2010)
Wen-tau Yih, Dan Roth, A Linear Programming Formulation for Global Inference in Natural Language Tasks conference on computational natural language learning. pp. 1- 8 ,(2004)
Thorsten Joachims, Making large scale SVM learning practical Technical reports. ,(1999) , 10.17877/DE290R-14262
Lise Getoor, Ben Taskar, Introduction to statistical relational learning MIT Press. ,(2007)
BSCH OLKOPF, C Burges, A Smola, Advances in kernel methods: support vector learning international conference on neural information processing. ,(1999) , 10.5555/299094
Mitch Marcus, Beatrice Santorini, Mary Ann Marcinkiewicz, None, Building a large annotated corpus of English: the penn treebank Computational Linguistics. ,vol. 19, pp. 313- 330 ,(1993) , 10.21236/ADA273556
Kai-Wei Chang, Dan Roth, Selective block minimization for faster convergence of limited memory large-scale linear models Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining - KDD '11. pp. 699- 707 ,(2011) , 10.1145/2020408.2020517
Thorsten Joachims, Thomas Finley, Chun-Nam John Yu, Cutting-plane training of structural SVMs Machine Learning. ,vol. 77, pp. 27- 59 ,(2009) , 10.1007/S10994-009-5108-8
Alekh Agarwal, John C. Duchi, Distributed delayed stochastic optimization 2012 IEEE 51st IEEE Conference on Decision and Control (CDC). pp. 5451- 5452 ,(2012) , 10.1109/CDC.2012.6426626