Discriminative models and dimensionality reduction for regression

作者: Vladimir Pavlovic , Minyoung Kim

DOI: 10.7282/T3PV6KQN

关键词:

摘要: Many prediction problems that arise in computer vision and robotics can be formulated within a regression framework. Unlike traditional problems, tasks are often characterized by varying number of output variables with complex dependency structures. The further aggravated the high dimensionality input. In this thesis, I address two challenging related to learning regressors such settings: (1) developing discriminative approaches handle structured variables, (2) reducing input while preserving statistical correlation output. A structure effectively captured probabilistic graphical models. contrast joint data modeling for models, propose conditional models approach directly ultimate objective. While as Conditional Random Fields (CRFs) has attracted significant interest past, setting been rarely explored. work first extend CRF discriminatively trained HMM methods problem. different based on directed undirected second parameter is cast convex optimization problem, accompanied new effective handles density integrability constraint. Experiments several problem domains, including human motion robot-arm state estimation, indicate yield accuracy comparable or better than state-of-the-art approaches. In part consider task finding low-dimensional representation covariates regressing This task, known reduction (DRR), particularly useful when visualizing high-dimensional data, efficiently designing reduced dimension, eliminating noise uncovering essential information predicting common many machine tasks, their use settings not widespread. A recent DRR have proposed statistics community but suffer from limitations, non-convexity need slicing potentially space. these issues proposing novel covariance operators reproducing kernel Hilbert spaces (RKHSes) provide closed-form solution without explicit slicing. benefits demonstrated comprehensive set evaluations important pattern recognition.

参考文章(51)
D Povey, PC Woodland, Large scale discriminative training for speech recognition Proc. ITW ASR, ISCA, 2000. ,(2000)
Bernhard Scholkopf, Ralf Herbrich, Alex Smola, Robert Williamson, A Generalized Representer Theorem european conference on computational learning theory. pp. 416- 426 ,(2001) , 10.1007/3-540-44581-1_27
Andrew McCallum, Dayne Freitag, Fernando C. N. Pereira, Maximum Entropy Markov Models for Information Extraction and Segmentation international conference on machine learning. pp. 591- 598 ,(2000)
Jacob C. Engwerda, AndréC.M. Ran, Arie L. Rijkeboer, Necessary and sufficient conditions for the existence of a positive definite solution of the matrix equation X + A*X-1A = Q Linear Algebra and its Applications. ,vol. 186, pp. 255- 275 ,(1993) , 10.1016/0024-3795(93)90295-Y
David A. Ross, Simon Osindero, Richard S. Zemel, Combining discriminative features to infer complex trajectories Proceedings of the 23rd international conference on Machine learning - ICML '06. pp. 761- 768 ,(2006) , 10.1145/1143844.1143940
Joshua B Tenenbaum, Vin de Silva, John C Langford, A Global Geometric Framework for Nonlinear Dimensionality Reduction Science. ,vol. 290, pp. 2319- 2323 ,(2000) , 10.1126/SCIENCE.290.5500.2319
J. N. Darroch, D. Ratcliff, Generalized Iterative Scaling for Log-Linear Models Annals of Mathematical Statistics. ,vol. 43, pp. 1470- 1480 ,(1972) , 10.1214/AOMS/1177692379
John Lafferty, Xiaojin Zhu, Yan Liu, Kernel conditional random fields: representation and clique selection international conference on machine learning. pp. 64- ,(2004) , 10.1145/1015330.1015337
Charles R. Baker, Joint measures and cross-covariance operators Transactions of the American Mathematical Society. ,vol. 186, pp. 273- 289 ,(1973) , 10.1090/S0002-9947-1973-0336795-3
Michael I Jordan, Robert A Jacobs, None, Hierarchical mixtures of experts and the EM algorithm Neural Computation. ,vol. 6, pp. 181- 214 ,(1994) , 10.1162/NECO.1994.6.2.181