Max-Margin Deep Generative Models for (Semi-)Supervised Learning

作者: Jun Zhu , Bo Zhang , Chongxuan Li

DOI:

关键词:

摘要: Deep generative models (DGMs) are effective on learning multilayered representations of complex data and performing inference input by exploring the ability. However, it is relatively insufficient to empower discriminative ability DGMs making accurate predictions. This paper presents max-margin deep (mmDGMs) a class-conditional variant (mmDCGMs), which explore strongly principle improve predictive performance in both supervised semi-supervised learning, while retaining capability. In we use predictions classifier as missing labels instead full posterior for efficiency; also introduce additional label-balance regularization terms unlabeled effectiveness. We develop an efficient doubly stochastic subgradient algorithm piecewise linear objectives different settings. Empirical results various datasets demonstrate that: (1) can significantly prediction meanwhile retain ability; (2) mmDGMs competitive best fully networks when employing convolutional neural recognition models; (3) mmDCGMs perform achieve state-of-the-art classification several benchmarks.

参考文章(53)
Geoffrey E. Hinton, Ruslan Salakhutdinov, Deep Boltzmann machines international conference on artificial intelligence and statistics. pp. 448- 455 ,(2009)
Arthur Szlam, Emily Denton, Rob Fergus, Soumith Chintala, Deep generative image models using a Laplacian pyramid of adversarial networks neural information processing systems. ,vol. 28, pp. 1486- 1494 ,(2015)
Shin Ishii, Shin-ichi Maeda, Ken Nakae, Takeru Miyato, Masanori Koyama, Distributional Smoothing with Virtual Adversarial Training arXiv: Machine Learning. ,(2015)
Gintare Karolina Dziugaite, Daniel M Roy, Zoubin Ghahramani, None, Training generative neural networks via maximum mean discrepancy optimization uncertainty in artificial intelligence. pp. 258- 267 ,(2015)
Jun Zhu, Bo Zhang, Chao Du, Learning Deep Generative Models with Doubly Stochastic MCMC arXiv: Learning. ,(2015)
Yichuan Tang, Deep Learning using Linear Support Vector Machines arXiv: Learning. ,(2013)
Frédéric Bastien, Pascal Lamblin, Razvan Pascanu, James Bergstra, Ian Goodfellow, Arnaud Bergeron, Nicolas Bouchard, David Warde-Farley, Yoshua Bengio, None, Theano: new features and speed improvements arXiv: Symbolic Computation. ,(2012)
Ruslan Salakhutdinov, Roger Grosse, Yuri Burda, Importance Weighted Autoencoders arXiv: Learning. ,(2015)
Ivo Danihelka, Daan Wierstra, Alex Graves, Danilo Jimenez Rezende, Karol Gregor, DRAW: A Recurrent Neural Network For Image Generation arXiv: Computer Vision and Pattern Recognition. ,(2015)
Alexey Dosovitskiy, Jost Tobias Springenberg, Thomas Brox, None, Learning to generate chairs with convolutional neural networks computer vision and pattern recognition. pp. 1538- 1546 ,(2015) , 10.1109/CVPR.2015.7298761