作者: Yue Wu , Yinpeng Chen , Lijuan Wang , Yuancheng Ye , Zicheng Liu
关键词: Linear model 、 Layer (object-oriented design) 、 Forgetting 、 Machine learning 、 Artificial intelligence 、 Scale (descriptive set theory) 、 Training set 、 Incremental learning 、 Face (geometry) 、 Deep learning 、 Missing data 、 Computer science
摘要: Modern machine learning suffers from \textit{catastrophic forgetting} when new classes incrementally. The performance dramatically degrades due to the missing data of old classes. Incremental methods have been proposed retain knowledge acquired classes, by using distilling and keeping a few exemplars However, these struggle \textbf{scale up large number classes}. We believe this is because combination two factors: (a) imbalance between (b) increasing visually similar Distinguishing an particularly challenging, training unbalanced. propose simple effective method address issue. found that last fully connected layer has strong bias towards can be corrected linear model. With parameters, our performs remarkably well on datasets: ImageNet (1000 classes) MS-Celeb-1M (10000 classes), outperforming state-of-the-art algorithms 11.1\% 13.2\% respectively.