作者: Jorge Sanchez , Florent Perronnin
DOI: 10.1109/CVPR.2011.5995504
关键词: Pattern recognition 、 Handwriting recognition 、 Data compression 、 Kernel (image processing) 、 Curse of dimensionality 、 Lossy compression 、 Computer science 、 Hash function 、 Contextual image classification 、 Artificial intelligence 、 Dimensionality reduction
摘要: We address image classification on a large-scale, i.e. when large number of images and classes are involved. First, we study accuracy as function the signature dimensionality training set size. show experimentally that larger set, higher impact accuracy. In other words, high-dimensional signatures important to obtain state-of-the-art results datasets. Second, tackle problem data compression very (on order 105 dimensions) using two lossy strategies: reduction technique known hash kernel an encoding based product quantizers. explain how gain in storage can be traded against loss and/or increase CPU cost. report databases — ImageNet dataset lM Flickr showing reduce our by factor 64 128 with little Integrating decompression classifier learning yields efficient scalable algorithm. On ILSVRC2010 74.3% at top-5, which corresponds 2.5% absolute improvement respect state-of-the-art. subset 10K top-1 16.7%, relative 160%