Issue No. 03 - March (2014 vol. 36)
Zeynep Akata , Xerox Res. Centre Eur., Meylan, France
Florent Perronnin , Xerox Res. Centre Eur., Meylan, France
Zaid Harchaoui , INRIA Grenoble Rhone-Alpes, Isere, France
Cordelia Schmid , INRIA Grenoble Rhone-Alpes, Isere, France
We benchmark several SVM objective functions for large-scale image classification. We consider one-versus-rest, multiclass, ranking, and weighted approximate ranking SVMs. A comparison of online and batch methods for optimizing the objectives shows that online methods perform as well as batch methods in terms of classification accuracy, but with a significant gain in training speed. Using stochastic gradient descent, we can scale the training to millions of images and thousands of classes. Our experimental evaluation shows that ranking-based algorithms do not outperform the one-versus-rest strategy when a large number of training examples are used. Furthermore, the gap in accuracy between the different algorithms shrinks as the dimension of the features increases. We also show that learning through cross-validation the optimal rebalancing of positive and negative examples can result in a significant improvement for the one-versus-rest strategy. Finally, early stopping can be used as an effective regularization strategy when training with online algorithms. Following these "good practices," we were able to improve the state of the art on a large subset of 10K classes and 9M images of ImageNet from 16.7 percent Top-1 accuracy to 19.1 percent.
Support vector machines, Training, Linear programming, Accuracy, Optimization, Visualization, Encoding,stochastic learning, Large scale, fine-grained visual categorization, image classification, ranking, SVM
Zeynep Akata, Florent Perronnin, Zaid Harchaoui, Cordelia Schmid, "Good Practice in Large-Scale Learning for Image Classification", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. 36, no. , pp. 507-520, March 2014, doi:10.1109/TPAMI.2013.146