2013 IEEE 13th International Conference on Data Mining (2013)
Dallas, TX, USA USA
Dec. 7, 2013 to Dec. 10, 2013
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/ICDM.2013.153
This paper introduces a supervised metric learning algorithm, called kernel density metric learning (KDML), which is easy to use and provides nonlinear, probability-based distance measures. KDML constructs a direct nonlinear mapping from the original input space into a feature space based on kernel density estimation. The nonlinear mapping in KDML embodies established distance measures between probability density functions, and leads to correct classification on datasets for which linear metric learning methods would fail. It addresses the severe challenge to kNN when features are from heterogeneous domains and, as a result, the Euclidean or Mahalanobis distance between original feature vectors is not meaningful. Existing metric learning algorithms can then be applied to the KDML features. We also propose an integrated optimization algorithm that learns not only the Mahalanobis matrix but also kernel bandwidths, the only hyper-parameters in the nonlinear mapping. KDML can naturally handle not only numerical features, but also categorical ones, which is rarely found in previous metric learning algorithms. Extensive experimental results on various datasets show that KDML significantly improves existing metric learning algorithms in terms of kNN classification accuracy.
Kernel, Euclidean distance, Vectors, Learning systems, Density measurement, Optimization
Y. He, W. Chen, Y. Chen and Y. Mao, "Kernel Density Metric Learning," 2013 IEEE 13th International Conference on Data Mining(ICDM), Dallas, TX, USA USA, 2013, pp. 271-280.