Issue No. 07 - July (2010 vol. 22)
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/TKDE.2009.126
Si Si , Nanyang Technological University, Singapore
Dacheng Tao , Nanyang Technological University, Singapore
Bo Geng , Nanyang Technological University, Singapore
The regularization principals [CHECK END OF SENTENCE] lead approximation schemes to deal with various learning problems, e.g., the regularization of the norm in a reproducing kernel Hilbert space for the ill-posed problem. In this paper, we present a family of subspace learning algorithms based on a new form of regularization, which transfers the knowledge gained in training samples to testing samples. In particular, the new regularization minimizes the Bregman divergence between the distribution of training samples and that of testing samples in the selected subspace, so it boosts the performance when training and testing samples are not independent and identically distributed. To test the effectiveness of the proposed regularization, we introduce it to popular subspace learning algorithms, e.g., principal components analysis (PCA) for cross-domain face modeling; and Fisher's linear discriminant analysis (FLDA), locality preserving projections (LPP), marginal Fisher's analysis (MFA), and discriminative locality alignment (DLA) for cross-domain face recognition and text categorization. Finally, we present experimental evidence on both face image data sets and text data sets, suggesting that the proposed Bregman divergence-based regularization is effective to deal with cross-domain learning problems.
Dimensionality reduction, regularization, and Bregman divergence.
B. Geng, S. Si and D. Tao, "Bregman Divergence-Based Regularization for Transfer Subspace Learning," in IEEE Transactions on Knowledge & Data Engineering, vol. 22, no. , pp. 929-942, 2009.