Bregman divergence-based regularization for transfer subspace learning

Publication Type:
Journal Article
Citation:
IEEE Transactions on Knowledge and Data Engineering, 2010, 22 (7), pp. 929 - 942
Issue Date:
2010-06-04
Filename Description Size
Thumbnail2011000232OK.pdf3.22 MB
Adobe PDF
Full metadata record
The regularization principals [31] lead approximation schemes to deal with various learning problems, e.g., the regularization of the norm in a reproducing kernel Hilbert space for the ill-posed problem. In this paper, we present a family of subspace learning algorithms based on a new form of regularization, which transfers the knowledge gained in training samples to testing samples. In particular, the new regularization minimizes the Bregman divergence between the distribution of training samples and that of testing samples in the selected subspace, so it boosts the performance when training and testing samples are not independent and identically distributed. To test the effectiveness of the proposed regularization, we introduce it to popular subspace learning algorithms, e.g., principal components analysis (PCA) for cross-domain face modeling; and Fisher's linear discriminant analysis (FLDA), locality preserving projections (LPP), marginal Fisher's analysis (MFA), and discriminative locality alignment (DLA) for cross-domain face recognition and text categorization. Finally, we present experimental evidence on both face image data sets and text data sets, suggesting that the proposed Bregman divergence-based regularization is effective to deal with cross-domain learning problems. © 2010 IEEE.
Please use this identifier to cite or link to this item: