Supervised Gaussian process latent variable model for dimensionality reduction
- Publication Type:
- Journal Article
- Citation:
- IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 2011, 41 (2), pp. 425 - 434
- Issue Date:
- 2011-04-01
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
![]() | 2011000246OK.pdf | 1.08 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
The Gaussian process latent variable model (GP-LVM) has been identified to be an effective probabilistic approach for dimensionality reduction because it can obtain a low-dimensional manifold of a data set in an unsupervised fashion. Consequently, the GP-LVM is insufficient for supervised learning tasks (e.g., classification and regression) because it ignores the class label information for dimensionality reduction. In this paper, a supervised GP-LVM is developed for supervised learning tasks, and the maximum a posteriori algorithm is introduced to estimate positions of all samples in the latent variable space. We present experimental evidences suggesting that the supervised GP-LVM is able to use the class label information effectively, and thus, it outperforms the GP-LVM and the discriminative extension of the GP-LVM consistently. The comparison with some supervised classification methods, such as Gaussian process classification and support vector machines, is also given to illustrate the advantage of the proposed method. © 2006 IEEE.
Please use this identifier to cite or link to this item: