Discriminative vanishing component analysis

Publication Type:
Conference Proceeding
Citation:
30th AAAI Conference on Artificial Intelligence, AAAI 2016, 2016, pp. 1666 - 1672
Issue Date:
2016-01-01
Filename Description Size
11968-55673-1-PB.pdfPublished version672.59 kB
Adobe PDF
Full metadata record
© Copyright 2016, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Vanishing Component Analysis (VCA) is a recently proposed prominent work in machine learning. It narrows the gap between tools and computational algebra: The vanishing ideal and its applications to classification problem. In this paper, we will analyze VCA in the kernel view, which is also another important research direction in machine learning. Under a very weak assumption, we provide a different point of view to VCA and make the kernel trick on VCA become possible. We demonstrate that the projection matrix derived by VCA is located in the same space as that of Kernel Principal Component Analysis (KPCA) with a polynomial kernel. Two groups of projections can express each other by linear transformation. Furthermore, we prove that KPCA and VCA have identical discriminative power, provided that the ratio trace criteria is employed as the measurement. We also show that the kernel formulated by the inner products of VCA's projections can be expressed by the KPCA's kernel linearly. Based on the analysis above, we proposed a novel Discriminative Vanishing Component Analysis (DVCA) approach. Experimental results are provided for demonstration.
Please use this identifier to cite or link to this item: