Semisupervised feature selection via spline regression for video semantic recognition

Publication Type:
Journal Article
IEEE Transactions on Neural Networks and Learning Systems, 2015, 26 (2), pp. 252 - 264
Issue Date:
Filename Description Size
06786497.pdfPublished Version3.75 MB
Adobe PDF
Full metadata record
© 2014 IEEE. To improve both the efficiency and accuracy of video semantic recognition, we can perform feature selection on the extracted video features to select a subset of features from the high-dimensional feature set for a compact and accurate video data representation. Provided the number of labeled videos is small, supervised feature selection could fail to identify the relevant features that are discriminative to target classes. In many applications, abundant unlabeled videos are easily accessible. This motivates us to develop semisupervised feature selection algorithms to better identify the relevant video features, which are discriminative to target classes by effectively exploiting the information underlying the huge amount of unlabeled video data. In this paper, we propose a framework of video semantic recognition by semisupervised feature selection via spline regression (S2FS2R). Two scatter matrices are combined to capture both the discriminative information and the local geometry structure of labeled and unlabeled training videos: A within-class scatter matrix encoding discriminative information of labeled training videos and a spline scatter output from a local spline regression encoding data distribution. An ℓ2,1-norm is imposed as a regularization term on the transformation matrix to ensure it is sparse in rows, making it particularly suitable for feature selection. To efficiently solve S2FS2R, we develop an iterative algorithm and prove its convergency. In the experiments, three typical tasks of video semantic recognition, such as video concept detection, video classification, and human action recognition, are used to demonstrate that the proposed S2FS2R achieves better performance compared with the state-of-the-art methods.
Please use this identifier to cite or link to this item: