Cost-sensitive feature selection via f-measure optimization reduction

Publication Type:
Conference Proceeding
31st AAAI Conference on Artificial Intelligence, AAAI 2017, 2017, pp. 2252 - 2258
Issue Date:
Filename Description Size
14175-66889-1-PB.pdfPublished version1.19 MB
Adobe PDF
Full metadata record
Copyright © 2017, Association for the Advancement of Artificial Intelligence ( All rights reserved. Feature selection aims to select a small subset from the high-dimensional features which can lead to better learning performance, lower computational complexity, and better model readability. The class imbalance problem has been neglected by traditional feature selection methods, therefore the selected features will be biased towards the majority classes. Because of the superiority of F-measure to accuracy for imbalanced data, we propose to use F-measure as the performance measure for feature selection algorithms. As a pseudo-linear function, the optimization of F-measure can be achieved by minimizing the total costs. In this paper, we present a novel cost-sensitive feature selection (CSFS) method which optimizes F-measure instead of accuracy to take class imbalance issue into account. The features will be selected according to optimal F-measure classifier after solving a series of cost-sensitive feature selection sub-problems. The features selected by our method will fully represent the characteristics of not only majority classes, but also minority classes. Extensive experimental results conducted on synthetic, multi-class and multi-label datasets validate the efficiency and significance of our feature selection method.
Please use this identifier to cite or link to this item: