Minority Oversampling in Kernel Adaptive Subspaces for Class Imbalanced Datasets

Publisher:
Institute of Electrical and Electronics Engineers
Publication Type:
Journal Article
Citation:
IEEE Transactions on Knowledge and Data Engineering, 2018, 30 (5), pp. 950 - 962
Issue Date:
2018-05-01
Full metadata record
Files in This Item:
Filename Description Size
656031E4-4318-45CC-8DEA-1D17E8DA8C01 am.pdfAccepted Manuscript Version1.83 MB
Adobe PDF
© 1989-2012 I EEE. The class imbalance problem in machine learning occurs when certain classes are underrepresented relative to the others, leading to a learning bias toward the majority classes. To cope with the skewed class distribution, many learning methods featuring minority oversampling have been proposed, which are proved to be effective. To reduce information loss during feature space projection, this study proposes a novel oversampling algorithm, named minority oversampling in kernel adaptive subspaces (MOKAS), which exploits the invariant feature extraction capability of a kernel version of the adaptive subspace self-organizing maps. The synthetic instances are generated from well-trained subspaces and then their pre-images are reconstructed in the input space. Additionally, these instances characterize nonlinear structures present in the minority class data distribution and help the learning algorithms to counterbalance the skewed class distribution in a desirable manner. Experimental results on both real and synthetic data show that the proposed MOKAS is capable of modeling complex data distribution and outperforms a set of state-of-the-art oversampling algorithms.
Please use this identifier to cite or link to this item: