TY - GEN
T1 - Unsupervised feature analysis with class margin optimization
AU - Wang, Sen
AU - Nie, Feiping
AU - Chang, Xiaojun
AU - Yao, Lina
AU - Li, Xue
AU - Sheng, Quan Z.
N1 - Publisher Copyright:
© Springer International Publishing Switzerland 2015.
PY - 2015
Y1 - 2015
N2 - Unsupervised feature selection has been attracting research attention in the communities of machine learning and data mining for decades. In this paper, we propose an unsupervised feature selection method seeking a feature coefficient matrix to select the most distinctive features. Specifically, our proposed algorithm integrates the Maximum Margin Criterion with a sparsity-based model into a joint framework, where the class margin and feature correlation are taken into account at the same time. To maximize the total data separability while preserving minimized within-class scatter simultaneously, we propose to embed Kmeans into the framework generating pseudo class label information in a scenario of unsupervised feature selection. Meanwhile, a sparsity-based model, ℓ2,p-norm, is imposed to the regularization term to effectively discover the sparse structures of the feature coefficient matrix. In this way, noisy and irrelevant features are removed by ruling out those features whose corresponding coefficients are zeros. To alleviate the local optimum problem that is caused by random initializations of K-means, a convergence guaranteed algorithm with an updating strategy for the clustering indicator matrix, is proposed to iteratively chase the optimal solution. Performance evaluation is extensively conducted over six benchmark data sets. From our comprehensive experimental results, it is demonstrated that our method has superior performance against all other compared approaches.
AB - Unsupervised feature selection has been attracting research attention in the communities of machine learning and data mining for decades. In this paper, we propose an unsupervised feature selection method seeking a feature coefficient matrix to select the most distinctive features. Specifically, our proposed algorithm integrates the Maximum Margin Criterion with a sparsity-based model into a joint framework, where the class margin and feature correlation are taken into account at the same time. To maximize the total data separability while preserving minimized within-class scatter simultaneously, we propose to embed Kmeans into the framework generating pseudo class label information in a scenario of unsupervised feature selection. Meanwhile, a sparsity-based model, ℓ2,p-norm, is imposed to the regularization term to effectively discover the sparse structures of the feature coefficient matrix. In this way, noisy and irrelevant features are removed by ruling out those features whose corresponding coefficients are zeros. To alleviate the local optimum problem that is caused by random initializations of K-means, a convergence guaranteed algorithm with an updating strategy for the clustering indicator matrix, is proposed to iteratively chase the optimal solution. Performance evaluation is extensively conducted over six benchmark data sets. From our comprehensive experimental results, it is demonstrated that our method has superior performance against all other compared approaches.
KW - Embedded K-means clustering
KW - Maximum margin criterion
KW - Sparse structure learning
KW - Unsupervised feature selection
UR - http://www.scopus.com/inward/record.url?scp=84984650296&partnerID=8YFLogxK
U2 - 10.1007/978-3-319-23528-8_24
DO - 10.1007/978-3-319-23528-8_24
M3 - 会议稿件
AN - SCOPUS:84984650296
SN - 9783319235271
SN - 9783319235271
SN - 9783319235271
SN - 9783319235271
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 383
EP - 398
BT - Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2015, Proceedings
A2 - Appice, Annalisa
A2 - Gama, João
A2 - Costa, Vitor Santos
A2 - Gama, João
A2 - Jorge, Alípio
A2 - Appice, Annalisa
A2 - Appice, Annalisa
A2 - Costa, Vitor Santos
A2 - Jorge, Alípio
A2 - Appice, Annalisa
A2 - Rodrigues, Pedro Pereira
A2 - Rodrigues, Pedro Pereira
A2 - Gama, João
A2 - Costa, Vitor Santos
A2 - Soares, Soares
A2 - Rodrigues, Pedro Pereira
A2 - Soares, Soares
A2 - Soares, Soares
A2 - Gama, João
A2 - Soares, Soares
A2 - Jorge, Alípio
A2 - Jorge, Alípio
A2 - Rodrigues, Pedro Pereira
A2 - Costa, Vitor Santos
PB - Springer Verlag
T2 - European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2015
Y2 - 7 September 2015 through 11 September 2015
ER -