TY - JOUR
T1 - Supervised Feature Selection via Multi-Center and Local Structure Learning
AU - Zhang, Canyu
AU - Nie, Feiping
AU - Wang, Rong
AU - Li, Xuelong
N1 - Publisher Copyright:
© 1989-2012 IEEE.
PY - 2024
Y1 - 2024
N2 - Feature selection has achieved unprecedented success in obtaining sparse discriminative features. However, the existing methods almost use the ell {2,p}ℓ2,p-norm constraint on transformation matrix to obtain sparse features, which introduces extra parameters and cannot obtain the features directly. In addition, existing algorithms only focused on the global structure and ignored the local structure, leading to poor performance when solving data with non-Gaussian distributions which a single center point cannot describe precisely. Based on above considerations, we propose a supervised feature selection via multi-center and local structure learning. We further introduce trace ratio criterion into our model in favor of improving the discriminant of features selected. In order to address the overlap problem, we use multiple center points to match the distribution of data and construct a kk-Nearest Neighbor graph to explore the local structure of the data. In addition, we also propose an efficient method to optimize the transformation matrix with the ell {2,0}ℓ2,0-norm constraint and can directly obtain the sparse features. We evaluate our method on Toy datasets and several real-world datasets, show improvement over state-of-the-art feature selection methods, and demonstrate the effectiveness of our model in dealing with non-Gaussian distributed data problems.
AB - Feature selection has achieved unprecedented success in obtaining sparse discriminative features. However, the existing methods almost use the ell {2,p}ℓ2,p-norm constraint on transformation matrix to obtain sparse features, which introduces extra parameters and cannot obtain the features directly. In addition, existing algorithms only focused on the global structure and ignored the local structure, leading to poor performance when solving data with non-Gaussian distributions which a single center point cannot describe precisely. Based on above considerations, we propose a supervised feature selection via multi-center and local structure learning. We further introduce trace ratio criterion into our model in favor of improving the discriminant of features selected. In order to address the overlap problem, we use multiple center points to match the distribution of data and construct a kk-Nearest Neighbor graph to explore the local structure of the data. In addition, we also propose an efficient method to optimize the transformation matrix with the ell {2,0}ℓ2,0-norm constraint and can directly obtain the sparse features. We evaluate our method on Toy datasets and several real-world datasets, show improvement over state-of-the-art feature selection methods, and demonstrate the effectiveness of our model in dealing with non-Gaussian distributed data problems.
KW - local structure
KW - multi-center
KW - sparse discriminant features
KW - Supervised feature selection
KW - ℓ2,0-norm constraint
UR - http://www.scopus.com/inward/record.url?scp=85187315493&partnerID=8YFLogxK
U2 - 10.1109/TKDE.2024.3372657
DO - 10.1109/TKDE.2024.3372657
M3 - 文章
AN - SCOPUS:85187315493
SN - 1041-4347
VL - 36
SP - 4930
EP - 4942
JO - IEEE Transactions on Knowledge and Data Engineering
JF - IEEE Transactions on Knowledge and Data Engineering
IS - 9
ER -