TY - JOUR
T1 - Outliers Robust Unsupervised Feature Selection for Structured Sparse Subspace
AU - Wang, Sisi
AU - Nie, Feiping
AU - Wang, Zheng
AU - Wang, Rong
AU - Li, Xuelong
N1 - Publisher Copyright:
© 1989-2012 IEEE.
PY - 2024/3/1
Y1 - 2024/3/1
N2 - Feature selection is one of the important topics of machine learning, and it has a wide range of applications in data preprocessing. At present, feature selection based on ℓ 2,1-norm regularization is a relatively mature method, but it is not enough to maximize the sparsity and parameter-tuning leads to increased costs. Later scholars found that the ℓ 2,0-norm constraint is more conductive to feature selection, but it is difficult to solve and lacks convergence guarantees. To address these problems, we creatively propose a novel Outliers Robust Unsupervised Feature Selection for structured sparse subspace (ORUFS), which utilizes ℓ2,0-norm constraint to learn a structured sparse subspace and avoid tuning the regularization parameter. Moreover, by adding binary weights, outliers are directly eliminated and the robustness of model is improved. More importantly, a Re-Weighted (RW) algorithm is exploited to solve our ℓp-norm problem. For the NP-hard problem of ℓ2,0-norm constraint, we develop an effective iterative optimization algorithm with strict convergence guarantees and closed-form solution. Subsequently, we provide theoretical analysis about convergence and computational complexity. Experimental results on real-world datasets illustrate that our method is superior to the state-of-the-art methods in clustering and anomaly detection tasks.
AB - Feature selection is one of the important topics of machine learning, and it has a wide range of applications in data preprocessing. At present, feature selection based on ℓ 2,1-norm regularization is a relatively mature method, but it is not enough to maximize the sparsity and parameter-tuning leads to increased costs. Later scholars found that the ℓ 2,0-norm constraint is more conductive to feature selection, but it is difficult to solve and lacks convergence guarantees. To address these problems, we creatively propose a novel Outliers Robust Unsupervised Feature Selection for structured sparse subspace (ORUFS), which utilizes ℓ2,0-norm constraint to learn a structured sparse subspace and avoid tuning the regularization parameter. Moreover, by adding binary weights, outliers are directly eliminated and the robustness of model is improved. More importantly, a Re-Weighted (RW) algorithm is exploited to solve our ℓp-norm problem. For the NP-hard problem of ℓ2,0-norm constraint, we develop an effective iterative optimization algorithm with strict convergence guarantees and closed-form solution. Subsequently, we provide theoretical analysis about convergence and computational complexity. Experimental results on real-world datasets illustrate that our method is superior to the state-of-the-art methods in clustering and anomaly detection tasks.
KW - anomaly detection
KW - clustering
KW - outlier robust
KW - unsupervised feature selection
KW - ℓ-norm constraint
UR - http://www.scopus.com/inward/record.url?scp=85165874063&partnerID=8YFLogxK
U2 - 10.1109/TKDE.2023.3297226
DO - 10.1109/TKDE.2023.3297226
M3 - 文章
AN - SCOPUS:85165874063
SN - 1041-4347
VL - 36
SP - 1234
EP - 1248
JO - IEEE Transactions on Knowledge and Data Engineering
JF - IEEE Transactions on Knowledge and Data Engineering
IS - 3
ER -