Fast Sparse Discriminative K-Means for Unsupervised Feature Selection

Research output: Contribution to journalArticlepeer-review

30 Scopus citations

Abstract

Embedded feature selection approach guides subsequent projection matrix (selection matrix) learning through the acquisition of pseudolabel matrix to conduct feature selection tasks. Yet the continuous pseudolabel matrix learned from relaxed problem based on spectral analysis deviates from reality to some extent. To cope with this issue, we design an efficient feature selection framework inspired by classical least-squares regression (LSR) and discriminative K-means (DisK-means), which is called the fast sparse discriminative K-means (FSDK) for the feature selection method. First, the weighted pseudolabel matrix with discrete trait is introduced to avoid trivial solution from unsupervised LSR. On this condition, any constraint imposed into pseudolabel matrix and selection matrix is dispensable, which is significantly beneficial to simplify the combinational optimization problem. Second, the ℓ2,p -norm regularizer is introduced to satisfy the row sparsity of selection matrix with flexible p. Consequently, the proposed FSDK model can be treated as a novel feature selection framework integrated from the DisK-means algorithm and ℓ2,p -norm regularizer to optimize the sparse regression problem. Moreover, our model is linearly correlated with the number of samples, which is speedy to handle the large-scale data. Comprehensive tests on various data terminally illuminate the effectiveness and efficiency of FSDK.

Original languageEnglish
Pages (from-to)9943-9957
Number of pages15
JournalIEEE Transactions on Neural Networks and Learning Systems
Volume35
Issue number7
DOIs
StatePublished - 2024

Keywords

  • embedded feature selection
  • least-square regression (LSR)
  • sparse discriminative K-means
  • trivial solution
  • weighted pseudolabel matrix
  • ℓ-norm regularizer

Fingerprint

Dive into the research topics of 'Fast Sparse Discriminative K-Means for Unsupervised Feature Selection'. Together they form a unique fingerprint.

Cite this