Semi-Supervised Feature Selection via Sparse Rescaled Linear Square Regression

Xiaojun Chen, Guowen Yuan, Feiping Nie, Zhong Ming

Research output: Contribution to journalArticlepeer-review

78 Scopus citations

Abstract

With the rapid increase of the data size, it has increasing demands for selecting features by exploiting both labeled and unlabeled data. In this paper, we propose a novel semi-supervised embedded feature selection method. The new method extends the least square regression model by rescaling the regression coefficients in the least square regression with a set of scale factors, which is used for evaluating the importance of features. An iterative algorithm is proposed to optimize the new model. It has been proved that solving the new model is equivalent to solving a sparse model with a flexible and adaptable ℓ2,pℓ2,p norm regularization. Moreover, the optimal solution of scale factors provides a theoretical explanation for why we can use w1 Vert2, Vert wdVert2w12,..,wd2 to evaluate the importance of features. Experimental results on eight benchmark data sets show the superior performance of the proposed method.

Original languageEnglish
Article number8528552
Pages (from-to)165-176
Number of pages12
JournalIEEE Transactions on Knowledge and Data Engineering
Volume32
Issue number1
DOIs
StatePublished - Jan 2020

Keywords

  • Feature selection
  • least square regression
  • semi-supervised feature selection
  • sparse feature selection

Fingerprint

Dive into the research topics of 'Semi-Supervised Feature Selection via Sparse Rescaled Linear Square Regression'. Together they form a unique fingerprint.

Cite this