Multiclass capped ℓp-norm SVM for robust classifications

Research output: Contribution to conferencePaperpeer-review

73 Scopus citations

Abstract

Support vector machine (SVM) model is one of most successful machine learning methods and has been successfully applied to solve numerous real-world application. Because the SVM methods use the hinge loss or squared hinge loss functions for classifications, they usually outperform other classification approaches, e.g. the least square loss function based methods. However, like most supervised learning algorithms, they learn classifiers based on the labeled data in training set without specific strategy to deal with the noise data. In many real-world applications, we often have data outliers in train set, which could misguide the classifiers learning, such that the classification performance is suboptimal. To address this problem, we proposed a novel capped ℓp-norm SVM classification model by utilizing the capped ℓp-norm based hinge loss in the objective which can deal with both light and heavy outliers. We utilize the new formulation to naturally build the multiclass capped ℓp-norm SVM. More importantly, we derive a novel optimization algorithms to efficiently minimize the capped ℓp-norm based objectives, and also rigorously prove the convergence of proposed algorithms. We present experimental results showing that employing the new capped ℓp-norm SVM method can consistently improve the classification performance, especially in the cases when the data noise level increases.

Original languageEnglish
Pages2415-2421
Number of pages7
StatePublished - 2017
Event31st AAAI Conference on Artificial Intelligence, AAAI 2017 - San Francisco, United States
Duration: 4 Feb 201710 Feb 2017

Conference

Conference31st AAAI Conference on Artificial Intelligence, AAAI 2017
Country/TerritoryUnited States
CitySan Francisco
Period4/02/1710/02/17

Fingerprint

Dive into the research topics of 'Multiclass capped ℓp-norm SVM for robust classifications'. Together they form a unique fingerprint.

Cite this