TY - GEN
T1 - Adaptive loss minimization for semi-supervised elastic embedding
AU - Nie, Feiping
AU - Wang, Hua
AU - Huang, Heng
AU - Ding, Chris
PY - 2013
Y1 - 2013
N2 - The semi-supervised learning usually only predict labels for unlabeled data appearing in training data, and cannot effectively predict labels for testing data never appearing in training set. To handle this outof- sample problem, many inductive methods make a constraint such that the predicted label matrix should be exactly equal to a linear model. In practice, this constraint is too rigid to capture the manifold structure of data. Motivated by this deficiency, we relax the rigid linear embedding constraint and propose to use an elastic embedding constraint on the predicted label matrix such that the manifold structure can be better explored. To solve our new objective and also a more general optimization problem, we study a novel adaptive loss with efficient optimization algorithm. Our new adaptive loss minimization method takes the advantages of both L1 norm and L2 norm, and is robust to the data outlier under Laplacian distribution and can efficiently learn the normal data under Gaussian distribution. Experiments have been performed on image classification tasks and our approach outperforms other state-of-the-art methods.
AB - The semi-supervised learning usually only predict labels for unlabeled data appearing in training data, and cannot effectively predict labels for testing data never appearing in training set. To handle this outof- sample problem, many inductive methods make a constraint such that the predicted label matrix should be exactly equal to a linear model. In practice, this constraint is too rigid to capture the manifold structure of data. Motivated by this deficiency, we relax the rigid linear embedding constraint and propose to use an elastic embedding constraint on the predicted label matrix such that the manifold structure can be better explored. To solve our new objective and also a more general optimization problem, we study a novel adaptive loss with efficient optimization algorithm. Our new adaptive loss minimization method takes the advantages of both L1 norm and L2 norm, and is robust to the data outlier under Laplacian distribution and can efficiently learn the normal data under Gaussian distribution. Experiments have been performed on image classification tasks and our approach outperforms other state-of-the-art methods.
UR - http://www.scopus.com/inward/record.url?scp=84896060974&partnerID=8YFLogxK
M3 - 会议稿件
AN - SCOPUS:84896060974
SN - 9781577356332
T3 - IJCAI International Joint Conference on Artificial Intelligence
SP - 1565
EP - 1571
BT - IJCAI 2013 - Proceedings of the 23rd International Joint Conference on Artificial Intelligence
T2 - 23rd International Joint Conference on Artificial Intelligence, IJCAI 2013
Y2 - 3 August 2013 through 9 August 2013
ER -