TY - JOUR
T1 - Semi-supervised Self-Training Algorithm for Density Peak Membership Optimization
AU - Liu, Xuewen
AU - Wang, Jikui
AU - Yang, Zhengguo
AU - Li, Bing
AU - Nie, Feiping
N1 - Publisher Copyright:
© 2022, Journal of Computer Engineering and Applications Beijing Co., Ltd.; Science Press. All rights reserved.
PY - 2022/9/1
Y1 - 2022/9/1
N2 - Most of data contain only a few labels because of high cost of obtaining them in reality. Compared with supervised learning and unsupervised learning, semi-supervised learning can obtain higher learning performance with less labeling cost by making full use of large amount of unlabeled data and small amount of labeled data in datasets. Self-Training algorithm is a classical semi-supervised learning algorithm. In the process of iteratively optimizing classifier, high-confidence samples are continuously selected from unlabeled samples and labeled by the base classifier. Then, these samples and pseudo-labels will be added into the training sets. Selecting high-confidence samples is a critical step in the Self-Training algorithm. Inspired by the density peaks clustering (DPC) algorithm, this paper proposes semi-supervised Self-Training algorithm for density peak membership optimization (STDPM), which uses den sity peak to select high-confidence samples. Firstly, STDPM takes density peak to discover the potential spatial structure information of the samples and constructs a prototype tree. Secondly, STDPM searches the unlabeled direct relatives of the labeled samples in the prototype tree, and defines the density peak of the unlabeled direct relatives that belong to different clusters as the clusters-peak. Then, clusters-peak is turned into the density peak membership after normalized. Finally, STDPM regards samples with membership greater than the set threshold as high-confidence samples that are labeled by the base classifier and added to the training set. STDPM makes full use of the density and distance information implied by the peak, which improves the selection quality of high-confidence samples and further improves the classification performance. Comparative experiments are conducted on 8 benchmark datasets, which verify the effectiveness of STDPM.
AB - Most of data contain only a few labels because of high cost of obtaining them in reality. Compared with supervised learning and unsupervised learning, semi-supervised learning can obtain higher learning performance with less labeling cost by making full use of large amount of unlabeled data and small amount of labeled data in datasets. Self-Training algorithm is a classical semi-supervised learning algorithm. In the process of iteratively optimizing classifier, high-confidence samples are continuously selected from unlabeled samples and labeled by the base classifier. Then, these samples and pseudo-labels will be added into the training sets. Selecting high-confidence samples is a critical step in the Self-Training algorithm. Inspired by the density peaks clustering (DPC) algorithm, this paper proposes semi-supervised Self-Training algorithm for density peak membership optimization (STDPM), which uses den sity peak to select high-confidence samples. Firstly, STDPM takes density peak to discover the potential spatial structure information of the samples and constructs a prototype tree. Secondly, STDPM searches the unlabeled direct relatives of the labeled samples in the prototype tree, and defines the density peak of the unlabeled direct relatives that belong to different clusters as the clusters-peak. Then, clusters-peak is turned into the density peak membership after normalized. Finally, STDPM regards samples with membership greater than the set threshold as high-confidence samples that are labeled by the base classifier and added to the training set. STDPM makes full use of the density and distance information implied by the peak, which improves the selection quality of high-confidence samples and further improves the classification performance. Comparative experiments are conducted on 8 benchmark datasets, which verify the effectiveness of STDPM.
KW - clusters-peak
KW - density peak membership
KW - direct relative node sets
KW - prototype tree
KW - self-training
UR - http://www.scopus.com/inward/record.url?scp=85146517299&partnerID=8YFLogxK
U2 - 10.3778/j.issn.1673-9418.2102018
DO - 10.3778/j.issn.1673-9418.2102018
M3 - 文章
AN - SCOPUS:85146517299
SN - 1673-9418
VL - 16
SP - 2078
EP - 2088
JO - Journal of Frontiers of Computer Science and Technology
JF - Journal of Frontiers of Computer Science and Technology
IS - 9
ER -