TY - JOUR
T1 - A parameter-free self-training algorithm based on the three successive confirmation rule
AU - Wang, Jikui
AU - Zhao, Wei
AU - Shang, Qingsheng
AU - Nie, Feiping
N1 - Publisher Copyright:
© 2025 Elsevier Ltd
PY - 2025/3/15
Y1 - 2025/3/15
N2 - Semi-supervised learning is a popular research topic today, and self-training is a classical semi-supervised learning framework. How to select high-confidence samples in self-training is a critical step. However, the existing algorithms do not consider both global and local information of the data. In the paper, we propose a parameter-free self-training algorithm based on the three successive confirmation rule, which integrates global and local information to identify high-confidence samples. Concretely, the local information is obtained by using k nearest neighbors and global information is derived from the three successive confirmation rule. This dual selection strategy helps to improve the quality of high-confidence samples and further improve the performance of classification. We conduct experiments on 14 benchmark datasets, comparing our method with other self-training algorithms. We use accuracy and F-score as performance metrics. The experimental results demonstrate that our algorithm significantly improves classification performance, proving its effectiveness and superiority in semi-supervised learning.
AB - Semi-supervised learning is a popular research topic today, and self-training is a classical semi-supervised learning framework. How to select high-confidence samples in self-training is a critical step. However, the existing algorithms do not consider both global and local information of the data. In the paper, we propose a parameter-free self-training algorithm based on the three successive confirmation rule, which integrates global and local information to identify high-confidence samples. Concretely, the local information is obtained by using k nearest neighbors and global information is derived from the three successive confirmation rule. This dual selection strategy helps to improve the quality of high-confidence samples and further improve the performance of classification. We conduct experiments on 14 benchmark datasets, comparing our method with other self-training algorithms. We use accuracy and F-score as performance metrics. The experimental results demonstrate that our algorithm significantly improves classification performance, proving its effectiveness and superiority in semi-supervised learning.
KW - High-confidence samples
KW - Self-training algorithm
KW - Semi-supervised learning
KW - Three successive confirmation rule
UR - http://www.scopus.com/inward/record.url?scp=85216518332&partnerID=8YFLogxK
U2 - 10.1016/j.engappai.2025.110165
DO - 10.1016/j.engappai.2025.110165
M3 - 文章
AN - SCOPUS:85216518332
SN - 0952-1976
VL - 144
JO - Engineering Applications of Artificial Intelligence
JF - Engineering Applications of Artificial Intelligence
M1 - 110165
ER -