A Convex Formulation for Fast Semi-Supervised Learning

  • Xinyi Fan
  • , Weizhong Yu
  • , Feiping Nie
  • , Zongcheng Miao
  • , Xuelong Li

Research output: Contribution to journalArticlepeer-review

Abstract

As a compromise between supervised and unsupervised learning, semi-supervised learning (SSL) harnesses both labeled and unlabeled data to enhance learning performance. Graph-based semi-supervised learning (GSSL) has emerged as a prominent approach owing to its versatility in representing sample interdependencies via graph structures. However, traditional GSSL methods face high time cost when computing matrix inverses, making them inefficient for large datasets. To address this, some researchers have introduced anchors as a bridge to accelerate the process. Nevertheless, most anchor-based models suffer from one or more of the following issues: (1) The anchor graph-based construction of the adjacency matrix has limitations; (2) The objective functions are typically non-convex, leading to local optima and requiring multiple runs to achieve good performance. To tackle these challenges, we develop a probability-driven approach to build the adjacency matrix, defining sample similarity as the probability of sharing the same anchor. Based on this strategy, we design a model (CFSL) with a strictly convex objective function, guaranteeing a globally optimal solution without iterative optimization. Experiments on multiple datasets indicate that our algorithm yields strong performance.

Original languageEnglish
Pages (from-to)6738-6749
Number of pages12
JournalIEEE Transactions on Knowledge and Data Engineering
Volume37
Issue number12
DOIs
StatePublished - 2025

Keywords

  • Semi-supervised learning
  • anchor-based graph
  • convex formulation
  • large-scale data

Fingerprint

Dive into the research topics of 'A Convex Formulation for Fast Semi-Supervised Learning'. Together they form a unique fingerprint.

Cite this