TY - JOUR
T1 - CLIP-guided continual novel class discovery
AU - Yan, Qingsen
AU - Yang, Yiting
AU - Dai, Yutong
AU - Zhang, Xing
AU - Wiltos, Katarzyna
AU - Woźniak, Marcin
AU - Dong, Wei
AU - Zhang, Yanning
N1 - Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2025/2/15
Y1 - 2025/2/15
N2 - Continual Novel Class Discovery (CNCD) aims to adapt a trained classification model to a new task while maintaining its performance on the old task. This presents two main challenges: (1) unsupervised learning of new tasks and (2) avoiding forgetting old classes when previous data is unavailable. Some prior works use task IDs to identify old and novel classes for parameter isolation, while others waive the requirement of task IDs by combining novel class discovery and old knowledge preservation into a single training process. However, this often leads to interference with feature space and presents difficulties in balancing old and new knowledge. This work proposes a method that does not require task IDs and argues that decoupling the training process is beneficial. We find that a simple semi-supervised learning strategy with prototype adaptation can unleash the strong generalization ability of the CLIP model to a small CNCD model for novel class discovery. However, this operation may deteriorate the performance of old classes. To address this issue, CutMix is utilized to improve the network's representation and preserve old knowledge. Compared to the baseline method, our method not only surpasses it on the novel classes to a significant margin (33.1% on the TinyImageNet) but also exhibits more accurate prediction on old classes (2.9% on the TinyImageNet). These advantages are further boosted when multiple novel class discovery steps are required (31.2%→56.1% on the TinyImageNet regarding the overall performance). Code will be made available.
AB - Continual Novel Class Discovery (CNCD) aims to adapt a trained classification model to a new task while maintaining its performance on the old task. This presents two main challenges: (1) unsupervised learning of new tasks and (2) avoiding forgetting old classes when previous data is unavailable. Some prior works use task IDs to identify old and novel classes for parameter isolation, while others waive the requirement of task IDs by combining novel class discovery and old knowledge preservation into a single training process. However, this often leads to interference with feature space and presents difficulties in balancing old and new knowledge. This work proposes a method that does not require task IDs and argues that decoupling the training process is beneficial. We find that a simple semi-supervised learning strategy with prototype adaptation can unleash the strong generalization ability of the CLIP model to a small CNCD model for novel class discovery. However, this operation may deteriorate the performance of old classes. To address this issue, CutMix is utilized to improve the network's representation and preserve old knowledge. Compared to the baseline method, our method not only surpasses it on the novel classes to a significant margin (33.1% on the TinyImageNet) but also exhibits more accurate prediction on old classes (2.9% on the TinyImageNet). These advantages are further boosted when multiple novel class discovery steps are required (31.2%→56.1% on the TinyImageNet regarding the overall performance). Code will be made available.
KW - Class-incremental learning
KW - Computer vision
KW - Knowledge distillation
KW - Novel class discovery
UR - http://www.scopus.com/inward/record.url?scp=85214482014&partnerID=8YFLogxK
U2 - 10.1016/j.knosys.2024.112920
DO - 10.1016/j.knosys.2024.112920
M3 - 文章
AN - SCOPUS:85214482014
SN - 0950-7051
VL - 310
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
M1 - 112920
ER -