TY - JOUR
T1 - Prototypical classifier with distribution consistency regularization for generalized category discovery
T2 - A strong baseline
AU - Hu, Zhanxuan
AU - Duan, Yu
AU - Zhang, Yaming
AU - Wang, Rong
AU - Nie, Feiping
N1 - Publisher Copyright:
© 2024 Elsevier Ltd
PY - 2025/2
Y1 - 2025/2
N2 - Generalized Category Discovery (GCD) addresses a more realistic and challenging setting in semi-supervised visual recognition, where unlabeled data contains samples from both known and novel categories. Recently, prototypical classifier has shown prominent performance on this issue, with the Softmax-based Cross-Entropy loss (SCE) commonly employed to optimize the distance between a sample and prototypes. However, the inherent non-bijectiveness of SCE prevents it from resolving intraclass relations among samples, resulting in semantic ambiguity. To mitigate this issue, we propose Distribution Consistency Regularization (DCR) for the prototypical classifier. By leveraging a simple intraclass consistency loss, we enforce the classifier to yield consistent distributions for samples belonging to the same class. In doing so, we equip the classifier to better capture local structures and alleviate semantic ambiguity. Additionally, we propose using partial labels, rather than hard pseudo labels, to explore potential positive pairs in unlabeled data, thereby reducing the risk of introducing noisy supervisory signals. DCR requires no external sophisticated module, rendering the enhanced model concise and efficient. Extensive experiments validate consistent performance benefits of DCR while achieving competitive or better performance on six benchmarks. Hence, our method can serve as a strong baseline for GCD. Our code is available at: https://github.com/yichenwang231/DCR.
AB - Generalized Category Discovery (GCD) addresses a more realistic and challenging setting in semi-supervised visual recognition, where unlabeled data contains samples from both known and novel categories. Recently, prototypical classifier has shown prominent performance on this issue, with the Softmax-based Cross-Entropy loss (SCE) commonly employed to optimize the distance between a sample and prototypes. However, the inherent non-bijectiveness of SCE prevents it from resolving intraclass relations among samples, resulting in semantic ambiguity. To mitigate this issue, we propose Distribution Consistency Regularization (DCR) for the prototypical classifier. By leveraging a simple intraclass consistency loss, we enforce the classifier to yield consistent distributions for samples belonging to the same class. In doing so, we equip the classifier to better capture local structures and alleviate semantic ambiguity. Additionally, we propose using partial labels, rather than hard pseudo labels, to explore potential positive pairs in unlabeled data, thereby reducing the risk of introducing noisy supervisory signals. DCR requires no external sophisticated module, rendering the enhanced model concise and efficient. Extensive experiments validate consistent performance benefits of DCR while achieving competitive or better performance on six benchmarks. Hence, our method can serve as a strong baseline for GCD. Our code is available at: https://github.com/yichenwang231/DCR.
KW - Clustering
KW - Generalized category discovery
KW - Semi-supervised learning
KW - Unsupervised learning
UR - http://www.scopus.com/inward/record.url?scp=85209560995&partnerID=8YFLogxK
U2 - 10.1016/j.neunet.2024.106908
DO - 10.1016/j.neunet.2024.106908
M3 - 文章
C2 - 39571384
AN - SCOPUS:85209560995
SN - 0893-6080
VL - 182
JO - Neural Networks
JF - Neural Networks
M1 - 106908
ER -