GBVSSL: Contrastive Semi-Supervised Learning Based on Generalized Bias-Variance Decomposition
Author:
Li Shu1, Han Lixin1, Wang Yang2ORCID, Zhu Jun3
Affiliation:
1. School of Computer and Information, Hohai University, Nanjing 211100, China 2. School of Computer and Information, Anqing Normal University, Anqing 246133, China 3. School of Computer and Software, Nanjing Vocational University of Industry Technology, Nanjing 210023, China
Abstract
Mainstream semi-supervised learning (SSL) techniques, such as pseudo-labeling and contrastive learning, exhibit strong generalization abilities but lack theoretical understanding. Furthermore, pseudo-labeling lacks the label enhancement from high-quality neighbors, while contrastive learning ignores the supervisory guidance provided by genuine labels. To this end, we first introduce a generalized bias-variance decomposition framework to investigate them. Then, this research inspires us to propose two new techniques to refine them: neighbor-enhanced pseudo-labeling, which enhances confidence-based pseudo-labels by incorporating aggregated predictions from high-quality neighbors; label-enhanced contrastive learning, which enhances feature representation by combining enhanced pseudo-labels and ground-truth labels to construct a reliable and complete symmetric adjacency graph. Finally, we combine these two new techniques to develop an excellent SSL method called GBVSSL. GBVSSL significantly surpasses previous state-of-the-art SSL approaches in standard benchmarks, such as CIFAR-10/100, SVHN, and STL-10. On CIFAR-100 with 400, 2500, and 10,000 labeled samples, GBVSSL outperforms FlexMatch by 3.46%, 2.72%, and 2.89%, respectively. On the real-world dataset Semi-iNat 2021, GBVSSL improves the Top-1 accuracy over CCSSL by 4.38%. Moreover, GBVSSL exhibits faster convergence and enhances unbalanced SSL. Extensive ablation and qualitative studies demonstrate the effectiveness and impact of each component of GBVSSL.
Funder
Natural Science Foundation of Colleges and Universities in Anhui Province of China
Reference42 articles.
1. Yang, F., Wu, K., Zhang, S., Jiang, G., Liu, Y., Zheng, F., Zhang, W., Wang, C., and Zeng, L. (2022, January 18–24). Class-aware contrastive semi-supervised learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA. 2. Fixmatch: Simplifying semi-supervised learning with consistency and confidence;Sohn;Adv. Neural Inf. Process. Syst.,2020 3. Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling;Zhang;Adv. Neural Inf. Process. Syst.,2021 4. Berthelot, D., Carlini, N., Cubuk, E.D., Kurakin, A., Sohn, K., Zhang, H., and Raffel, C. (May, January 26). ReMixMatch: Semi-Supervised Learning with Distribution Alignment and Augmentation Anchoring. Proceedings of the International Conference on Learning Representations, Online. 5. Chen, H., Tao, R., Fan, Y., Wang, Y., Wang, J., Schiele, B., Xie, X., Raj, B., and Savvides, M. (2023). Softmatch: Addressing the quantity-quality trade-off in semi-supervised learning. arXiv.
|
|