Improving Semi-Supervised Text Classification with Dual Meta-Learning

Author:

Li Shujie1ORCID,Yuan Guanghu2ORCID,Yang Min3ORCID,Shen Ying4ORCID,Li Chengming4ORCID,Xu Ruifeng5ORCID,Zhao Xiaoyan6ORCID

Affiliation:

1. Shenzhen Institutes of Advanced Technology, Chinese Academy of Science, Shenzhen, China

2. Shenzhen Institutes of Advanced Technology, Chinese Academy of Science, Shenzhen China

3. Shenzhen Institutes of Advanced Technology Chinese Academy of Sciences, Shenzhen, China

4. School of Intelligent Systems Engineering, Sun Yat-Sen University, Guangdong China

5. Harbin Institute of Technology (Shenzhen), Shenzhen, China

6. Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China

Abstract

The goal of semi-supervised text classification (SSTC) is to train a model by exploring both a small number of labeled data and a large number of unlabeled data, such that the learned semi-supervised classifier performs better than the supervised classifier trained on solely the labeled samples. Pseudo-labeling is one of the most widely used SSTC techniques, which trains a teacher classifier with a small number of labeled examples to predict pseudo labels for the unlabeled data. The generated pseudo-labeled examples are then utilized to train a student classifier, such that the learned student classifier can outperform the teacher classifier. Nevertheless, the predicted pseudo labels may be inaccurate, making the performance of the student classifier degraded. The student classifier may perform even worse than the teacher classifier. To alleviate this issue, in this paper, we introduce a dual meta-learning ( DML ) technique for semi-supervised text classification, which improves the teacher and student classifiers simultaneously in an iterative manner. Specifically, we propose a meta-noise correction method to improve the student classifier by proposing a Noise Transition Matrix (NTM) with meta-learning to rectify the noisy pseudo labels. In addition, we devise a meta pseudo supervision method to improve the teacher classifier. Concretely, we exploit the feedback performance from the student classifier to further guide the teacher classifier to produce more accurate pseudo labels for the unlabeled data. In this way, both teacher and student classifiers can co-evolve in the iterative training process. Extensive experiments on four benchmark datasets highlight the effectiveness of our DML method against existing state-of-the-art methods for semi-supervised text classification. We release our code and data of this paper publicly at https://github.com/GRIT621/DML.

Funder

National Key Research and Development Program of China

National Natural Science Foundation of China

Shenzhen Science and Technology Innovation Program

Shenzhen Basic Research Foundation

Publisher

Association for Computing Machinery (ACM)

Reference62 articles.

1. RECORD: Resource Constrained Semi-Supervised Learning under Distribution Shift

2. Semi-supervised Collaborative Filtering by Text-enhanced Domain Adaptation

3. Knowledge Preserving and Distribution Alignment for Heterogeneous Domain Adaptation

4. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results;Tarvainen Antti;NIPS,2017

5. Hengduo Li, Zuxuan Wu, Abhinav Shrivastava, and Larry S. Davis. 2022. Rethinking pseudo labels for semi-supervised object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36. 1314–1322.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3