A Chinese Few-Shot Text Classification Method Utilizing Improved Prompt Learning and Unlabeled Data
-
Published:2023-03-06
Issue:5
Volume:13
Page:3334
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Hu Tingkai1ORCID, Chen Zuqin2, Ge Jike1, Yang Zhaoxu1, Xu Jichao1
Affiliation:
1. College of Intelligent Technology and Engineering, Chongqing University of Science and Technology, No. 20 University East Road, Chongqing 401331, China 2. School of Library, Chongqing University of Science and Technology, No. 20 University East Road, Chongqing 401331, China
Abstract
Insufficiently labeled samples and low-generalization performance have become significant natural language processing problems, drawing significant concern for few-shot text classification (FSTC). Advances in prompt learning have significantly improved the performance of FSTC. However, prompt learning methods typically require the pre-trained language model and tokens of the vocabulary list for model training, while different language models have different token coding structures, making it impractical to build effective Chinese prompt learning methods from previous approaches related to English. In addition, a majority of current prompt learning methods do not make use of existing unlabeled data, thus often leading to unsatisfactory performance in real-world applications. To address the above limitations, we propose a novel Chinese FSTC method called CIPLUD that combines an improved prompt learning method and existing unlabeled data, which are used for the classification of a small amount of Chinese text data. We used the Chinese pre-trained language model to build two modules: the Multiple Masks Optimization-based Prompt Learning (MMOPL) module and the One-Class Support Vector Machine-based Unlabeled Data Leveraging (OCSVM-UDL) module. The former generates prompt prefixes with multiple masks and constructs suitable prompt templates for Chinese labels. It optimizes the random token combination problem during label prediction with joint probability and length constraints. The latter, by establishing an OCSVM model in the trained text vector space, selects reasonable pseudo-label data for each category from a large amount of unlabeled data. After selecting the pseudo-label data, we mixed them with the previous few-shot annotated data to obtain brand new training data and then repeated the steps of the two modules as an iterative semi-supervised optimization process. The experimental results on the four Chinese FSTC benchmark datasets demonstrate that our proposed solution outperformed other prompt learning methods with an average accuracy improvement of 2.3%.
Funder
National Social Science Foundation Western Project of China
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference40 articles.
1. Survey on supervised machine learning techniques for automatic text classification;Kadhim;Artif. Intell. Rev.,2019 2. Chen, W., Xu, Z., Zheng, X., Yu, Q., and Luo, Y. (2020). Research on Sentiment Classification of Online Travel Review Text. Appl. Sci., 10. 3. Xu, G., Liu, P., Zhu, Z., Liu, J., and Xu, F. (2021). Attention-Enhanced Graph Convolutional Networks for Aspect-Based Sentiment Classification with Multi-Head Attention. Appl. Sci., 11. 4. Wang, Y., Guo, J., Yuan, C., and Li, B. (2022). Sentiment Analysis of Twitter Data. Appl. Sci., 12. 5. A new similarity measure for vector space models in text classification and information retrieval;Eminagaoglu;J. Inf. Sci.,2022
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|