Improving Intent Classification Using Unlabeled Data from Large Corpora
-
Published:2023-02-03
Issue:3
Volume:11
Page:769
-
ISSN:2227-7390
-
Container-title:Mathematics
-
language:en
-
Short-container-title:Mathematics
Author:
Bercaru Gabriel12ORCID, Truică Ciprian-Octavian12ORCID, Chiru Costin-Gabriel12ORCID, Rebedea Traian2ORCID
Affiliation:
1. SoftTehnica, RO-030128 Bucharest, Romania 2. Computer Science and Engineering Department, Faculty of Automatic Control and Computers, University Politehnica of Bucharest, RO-060042 Bucharest, Romania
Abstract
Intent classification is a central component of a Natural Language Understanding (NLU) pipeline for conversational agents. The quality of such a component depends on the quality of the training data, however, for many conversational scenarios, the data might be scarce; in these scenarios, data augmentation techniques are used. Having general data augmentation methods that can generalize to many datasets is highly desirable. The work presented in this paper is centered around two main components. First, we explore the influence of various feature vectors on the task of intent classification using RASA’s text classification capabilities. The second part of this work consists of a generic method for efficiently augmenting textual corpora using large datasets of unlabeled data. The proposed method is able to efficiently mine for examples similar to the ones that are already present in standard, natural language corpora. The experimental results show that using our corpus augmentation methods enables an increase in text classification accuracy in few-shot settings. Particularly, the gains in accuracy raise up to 16% when the number of labeled examples is very low (e.g., two examples). We believe that our method is important for any Natural Language Processing (NLP) or NLU task in which labeled training data are scarce or expensive to obtain. Lastly, we give some insights into future work, which aims at combining our proposed method with a semi-supervised learning approach.
Funder
Romanian Ministry of European Investments and Projects
Subject
General Mathematics,Engineering (miscellaneous),Computer Science (miscellaneous)
Reference34 articles.
1. A Comprehensive Analysis of Transformer-Deep Neural Network Models in Twitter Disaster Detection;Balakrishnan;Mathematics,2022 2. Larson, S., Mahendran, A., Peper, J.J., Clarke, C., Lee, A., Hill, P., Kummerfeld, J.K., Leach, K., Laurenzano, M.A., and Tang, L. (2019, January 3–7). An Evaluation Dataset for Intent Classification and Out-of-Scope Prediction. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China. 3. Casanueva, I., Temčinas, T., Gerz, D., Henderson, M., and Vulić, I. (2020, January 9). Efficient Intent Detection with Dual Sentence Encoders. Proceedings of the 2nd Workshop on Natural Language Processing for Conversational AI, Online. 4. Liu, X., Eshghi, A., Swietojanski, P., and Rieser, V. (2021). Benchmarking Natural Language Understanding Services for Building Conversational Agents, Springer. Lecture Notes in Electrical Engineering. 5. Ahmadvand, A., Choi, J.I., and Agichtein, E. (2019, January 21–25). Contextual Dialogue Act Classification for Open-Domain Conversational Agents. Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|