Use of prompt-based learning for code-mixed and code-switched text classification
-
Published:2024-09
Issue:5
Volume:27
Page:
-
ISSN:1386-145X
-
Container-title:World Wide Web
-
language:en
-
Short-container-title:World Wide Web
Author:
Udawatta Pasindu,Udayangana Indunil,Gamage Chathulanka,Shekhar Ravi,Ranathunga Surangika
Abstract
AbstractCode-mixing and code-switching (CMCS) are prevalent phenomena observed in social media conversations and various other modes of communication. When developing applications such as sentiment analysers and hate-speech detectors that operate on this social media data, CMCS text poses challenges. Recent studies have demonstrated that prompt-based learning of pre-trained language models outperforms full fine-tuning across various tasks. Despite the growing interest in classifying CMCS text, the effectiveness of prompt-based learning for the task remains unexplored. This paper presents an extensive exploration of prompt-based learning for CMCS text classification and the first comprehensive analysis of the impact of the script on classifying CMCS text. Our study reveals that the performance in classifying CMCS text is significantly influenced by the inclusion of multiple scripts and the intensity of code-mixing. In response, we introduce a novel method, Dynamic+AdapterPrompt, which employs distinct models for each script, integrated with adapters. While DynamicPrompt captures the script-specific representation of the text, AdapterPrompt emphasizes capturing the task-oriented functionality. Our experiments on Sinhala-English, Kannada-English, and Hindi-English datasets for sentiment classification, hate-speech detection, and humour detection tasks show that our method outperforms strong fine-tuning baselines and basic prompting strategies.
Publisher
Springer Science and Business Media LLC
Reference46 articles.
1. Bali, K., Sharma, J., Choudhury, M., Vyas, Y.: I am borrowing ya mixing? an analysis of English-Hindi code mixing in Facebook. In: Diab, M., Hirschberg, J., Fung, P., Solorio, T. (eds.) Proceedings of the First Workshop on Computational Approaches to Code Switching, pp. 116–126. Association for Computational Linguistics, Doha, Qatar (2014). https://doi.org/10.3115/v1/W14-3914 . https://aclanthology.org/W14-3914 2. Gundapu, S., Mamidi, R.: Word level language identification in English Telugu code mixed data. In: Proceedings of the 32nd Pacific Asia Conference on Language, Information and Computation. Association for Computational Linguistics, Hong Kong (2018). https://aclanthology.org/Y18-1021 3. Zirker, K.A.H.: Intrasentential vs. intersentential code switching in early and late bilinguals (2007). https://api.semanticscholar.org/CorpusID:60154198 4. Hande, A., Puranik, K., Yasaswini, K., Priyadharshini, R., Thavareesan, S., Sampath, A., Shanmugavadivel, K., Thenmozhi, D., Chakravarthi, B.R.: Offensive Language Identification in Low-resourced Code-mixed Dravidian languages using Pseudo-labeling (2021). https://doi.org/10.48550/arXiv.2108.12177 5. Srivastava, V., Singh, M.: Code-mixed nlg: Resources, metrics, and challenges. In: 5th Joint international conference on data science & management of data (9th ACM IKDD CODS and 27th COMAD). CODS-COMAD 2022, pp. 328–332. Association for Computing Machinery, New York, NY, USA (2022). https://doi.org/10.1145/3493700.3493766
|
|