Enhancing Cross-Lingual Sarcasm Detection by a Prompt Learning Framework with Data Augmentation and Contrastive Learning

Author:

An Tianbo123,Yan Pingping123,Zuo Jiaai4,Jin Xing56,Liu Mingliang1,Wang Jingrui123ORCID

Affiliation:

1. College of Computer Science and Technology, Changchun University, Changchun 130022, China

2. Key Laboratory of Intelligent Rehabilitation and Barrier-Free for the Disabled, Ministry of Education, Changchun University, Changchun 130022, China

3. Jilin Provincial Key Laboratory of Human Health Status Identification & Function Enhancement, Changchun 130022, China

4. College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China

5. School of Cyberspace, Hangzhou Dianzi University, Hangzhou 310018, China

6. Experimental Center of Data Science and Intelligent Decision-Making, Hangzhou Dianzi University, Hangzhou 310018, China

Abstract

Given their intricate nature and inherent ambiguity, sarcastic texts often mask deeper emotions, making it challenging to discern the genuine feelings behind the words. The proposal of the sarcasm detection task is to assist us with more accurately understanding the true intention of the speaker. Advanced methods, such as deep learning and neural networks, are widely used in the field of sarcasm detection. However, most research mainly focuses on sarcastic texts in English, as other languages lack corpora and annotated datasets. To address the challenge of low-resource languages in sarcasm detection tasks, a zero-shot cross-lingual transfer learning method is proposed in this paper. The proposed approach is based on prompt learning and aims to assist the model with understanding downstream tasks through prompts. Specifically, the model uses prompt templates to construct training data into cloze-style questions and then trains them using a pre-trained cross-lingual language model. Combining data augmentation and contrastive learning can further improve the capacity of the model for cross-lingual transfer learning. To evaluate the performance of the proposed model, we utilize a publicly accessible sarcasm dataset in English as training data in a zero-shot cross-lingual setting. When tested with Chinese as the target language for transfer, our model achieves F1-scores of 72.14% and 76.7% on two test datasets, outperforming the strong baselines by significant margins.

Funder

Jilin Provincial Department of Science and Technology

Education Department of Jilin Province

Publisher

MDPI AG

Reference68 articles.

1. Maynard, D.G., and Greenwood, M.A. (2014, January 26–31). Who cares about sarcastic tweets? Investigating the impact of sarcasm on sentiment analysis. Proceedings of the Lrec 2014 Proceedings, ELRA, Reykjavik, Iceland.

2. Merriam-Webster, I. (1995). The Merriam-Webster Dictionary, Merriam-Webster.

3. Context-based feature technique for sarcasm identification in benchmark datasets using deep learning and BERT model;Eke;IEEE Access,2021

4. Sentiment and sarcasm classification with multitask learning;Majumder;IEEE Intell. Syst.,2019

5. Ensemble transfer learning-based multimodal sentiment analysis using weighted convolutional neural networks;Ghorbanali;Inf. Process. Manag.,2022

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3