MVACLNet: A Multimodal Virtual Augmentation Contrastive Learning Network for Rumor Detection
-
Published:2024-05-08
Issue:5
Volume:17
Page:199
-
ISSN:1999-4893
-
Container-title:Algorithms
-
language:en
-
Short-container-title:Algorithms
Author:
Liu Xin1, Pang Mingjiang1ORCID, Li Qiang2, Zhou Jiehan3ORCID, Wang Haiwen1, Yang Dawei1
Affiliation:
1. Qingdao Institute of Software, College of Computer Science and Technology, China University of Petroleum (East China), No. 66, West Changjiang Road, Huangdao District, Qingdao 266580, China 2. Scientific and Technological Innovation Center of ARI, Beijing 100020, China 3. Information Technology and Electrical Engineering, University of Oulu, 90570 Oulu, Finland
Abstract
In today’s digital era, rumors spreading on social media threaten societal stability and individuals’ daily lives, especially multimodal rumors. Hence, there is an urgent need for effective multimodal rumor detection methods. However, existing approaches often overlook the insufficient diversity of multimodal samples in feature space and hidden similarities and differences among multimodal samples. To address such challenges, we propose MVACLNet, a Multimodal Virtual Augmentation Contrastive Learning Network. In MVACLNet, we first design a Hierarchical Textual Feature Extraction (HTFE) module to extract comprehensive textual features from multiple perspectives. Then, we fuse the textual and visual features using a modified cross-attention mechanism, which operates from different perspectives at the feature value level, to obtain authentic multimodal feature representations. Following this, we devise a Virtual Augmentation Contrastive Learning (VACL) module as an auxiliary training module. It leverages ground-truth labels and extra-generated virtual multimodal feature representations to enhance contrastive learning, thus helping capture more crucial similarities and differences among multimodal samples. Meanwhile, it performs a Kullback–Leibler (KL) divergence constraint between predicted probability distributions of the virtual multimodal feature representations and their corresponding virtual labels to help extract more content-invariant multimodal features. Finally, the authentic multimodal feature representations are input into a rumor classifier for detection. Experiments on two real-world datasets demonstrate the effectiveness and superiority of MVACLNet on multimodal rumor detection.
Funder
Natural Science Foundation of Shandong Province Key Research and Development (R&D) Plan of Shandong Province
Reference53 articles.
1. An overview of online fake news: Characterization, detection, and discussion;Zhang;Inf. Process. Manag.,2020 2. An exploration of how fake news is taking over social media and putting public health at risk;Naeem;Health Inf. Libr. J.,2021 3. Castillo, C., Mendoza, M., and Poblete, B. (April, January 28). Information credibility on twitter. Proceedings of the 20th International World Wide Web Conference, Hyderabad, India. 4. Zhao, Z., Resnick, P., and Mei, Q. (2015, January 18–22). Enquiring Minds: Early Detection of Rumors in Social Media from Enquiry Posts. Proceedings of the 24th International World Wide Web Conference, Florence, Italy. 5. Jin, Z., Cao, J., Zhang, Y., and Luo, J. (2016, January 12–17). News verification by exploiting conflicting social viewpoints in microblogs. Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA.
|
|