Author:
Bornea Mihaela,Pan Lin,Rosenthal Sara,Florian Radu,Sil Avirup
Abstract
Prior work on multilingual question answering has mostly focused on using large multilingual pre-trained language models (LM) to perform zero-shot language-wise learning: train a QA model on English and test on other languages. In this work, we explore strategies that improve cross-lingual transfer by bringing the multilingual embeddings closer in the semantic space.
Our first strategy augments the original English training data with machine translation-generated data. This results in a corpus of multilingual silver-labeled QA pairs that is 14 times larger than the original training set. In addition, we propose two novel strategies, language adversarial training and language arbitration framework, which significantly improve the (zero-resource) cross-lingual transfer performance and result in LM embeddings that are less language-variant. Empirically, we show that the proposed models outperform the previous zero-shot baseline on the recently introduced multilingual MLQA and TyDiQA datasets.
Publisher
Association for the Advancement of Artificial Intelligence (AAAI)
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. From text to multimodal: a survey of adversarial example generation in question answering systems;Knowledge and Information Systems;2024-08-09
2. Multilingual Event Causality Identification via Meta-learning with Knowledge;Proceedings of the 2024 International Conference on Generative Artificial Intelligence and Information Security;2024-05-10
3. Data Augmentation for Conversational AI;Proceedings of the 32nd ACM International Conference on Information and Knowledge Management;2023-10-21
4. Multilingual Text Summarization for German Texts Using Transformer Models;Information;2023-05-25
5. Multi-Aspect Heterogeneous Graph Augmentation;Proceedings of the ACM Web Conference 2023;2023-04-30