Affiliation:
1. Xinjiang Technical Institute of Physics & Chemistry, Chinese Academy of Sciences, China and University of Chinese Academy of Sciences, China and Xinjiang Laboratory of Minority Speech and Language Information Processing, China
2. Xinjiang Technical Institute of Physics & Chemistry, Chinese Academy of Sciences, China and University of Chinese Academy of Sciences, China and Xinjiang Laboratory of Minority Speech and Language Information Processing, China
Abstract
Multilingual pre-trained language models (mPLMs) have achieved remarkable performance on zero-shot cross-lingual transfer learning. However, most mPLMs implicitly encourage cross-lingual alignment in pre-training stage, making it hard to capture accurate word alignment across languages. In this paper, we propose Word-align ADapters for Cross-lingual transfer (WAD-X) to explicitly align word representations of mPLMs via language-specific subspace. Taking a mPLM as the backbone model, WAD-X constructs subspace for each source-target language pair via adapters. The adapters use statistical alignment as the prior knowledge to guide word-level aligning in the corresponding bilingual semantic subspace. We evaluate our model across a set of target languages on three zero-shot cross-lingual transfer tasks: part-of-speech tagging (POS), dependency parsing (DP), and sentiment analysis (SA). Experimental results demonstrate that our proposed model improves zero-shot cross-lingual transfer on three benchmarks, with improvements of 2.19, 2.50, and 1.61 points in POS, DP, and SA tasks over strong baselines.
Funder
Youth Innovation Promotion Association of Chinese Academy of Sciences
National Natural Science Foundation of China
Tianshan Innovative Research Team of Xinjiang
Natural Science Foundation of Xinjiang Uyghur Autonomous Region
Tianshan Elite” Science and Technology Innovation Leading Talents Program
West Light Foundation of Chinese Academy of Sciences
Key Research and Development Program of Xinjiang Uyghur Autonomous Region
Publisher
Association for Computing Machinery (ACM)
Reference68 articles.
1. Mikel Artetxe, Gorka Labaka, and Eneko Agirre. 2018. Generalizing and improving bilingual word embedding mappings with a multi-step framework of linear transformations. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th Innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018, Sheila A. McIlraith and Kilian Q. Weinberger (Eds.). AAAI Press, 5012–5019. https://www.aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16935
2. On the Cross-lingual Transferability of Monolingual Representations
3. Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond
4. TRSAv1: A new benchmark dataset for classifying user reviews on Turkish e-commerce websites;Aydoğan Murat;Journal of Information Science,2022
5. Simple, Scalable Adaptation for Neural Machine Translation
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献