Affiliation:
1. Columbia University, USA. teffland@cs.columbia.edu
2. Google Research, USA. mjcollins@google.com
Abstract
Abstract
We present Expected Statistic Regulariza tion (ESR), a novel regularization technique that utilizes low-order multi-task structural statistics to shape model distributions for semi- supervised learning on low-resource datasets. We study ESR in the context of cross-lingual transfer for syntactic analysis (POS tagging and labeled dependency parsing) and present several classes of low-order statistic functions that bear on model behavior. Experimentally, we evaluate the proposed statistics with ESR for unsupervised transfer on 5 diverse target languages and show that all statistics, when estimated accurately, yield improvements to both POS and LAS, with the best statistic improving POS by +7.0 and LAS by +8.5 on average. We also present semi-supervised transfer and learning curve experiments that show ESR provides significant gains over strong cross-lingual-transfer-plus-fine-tuning baselines for modest amounts of label data. These results indicate that ESR is a promising and complementary approach to model-transfer approaches for cross-lingual parsing.1
Subject
Artificial Intelligence,Computer Science Applications,Linguistics and Language,Human-Computer Interaction,Communication
Reference51 articles.
1. Cross-lingual dependency parsing with unlabeled auxiliary languages;Ahmad,2019
2. Many languages, one parser;Ammar;Transactions of the Association for Computational Linguistics,2016
3. Multiple instance learning networks for fine-grained sentiment analysis;Angelidis;Transactions of the Association for Computational Linguistics,2018
4. Learning extractors from unlabeled text using relevant databases;Bellare,2007
5. Transfer learning between related tasks using expected label proportions;Noach,2019
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献