Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks

Author:

Naik Aakanksha12,Lehman Jill3,Rosé Carolyn45

Affiliation:

1. Language Technologies Institute, Carnegie Mellon University, USA. anaik@andrew.cmu.edu

2. Rehabilitation Medicine Department, Clinical Center, National Institutes of Health, USA

3. Human-Computer Interaction Institute, Carnegie Mellon University, USA. jfl@andrew.cmu.edu

4. Language Technologies Institute, Carnegie Mellon University, USA

5. Rehabilitation Medicine Department, Clinical Center, National Institutes of Health, USA. cp3a@andrew.cmu.edu

Abstract

Abstract Natural language understanding (NLU) has made massive progress driven by large benchmarks, but benchmarks often leave a long tail of infrequent phenomena underrepresented. We reflect on the question: Have transfer learning methods sufficiently addressed the poor performance of benchmark-trained models on the long tail? We conceptualize the long tail using macro-level dimensions (underrepresented genres, topics, etc.), and perform a qualitative meta-analysis of 100 representative papers on transfer learning research for NLU. Our analysis asks three questions: (i) Which long tail dimensions do transfer learning studies target? (ii) Which properties of adaptation methods help improve performance on the long tail? (iii) Which methodological gaps have greatest negative impact on long tail performance? Our answers highlight major avenues for future research in transfer learning for the long tail. Lastly, using our meta-analysis framework, we perform a case study comparing the performance of various adaptation methods on clinical narratives, which provides interesting insights that may enable us to make progress along these future avenues.

Publisher

MIT Press

Subject

Artificial Intelligence,Computer Science Applications,Linguistics and Language,Human-Computer Interaction,Communication

Reference100 articles.

1. Model adaptation for personalized opinion analysis;Boni,2015

2. Domain adaptation with adversarial training and graph embeddings;Alam,2018

3. Exploiting feature hierarchy for transfer learning in named entity recognition;Arnold,2008

4. On achieving and evaluating language-independence in NLP;Bender;Linguistic Issues in Language Technology,2011

5. Biographies, Bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification;Blitzer,2007

Cited by 2 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Boosting Long-Tail Data Classification with Sparse Prototypical Networks;Lecture Notes in Computer Science;2024

2. General then Personal: Decoupling and Pre-training for Personalized Headline Generation;Transactions of the Association for Computational Linguistics;2023

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3