Alexa Teacher Model

Author:

FitzGerald Jack1,Ananthakrishnan Shankar2,Arkoudas Konstantine3,Bernardi Davide4,Bhagia Abhishek5,Delli Bovi Claudio6,Cao Jin3,Chada Rakesh5,Chauhan Amit5,Chen Luoxin2,Dwarakanath Anurag7,Dwivedi Satyam7,Gojayev Turan6,Gopalakrishnan Karthik8,Gueudre Thomas6,Hakkani-Tur Dilek9,Hamza Wael3,Hüser Jonathan J.6,Jose Kevin Martin6,Khan Haidar3,Liu Beiye3,Lu Jianhua2,Manzotti Alessandro10,Natarajan Pradeep11,Owczarzak Karolina2,Oz Gokmen2,Palumbo Enrico12,Peris Charith2,Prakash Chandana Satya2,Rawls Stephen3,Rosenbaum Andy2,Shenoy Anjali7,Soltan Saleh3,Sridhar Mukund Harakere2,Tan Lizhen2,Triefenbach Fabian6,Wei Pan2,Yu Haiyang2,Zheng Shuai5,Tur Gokhan9,Natarajan Prem13

Affiliation:

1. Amazon, Denver, CO, USA

2. Amazon, Cambridge, MA, USA

3. Amazon, New York, NY, USA

4. Amazon, Turin, AA, USA

5. Amazon, Seattle, WA, USA

6. Amazon, Aachen, Germany

7. Amazon, Bangalore, India

8. Amazon, Santa Clara, CA, USA

9. Amazon, Sunnyvale, CA, USA

10. Amazon, Turin, Italy

11. Amazon, Chicago, IL, USA

12. Spotify, Turin, Italy

13. Amazon, Los Angeles, CA, USA

Publisher

ACM

Reference55 articles.

1. Armen Aghajanyan , Anchit Gupta , Akshat Shrivastava , Xilun Chen , Luke Zettlemoyer , and Sonal Gupta . 2021 . Muppet: Massive Multi-task Representations with Pre-Finetuning. , 5799--5811 pages. https://doi.org/10.18653/v1/2021.emnlp-main.468 Armen Aghajanyan, Anchit Gupta, Akshat Shrivastava, Xilun Chen, Luke Zettlemoyer, and Sonal Gupta. 2021. Muppet: Massive Multi-task Representations with Pre-Finetuning. , 5799--5811 pages. https://doi.org/10.18653/v1/2021.emnlp-main.468

2. Jimmy Ba and Rich Caruana . 2014 . Do Deep Nets Really Need to be Deep?. In Advances in Neural Information Processing Systems, Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K. Q . Weinberger (Eds.) , Vol. 27 . Curran Associates, Inc. https://proceedings.neurips.cc/paper/ 2014/file/ea8fcd92d59581717e06eb187f10666d-Paper.pdf Jimmy Ba and Rich Caruana. 2014. Do Deep Nets Really Need to be Deep?. In Advances in Neural Information Processing Systems, Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K. Q. Weinberger (Eds.), Vol. 27. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2014/file/ea8fcd92d59581717e06eb187f10666d-Paper.pdf

3. Tom Brown , Benjamin Mann , Nick Ryder , Melanie Subbiah , Jared D Kaplan , Prafulla Dhariwal , Arvind Neelakantan , Pranav Shyam , Girish Sastry , Amanda Askell , Sandhini Agarwal , Ariel Herbert-Voss , Gretchen Krueger , Tom Henighan , Rewon Child , Aditya Ramesh , Daniel Ziegler , Jeffrey Wu , Clemens Winter , Chris Hesse , Mark Chen , Eric Sigler , Mateusz Litwin , Scott Gray , Benjamin Chess , Jack Clark , Christopher Berner , Sam McCandlish , Alec Radford , Ilya Sutskever , and Dario Amodei . 2020. Language Models are Few-Shot Learners . , Vol. 33 ( 2020 ), 1877--1901. https://proceedings.neurips.cc/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language Models are Few-Shot Learners. , Vol. 33 (2020), 1877--1901. https://proceedings.neurips.cc/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf

4. Cristian Buciluundefined , Rich Caruana , and Alexandru Niculescu-Mizil . 2006 . Model Compression. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining ( Philadelphia, PA, USA) (KDD '06). Association for Computing Machinery, New York, NY, USA, 535--541. https://doi.org/10.1145/1150402.1150464 Cristian Buciluundefined, Rich Caruana, and Alexandru Niculescu-Mizil. 2006. Model Compression. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (Philadelphia, PA, USA) (KDD '06). Association for Computing Machinery, New York, NY, USA, 535--541. https://doi.org/10.1145/1150402.1150464

5. Jin Cao Jun Wang Wael Hamza Kelly Vanee and Shang-Wen Li. 2020. Style Attuned Pre-training and Parameter Efficient Fine-tuning for Spoken Language Understanding. (2020). Jin Cao Jun Wang Wael Hamza Kelly Vanee and Shang-Wen Li. 2020. Style Attuned Pre-training and Parameter Efficient Fine-tuning for Spoken Language Understanding. (2020).

Cited by 6 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. A Lightweight and Effective Multi-View Knowledge Distillation Framework for Text-Image Retrieval;2024 International Joint Conference on Neural Networks (IJCNN);2024-06-30

2. A criteria-based classification model using augmentation and contrastive learning for analyzing imbalanced statement data;Heliyon;2024-06

3. Initial Development and Performance Evaluation of a Bengali Voice-Operated Virtual Assistant for Personal Computer Control;2023 IEEE 64th International Scientific Conference on Information Technology and Management Science of Riga Technical University (ITMS);2023-10-05

4. A Mixed-Methods Approach to Understanding User Trust after Voice Assistant Failures;Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems;2023-04-19

5. Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing;Transactions of the Association for Computational Linguistics;2023

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3