Capture Salient Historical Information: A Fast and Accurate Non-autoregressive Model for Multi-turn Spoken Language Understanding

Author:

Cheng Lizhi1ORCID,Jia Weijia2ORCID,Yang Wenmian3ORCID

Affiliation:

1. Shanghai Jiao Tong University, Minhang District, Shanghai, PR China

2. BNU-UIC Institute of Artificial Intelligence and Future Networks, Beijing Normal University (Zhuhai), Guangdong Key Lab of AI and Multi-Modal Data Processing, BNU-HKBU United International College, Jintong Road, Tangjiawan, Zhuhai, Guangdong, PR China

3. Nanyang Technological University, Singapore City, Singapore

Abstract

Spoken Language Understanding (SLU), a core component of the task-oriented dialogue system, expects a shorter inference facing the impatience of human users. Existing work increases inference speed by designing non-autoregressive models for single-turn SLU tasks but fails to apply to multi-turn SLU in confronting the dialogue history. The intuitive idea is to concatenate all historical utterances and utilize the non-autoregressive models directly. However, this approach seriously misses the salient historical information and suffers from the uncoordinated-slot problems. To overcome those shortcomings, we propose a novel model for multi-turn SLU named Salient History Attention with Layer-Refined Transformer (SHA-LRT), which comprises a SHA module, a Layer-Refined Mechanism (LRM), and a Slot Label Generation (SLG) task. SHA captures salient historical information for the current dialogue from both historical utterances and results via a well-designed history-attention mechanism. LRM predicts preliminary SLU results from Transformer’s middle states and utilizes them to guide the final prediction, and SLG obtains the sequential dependency information for the non-autoregressive encoder. Experiments on public datasets indicate that our model significantly improves multi-turn SLU performance (17.5% on Overall) with accelerating (nearly 15 times) the inference process over the state-of-the-art baseline as well as effective on the single-turn SLU tasks.

Funder

Guangdong Key Lab of AI and Multi-modal Data Processing, United International College (UIC), Zhuhai

Chinese National Research Fund

Beijing Normal University (Zhuhai) Guangdong

Zhuhai Science-Tech Innovation Bureau

Publisher

Association for Computing Machinery (ACM)

Subject

Computer Science Applications,General Business, Management and Accounting,Information Systems

Reference56 articles.

1. Layer normalization;Ba Jimmy Lei;arXiv preprint arXiv:1607.06450,2016

2. Memory Consolidation for Contextual Spoken Language Understanding with Dialogue Logistic Inference

3. Sequential Dialogue Context Modeling for Spoken Language Understanding

4. Towards Zero-Shot Frame Semantic Parsing for Domain Scaling

5. Learning end-to-end goal-oriented dialog;Bordes Antoine;arXiv preprint arXiv:1605.07683,2016

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. A Survey on Non-Autoregressive Generation for Neural Machine Translation and Beyond;IEEE Transactions on Pattern Analysis and Machine Intelligence;2023

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3