Author:
Chen Haowei,Li Chen,Liang Jiajing,Tian Lihua
Abstract
AbstractWith the continuous advancement of social information, the number of texts in the form of dialogue between individuals has exponentially increased. However, it is very challenging to review the previous dialogue content before initiating a new conversation. In view of the above background, a new dialogue summarization algorithm based on multi-task learning is first proposed in the paper. Specifically, Minimum Risk Training is used as the loss function to alleviate the problem of inconsistent goals between the training phase and the testing phase. Then, in order to deal with the problem that the model cannot effectively distinguish gender pronouns, a gender pronoun discrimination auxiliary task based on contrast learning is designed to help the model learn to distinguish different gender pronouns. Finally, an auxiliary task of reducing exposure bias is introduced, which involves incorporating the summary generated during inference into another round of training to reduce the difference between the decoder inputs during the training and testing stages. Experimental results show that our model outperforms strong baselines on three public dialogue summarization datasets: SAMSUM, DialogSum, and CSDS.
Publisher
Springer Science and Business Media LLC
Reference27 articles.
1. Liu Z, Chen N (2019) Reading turn by turn: hierarchical attention architecture for spoken dialogue comprehension. In: Proceedings of the 57th annual meeting of the association for computational linguistics. pp 5460–5466
2. Gao S, Chen X, Ren Z, Zhao D, Yan R (2020) From standard summarization to new tasks and beyond: summarization with manifold information. arXiv:2005.04684
3. Park S, Lee J (2022) Unsupervised abstractive dialogue summarization with word graphs and POV conversion. arXiv:2205.13108
4. Liu Y, Liu P (2021) Simcls: A simple framework for contrastive learning of abstractive summarization. arXiv:2106.01890
5. Lee S, Behpour S, Eaton E (2021) Sharing less is more: lifelong learning in deep networks with selective layer transfer. In: International conference on machine learning. PMLR, pp 6065–6075