Affiliation:
1. School of Electronics and Information Technology, Sun Yat-sen University, Guangzhou, China
Abstract
Multimodal sequence analysis aims to draw inferences from visual, language, and acoustic sequences. A majority of existing works focus on the aligned fusion of three modalities to explore inter-modal interactions, which is impractical in real-world scenarios. To overcome this issue, we seek to focus on analyzing unaligned sequences, which is still relatively underexplored and also more challenging. We propose Multimodal Graph, whose novelty mainly lies in transforming the sequential learning problem into graph learning problem. The graph-based structure enables parallel computation in time dimension (as opposed to recurrent neural network) and can effectively learn longer intra- and inter-modal temporal dependency in unaligned sequences. First, we propose multiple ways to construct the adjacency matrix for sequence to perform sequence to graph transformation. To learn intra-modal dynamics, a graph convolution network is employed for each modality based on the defined adjacency matrix. To learn inter-modal dynamics, given that the unimodal sequences are unaligned, the commonly considered word-level fusion does not pertain. To this end, we innovatively devise graph pooling algorithms to automatically explore the associations between various time slices from different modalities and learn high-level graph representation hierarchically. Multimodal Graph outperforms state-of-the-art models on three datasets under the same experimental setting.
Funder
National Natural Science Foundation of China
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications,Hardware and Architecture
Reference70 articles.
1. Shaojie Bai, J. Kolter, and Vladlen Koltun. 2019. Trellis networks for sequence modeling. In Proceedings of the International Conference on Learning Representations.
2. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling;Bai Shaojie;arXiv: 1803.01271,2018
3. Multimodal Machine Learning: A Survey and Taxonomy
4. Learning long-term dependencies with gradient descent is difficult
5. IEMOCAP: interactive emotional dyadic motion capture database
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献