Affiliation:
1. School of Computing and Digital Technology, Birmingham City University, Birmingham B5 5JU, UK
Abstract
Human communication is predominantly expressed through speech and writing, which are powerful mediums for conveying thoughts and opinions. Researchers have been studying the analysis of human sentiments for a long time, including the emerging area of bimodal sentiment analysis in natural language processing (NLP). Bimodal sentiment analysis has gained attention in various areas such as social opinion mining, healthcare, banking, and more. However, there is a limited amount of research on bimodal conversational sentiment analysis, which is challenging due to the complex nature of how humans express sentiment cues across different modalities. To address this gap in research, a comparison of multiple data modality models has been conducted on the widely used MELD dataset, which serves as a benchmark for sentiment analysis in the research community. The results show the effectiveness of combining acoustic and linguistic representations using a proposed neural-network-based ensemble learning technique over six transformer and deep-learning-based models, achieving state-of-the-art accuracy.
Subject
Artificial Intelligence,Computer Science Applications,Information Systems,Management Information Systems
Reference96 articles.
1. Venkataramanan, K., and Rajamohan, H.R. (2019). Emotion recognition from speech. arXiv.
2. Hendler, J., and Mulvehill, A.M. (2016). Social Machines: The Coming Collision of Artificial Intelligence, Social Networking, and Humanity, Apress.
3. The data deluge: An e-science perspective;Hey;Grid Comput. Mak. Glob. Infrastruct. Real.,2003
4. Picard, R.W. (2000). Affective Computing, MIT Press.
5. Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep Learning, MIT Press.
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献