Affiliation:
1. School of Information and Communication Engineering, Communication University of China, Beijing 100024, China
Abstract
Traditional broadcasting methods often result in fatigue and decision-making errors when dealing with complex and diverse live content. Current research on intelligent broadcasting primarily relies on preset rules and model-based decisions, which have limited capabilities for understanding emotional dynamics. To address these issues, this study proposed and developed an emotion-driven intelligent broadcasting system, EmotionCast, to enhance the efficiency of camera switching during live broadcasts through decisions based on multimodal emotion recognition technology. Initially, the system employs sensing technologies to collect real-time video and audio data from multiple cameras, utilizing deep learning algorithms to analyze facial expressions and vocal tone cues for emotion detection. Subsequently, the visual, audio, and textual analyses were integrated to generate an emotional score for each camera. Finally, the score for each camera shot at the current time point was calculated by combining the current emotion score with the optimal scores from the preceding time window. This approach ensured optimal camera switching, thereby enabling swift responses to emotional changes. EmotionCast can be applied in various sensing environments such as sports events, concerts, and large-scale performances. The experimental results demonstrate that EmotionCast excels in switching accuracy, emotional resonance, and audience satisfaction, significantly enhancing emotional engagement compared to traditional broadcasting methods.
Funder
National Key R&D Program of China
Reference38 articles.
1. Wang, C., Qiu, C., Xu, W., Zhu, W., Luo, H., Chen, L., Yu, Z., Fu, L., and Chen, X. (2023, January 12–15). Intelligent Directing System for Music Concert Scene Based on Visual and Auditory Information. Proceedings of the 2023 ACM International Conference on Interactive Media Experiences Workshops (IMXw’23), Nantes, France.
2. Artificial Intelligence in News Media: Current Perceptions and Future Outlook;Ceron;J. Media,2022
3. A Semantic and Emotion-Based Dual Latent Variable Generation Model for a Dialogue System;Yan;CAAI Trans. Intell. Technol.,2023
4. Qiu, K., Zhang, Y., Zhao, J., Zhang, S., Wang, Q., and Chen, F. (2024). A Multimodal Sentiment Analysis Approach Based on a Joint Chained Interactive Attention Mechanism. Electronics, 13.
5. Video Sentiment Analysis with Bimodal Information-Augmented Multi-Head Attention;Wu;Knowl.-Based Syst.,2022