1. 1) M. Ayadi, M. S. Kamel and F. Karray, ``Survey on speech emotion recognition: Features, classification schemes, and databases,'' Pattern Recognit., 44, 572-587 (2011).
2. 2) R. Cowie, E. Douglas-Cowie, N. Tsapatsoulis, G. Votsis, S. Kollias, W. Fellenz and J. G. Taylor, ``Emotion recognition in human-computer interaction,'' IEEE Signal Process. Mag., 18, 32-80 (2001).
3. 3) P. Gupta and N. Rajput, ``Two-stream emotion recognition for call center monitoring,'' Proc. Interspeech 2007, pp. 2241-2244 (2007).
4. 4) J. Adelhardt, R. Shi, C. Frank, V. Zeißler, A. Batliner, E. Nöth and H. Niemann, ``Multimodal user state recognition in a modern dialogue system,'' Proc. Annu. Conf. Artif. Intell., 2003, pp. 591-605 (2003).
5. 5) H. Ai, D. J. Litman, K. Forbes-Riley, M. Rotaru, J. Tetreault and A. Purandare, ``Using system and user performance features to improve emotion detection in spoken tutoring dialogs,'' Proc. Interspeech 2006, pp. 797-800 (2006).