Author:
KULKARNI SMITA SUNIL,Jadhav Sangeeta
Abstract
This paper is about recognizing multiple person actions occurring in videos, including individual actions, interactions,and group activities. In an environment, multiple people perform group actions such as walking in groupsand talking by facing each other. The model develops by retrieving individual person action from video sequencesby representing interactive contextual features among multiple people. The novelty of the proposed frameworkis the development of interactive action context descriptors (IAC) and classifying group activities using MachineLearning. Each individual person and other nearby people’s relative action score are encoded by IAC in thevideo frame. Individual person action descriptors are important clues for recognition of multiple person activityby developing interaction context. An action retrieval technique was formulated based on KNN for individualaction classification scores. This model also introduces Fully Connected Conditional Random Field (FCCRF) tolearn interaction context information among multiple people. FCCRF regularizes activity categorization by thespatial-temporal model. This paper also presents threshold processing to improve the performance of contextdescriptors. The experimental results compared to state-of-the-art approaches and demonstrated improvement inperformance for group activity recognition.
Publisher
Perpetual Innovation Media Pvt. Ltd.
Reference21 articles.
1. Ryoo, M. S., and J. K. Aggarwal. "Recognition of high-level group activities based on activities of individual members." In 2008 IEEE Workshop on Motion and video Computing, pp. 1-8. IEEE, 2008.
2. Choi, Wongun, Khuram Shahid, and Silvio Savarese. "What are they doing?: Collective activity classification using spatio-temporal relationship among people." In 2009 IEEE 12th international conference on computer vision workshops, ICCV Workshops, pp. 1282-1289. IEEE, 2009.
3. Lan, Tian, Yang Wang, Weilong Yang, Stephen N. Robinovitch, and Greg Mori. "Discriminative latent models for recognizing contextual group activities." IEEE transactions on pattern analysis and machine intelligence 34, no. 8 (2011): 1549-1562.
4. Choi, Wongun, Khuram Shahid, and Silvio Savarese. "Learning context for collective activity recognition." In CVPR 2011, pp. 3273-3280. IEEE, 2011.
5. Kaneko, Takuhiro, Masamichi Shimosaka, Shigeyuki Odashima, Rui Fukui, and Tomomasa Sato. "Viewpoint invariant collective activity recognition with relative action context." In European Conference on Computer Vision, pp. 253-262. Springer, Berlin, Heidelberg, 2012.