Author:
Wu Lifang,Lang Xianglong,Xiang Ye,Wang Qi,Tian Meng
Abstract
Group activity recognition that infers the activity of a group of people is a challenging task and has received a great deal of interest in recent years. Different from individual action recognition, group activity recognition needs to model not only the visual cues of individuals but also the relationships between them. The existing approaches inferred relations based on the holistic features of the individual. However, parts of the human body, such as the head, hands, legs, and their relationships, are the critical cues in most group activities. In this paper, we establish the part-based graphs from different viewpoints. The intra-actor part graph is designed to model the spatial relations of different parts for an individual, and the inter-actor part graph is proposed to explore part-level relations among actors, in which visual relation and location relation are both considered. Furthermore, a two-branch framework is utilized to capture the static spatial and dynamic temporal representations simultaneously. On the Volleyball Dataset, our approach obtains a classification accuracy of 94.8%, achieving very competitive performance in comparison with the state of the art. As for the Collective Activity Dataset, our approach improves the accuracy by 0.3% compared with the state-of-the-art results.
Funder
Chaoyang District Postdoctoral Research Foundation
National Natural Science Foundation of China
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference62 articles.
1. HiGCIN: Hierarchical Graph-based Cross Inference Network for Group Activity Recognition
2. stagNet: An Attentive Semantic RNN for Group Activity and Individual Action Recognition
3. A Comprehensive Review of Group Activity Recognition in Videos
4. Cost-sensitive top-down/bottom-up inference for multiscale activity recognition;Amer;Proceedings of the European Conference on Computer Vision,2012
5. Hirf: Hierarchical random field for collective activity recognition in videos;Amer;Proceedings of the European Conference on Computer Vision,2014
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献