Author:
Yang Qi,Lu Tongwei,Zhou Huabing
Abstract
Temporal modeling is the key for action recognition in videos, but traditional 2D CNNs do not capture temporal relationships well. 3D CNNs can achieve good performance, but are computationally intensive and not well practiced on existing devices. Based on these problems, we design a generic and effective module called spatio-temporal motion network (SMNet). SMNet maintains the complexity of 2D and reduces the computational effort of the algorithm while achieving performance comparable to 3D CNNs. SMNet contains a spatio-temporal excitation module (SE) and a motion excitation module (ME). The SE module uses group convolution to fuse temporal information to reduce the number of parameters in the network, and uses spatial attention to extract spatial information. The ME module uses the difference between adjacent frames to extract feature-level motion patterns between adjacent frames, which can effectively encode motion features and help identify actions efficiently. We use ResNet-50 as the backbone network and insert SMNet into the residual blocks to form a simple and effective action network. The experiment results on three datasets, namely Something-Something V1, Something-Something V2, and Kinetics-400, show that it out performs state-of-the-arts motion recognition networks.
Funder
National Natural Science Foundation of China
Hubei Technology Innovation Project
Subject
General Physics and Astronomy
Reference46 articles.
1. Two-stream convolutional networks for action recognition in videos;Simonyan;Adv. Neural Inf. Process. Syst.,2014
2. Temporal segment networks: Towards good practices for deep action recognition;Wang;Comput. Vis.,2016
3. Towards Good Practices for Very Deep Two-Stream ConvNets;Wang;arXiv,2015
4. 3D Convolutional Neural Networks for Human Action Recognition
Cited by
10 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献