Affiliation:
1. Research Scholar, JJTU, Jhunjhunu, Rajasthan, India
2. Research Guide, JJTU Jhunjhunu, Rajasthan, India
3. Research Co-Guide, DYPCOE, Pune, Maharashtra, India
Abstract
With the use of deep learning algorithms from artificial intelligence (AI), several types of research have been conducted on video data. Object localization, behaviour analysis, scene understanding, scene labelling, human activity recognition (HAR), and event recognition make up the majority of them. Among all of them, HAR is one of the most difficult jobs and key areas of research in video data processing. HAR can be used in a variety of fields, including robotics, human-computer interaction, video surveillance, and human behaviour categorization. This research seeks to compare deep learning approaches on several benchmark video datasets for vision-based human activity detection. We suggest a brand-new taxonomy for dividing up the literature into CNN- and RNN-based methods. We further categorise these approaches into four subgroups and show several methodologies, their effectiveness, and experimental datasets. To illustrate the development of HAR techniques, a brief comparison is also provided with the handcrafted feature-based approach and its merger with deep learning. Finally, we go over potential future research areas and some unresolved issues with recognising human activities. This survey's goal is to present the most recent developments in HAR techniques for vision-based deep learning using the most recent literature analysis.
Subject
General Earth and Planetary Sciences,General Environmental Science
Reference25 articles.
1. Chen, D., P. Wang, L. Yue, Y. Zhang, and T. Jia. 2020. Anomaly Detection in Surveillance Video Based on Bidirectional Prediction. Image and Vision Computing 98:103915. doi:10.1016/j.imavis.2020.103915.
2. Wang, S., Y. Liu, J. Wang, S. Gao, and W. Yang. 2021. A Moving Track Data-Based Method for Gathering Behavior Prediction at Early Stage. Applied Intelligence 51 (11):8498–518. doi: 10.1007/s10489-021-02244-2.
3. Wenjin, T., Z. Hao Lai, M. C. Leu, and Z. Yin. 2018. Worker Activity Recognition in Smart Manufacturing Using IMU and SEMG Signals with Convolutional Neural Networks. Procedia Manufacturing 26:1159–66. Elsevier B.V. doi:10.1016/j.promfg.2018.07.152.
4. Deng, J., R. S. Wei Dong, L. Li-Jia, L. Kai, and L. Fei-Fei. 2009. ImageNet: A Large-Scale Hierarchical Image Database. 2009 IEEE Conference on Computer Vision and Pattern Recognition, 248–55. doi: 10.1109/cvprw.2009.5206848.
5. Kay, W., J. Carreira, K. Simonyan, B. Zhang, C. Hillier, S. Vijayanarasimhan, F. Viola, T. Green, T. Back, P. Natsev, M. Suleyman, and A. Zisserman. 2017. The Kinetics Human Action Video Dataset. ArXiv.