Affiliation:
1. Department of Telecommunication Engineering, University of Engineering and Technology, Taxila, Pakistan
2. Department of Computer Engineering, University of Engineering and Technology, Taxila, Pakistan
3. Swarm Robotic Lab-National Centre for Robotics and Automation (NCRA), University of Engineering and Technology, Taxila, Pakistan
Abstract
In the current era of technological development, human actions can be recorded in public places like airports, shopping malls, and educational institutes, etc., to monitor suspicious activities like terrorism, fighting, theft, and vandalism. Surveillance videos contain adequate visual and motion information for events that occur within a camera’s view. Our study focuses on the concept that actions are a sequence of moving body parts. In this paper, a new descriptor is proposed that formulates human poses and tracks the relative motion of human body parts along with the video frames, and extracts the position and orientation of body parts. We used Part Affinity Fields (PAFs) to acquire the associated body parts of the people present in the frame. The architecture jointly learns the body parts and their associations with other body parts in a sequential process, such that a pose can be formulated step by step. We can obtain the complete pose with a limited number of points as it moves along the video and we can conclude with a defined action. Later, these feature points are classified with a Support Vector Machine (SVM). The proposed work was evaluated on the benchmark datasets, namely, UT-interaction, UCF11, CASIA, and HCA datasets. Our proposed scheme was evaluated on the aforementioned datasets, which contained criminal/suspicious actions, such as kick, punch, push, gun shooting, and sword-fighting, and achieved an accuracy of 96.4% on UT-interaction, 99% on UCF11, 98% on CASIA and 88.72% on HCA.
Funder
Higher Education Commission, Pakistan
Subject
General Engineering,General Mathematics
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Identification of Human Activity by Utilizing YOLOv5s Approaches;2024 IEEE 3rd International Conference on Computing and Machine Intelligence (ICMI);2024-04-13
2. Temporal Relations of Informative Frames in Action Recognition;2024 13th Iranian/3rd International Machine Vision and Image Processing Conference (MVIP);2024-03-06
3. A Deep Autoencoder-Based Approach for Suspicious Action Recognition in Surveillance Videos;Arabian Journal for Science and Engineering;2023-07-08
4. Lightweight CNN and GRU Network for Real-Time Action Recognition;2022 12th International Conference on Pattern Recognition Systems (ICPRS);2022-06-07