Analysis of Movement and Activities of Handball Players Using Deep Neural Networks
-
Published:2023-04-13
Issue:4
Volume:9
Page:80
-
ISSN:2313-433X
-
Container-title:Journal of Imaging
-
language:en
-
Short-container-title:J. Imaging
Author:
Host Kristina12, Pobar Miran12ORCID, Ivasic-Kos Marina12ORCID
Affiliation:
1. Faculty of Informatics and Digital Technologies, University of Rijeka, 51000 Rijeka, Croatia 2. Centre for Artificial Intelligence and Cybersecurity, University of Rijeka, 51000 Rijeka, Croatia
Abstract
This paper focuses on image and video content analysis of handball scenes and applying deep learning methods for detecting and tracking the players and recognizing their activities. Handball is a team sport of two teams played indoors with the ball with well-defined goals and rules. The game is dynamic, with fourteen players moving quickly throughout the field in different directions, changing positions and roles from defensive to offensive, and performing different techniques and actions. Such dynamic team sports present challenging and demanding scenarios for both the object detector and the tracking algorithms and other computer vision tasks, such as action recognition and localization, with much room for improvement of existing algorithms. The aim of the paper is to explore the computer vision-based solutions for recognizing player actions that can be applied in unconstrained handball scenes with no additional sensors and with modest requirements, allowing a broader adoption of computer vision applications in both professional and amateur settings. This paper presents semi-manual creation of custom handball action dataset based on automatic player detection and tracking, and models for handball action recognition and localization using Inflated 3D Networks (I3D). For the task of player and ball detection, different configurations of You Only Look Once (YOLO) and Mask Region-Based Convolutional Neural Network (Mask R-CNN) models fine-tuned on custom handball datasets are compared to original YOLOv7 model to select the best detector that will be used for tracking-by-detection algorithms. For the player tracking, DeepSORT and Bag of tricks for SORT (BoT SORT) algorithms with Mask R-CNN and YOLO detectors were tested and compared. For the task of action recognition, I3D multi-class model and ensemble of binary I3D models are trained with different input frame lengths and frame selection strategies, and the best solution is proposed for handball action recognition. The obtained action recognition models perform well on the test set with nine handball action classes, with average F1 measures of 0.69 and 0.75 for ensemble and multi-class classifiers, respectively. They can be used to index handball videos to facilitate retrieval automatically. Finally, some open issues, challenges in applying deep learning methods in such a dynamic sports environment, and direction for future development will be discussed.
Funder
Croatian Science Foundation “Automatic recognition of actions and activities in multimedia content from the sports domain” University of Rijeka
Subject
Electrical and Electronic Engineering,Computer Graphics and Computer-Aided Design,Computer Vision and Pattern Recognition,Radiology, Nuclear Medicine and imaging
Reference39 articles.
1. Acuna, D. (2017, January 4–9). Towards Real-Time Detection and Tracking of Basketball Players Using Deep Neural Networks. Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA. 2. Redmon, J., and Farhadi, A. (2017, January 21–26). YOLO9000: Better, Faster, Stronger. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA. 3. Bewley, A., Ge, Z., Ott, L., Ramos, F., and Upcroft, B. (2016, January 25–28). Simple Online and Realtime Tracking. Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), IEEE, Phoenix, AZ, USA. 4. Ramanathan, V., Huang, J., Abu-El-Haija, S., Gorban, A., Murphy, K., and Fei-Fei, L. (2016, January 27–30). Detecting Events and Key Actors in Multi-Person Videos. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, Las Vegas, NV, USA. 5. Zakharchenko, I. (2020). Basketball Pose-Based Action Recognition, Ukrainian Catholic University, Faculty of Applied Sciences.
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. FineTea: A Novel Fine-Grained Action Recognition Video Dataset for Tea Ceremony Actions;Journal of Imaging;2024-08-31 2. Feature Extraction and Personalized Sports Training for Athletes Using Variational Autoencoder (VAE);2024 International Conference on Machine Intelligence and Digital Applications;2024-05-30 3. Super Long Range CNN For Video Enhancement in Handball Action Recognition;2024 3rd International Conference on Artificial Intelligence For Internet of Things (AIIoT);2024-05-03 4. Collision Detection and Prevention for Automobiles using Machine Learning;2024 International Conference on Trends in Quantum Computing and Emerging Business Technologies;2024-03-22 5. Exploring Dynamic Scenes: 3D CNN-Based Activity Recognition;2024 11th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO);2024-03-14
|
|