Affiliation:
1. Ministry of Sports, Xi’an Academy of Fine Arts, Xi’an 710065, Shaanxi, China
2. School of Sciences, Xi’an Technological University, Xi’an 710021, Shaanxi, China
Abstract
The team sports game video features complex background, fast target movement, and mutual occlusion between targets, which poses great challenges to multiperson collaborative video analysis. This paper proposes a video semantic extraction method that integrates domain knowledge and in-depth features, which can be applied to the analysis of a multiperson collaborative basketball game video, where the semantic event is modeled as an adversarial relationship between two teams of players. We first designed a scheme that combines a dual-stream network and learnable spatiotemporal feature aggregation, which can be used for end-to-end training of video semantic extraction to bridge the gap between low-level features and high-level semantic events. Then, an algorithm based on the knowledge from different video sources is proposed to extract the action semantics. The algorithm gathers local convolutional features in the entire space-time range, which can be used to track the ball/shooter/hoop to realize automatic semantic extraction of basketball game videos. Experiments show that the scheme proposed in this paper can effectively identify the four categories of short, medium, long, free throw, and scoring events and the semantics of athletes’ actions based on the video footage of the basketball game.
Subject
Computer Science Applications,Software
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. A novel multi-modal feature extraction system for news video;International Conference on Automation Control, Algorithm, and Intelligent Bionics (ACAIB 2023);2023-08-10
2. Deep Learning Algorithm-Based Target Detection and Fine Localization of Technical Features in Basketball;Computational Intelligence and Neuroscience;2022-05-23