Deep Crowd Anomaly Detection by Fusing Reconstruction and Prediction Networks
-
Published:2023-03-23
Issue:7
Volume:12
Page:1517
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Sharif Md. Haidar1ORCID, Jiao Lei1ORCID, Omlin Christian W.1ORCID
Affiliation:
1. Department of Information and Communication Technology, University of Agder, 4879 Grimstad, Norway
Abstract
Abnormal event detection is one of the most challenging tasks in computer vision. Many existing deep anomaly detection models are based on reconstruction errors, where the training phase is performed using only videos of normal events and the model is then capable to estimate frame-level scores for an unknown input. It is assumed that the reconstruction error gap between frames of normal and abnormal scores is high for abnormal events during the testing phase. Yet, this assumption may not always hold due to superior capacity and generalization of deep neural networks. In this paper, we design a generalized framework (rpNet) for proposing a series of deep models by fusing several options of a reconstruction network (rNet) and a prediction network (pNet) to detect anomaly in videos efficiently. In the rNet, either a convolutional autoencoder (ConvAE) or a skip connected ConvAE (AEc) can be used, whereas in the pNet, either a traditional U-Net, a non-local block U-Net, or an attention block U-Net (aUnet) can be applied. The fusion of both rNet and pNet increases the error gap. Our deep models have distinct degree of feature extraction capabilities. One of our models (AEcaUnet) consists of an AEc with our proposed aUnet has capability to confirm better error gap and to extract high quality of features needed for video anomaly detection. Experimental results on UCSD-Ped1, UCSD-Ped2, CUHK-Avenue, ShanghaiTech-Campus, and UMN datasets with rigorous statistical analysis show the effectiveness of our models.
Funder
Research Council of Norway
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference143 articles.
1. Hasan, M., Choi, J., Neumann, J., Chowdhury, A.K.R., and Davis, L.S. (2016, January 27–30). Learning Temporal Regularity in Video Sequences. Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA. 2. Liu, W., Luo, W., Lian, D., and Gao, S. (2018, January 18–23). Future Frame Prediction for Anomaly Detection—A New Baseline. Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA. 3. A cascade reconstruction model with generalization ability evaluation for anomaly detection in videos;Zhong;Pattern Recognit.,2022 4. Gong, D., Liu, L., Le, V., Saha, B., Mansour, M.R., Venkatesh, S., and van den Hengel, A. (November, January 27). Memorizing Normality to Detect Anomaly: Memory-Augmented Deep Autoencoder for Unsupervised Anomaly Detection. Proceedings of the International Conference on Computer Vision (ICCV), Seoul, Republic of Korea. 5. Park, H., Noh, J., and Ham, B. (2020, January 13–19). Learning Memory-Guided Normality for Anomaly Detection. Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|