Contrastive-regularized U-Net for video anomaly detection
Document Type
Article
Publication Date
1-1-2023
Abstract
Video anomaly detection aims to identify anomalous segments in a video. It is typically trained with weakly supervised video-level labels. This paper focuses on two crucial factors affecting the performance of video anomaly detection models. First, we explore how to capture the local and global temporal dependencies more effectively. Previous architectures are effective at capturing either local and global information, but not both. We propose to employ a U-Net like structure to model both types of dependencies in a unified structure where the encoder learns global dependencies hierarchically on top of local ones; then the decoder propagates this global information back to the segment level for classification. Second, overfitting is a non-trivial issue for video anomaly detection due to limited training data. We propose weakly supervised contrastive regularization which adopts a feature-based approach to regularize the network. Contrastive regularization learns more generalizable features by enforcing inter-class separability and intra-class compactness. Extensive experiments on the UCF-Crime dataset shows that our approach outperforms several state-of-the-art methods.
Keywords
Feature extraction, Deep learning, Anomaly detection, Image reconstruction, Image segmentation, Supervised learning, Training data, video anomaly detection, weakly supervised learning, contrastive-based regularization, multi-instance learning, deep learning
Divisions
sch_ecs
Funders
Fundamental Research Grant Scheme (FRGS) through the Ministry of Higher Education (MOHE) of Malaysia (FRGS/1/2018/ICT02/UTAR/02/03)
Publication Title
IEEE Access
Volume
11
Publisher
Institute of Electrical and Electronics Engineers
Publisher Location
445 HOES LANE, PISCATAWAY, NJ 08855-4141 USA