论文标题
LGN-NET:视频异常检测的本地全球正常网络
LGN-Net: Local-Global Normality Network for Video Anomaly Detection
论文作者
论文摘要
视频异常检测(VAD)多年来一直在深入研究,因为它在智能视频系统中的潜在应用。现有的无监督VAD方法倾向于从仅由正常视频组成的训练集学习正态性,并考虑偏离异常范围的实例。但是,他们通常仅考虑时间维度的本地或全球正态性。他们中的一些人专注于学习从连续帧的本地时空表示,以增强正常事件的表示。但是强大的表示允许这些方法表示某些异常并导致错过检测。相比之下,其他方法致力于记住整个训练视频的原型正常模式,以削弱异常的概括,这也限制了它们代表不同的正常模式并引起错误警报。为此,我们提出了一个两个分支机构模型,局部 - 全球正态网络(LGN-NET),以同时学习本地和全球正常性。具体而言,一个分支通过时空预测网络了解了连续框架的外观和运动的演变规律性,而另一个分支则将整个视频的原型特征作为内存模块的全局态度记住。 LGN-NET通过融合本地和全球正常性来取得代表正常和异常实例的平衡。此外,融合的正态性使LGN-NET能够概括到各种场景,而不是利用单个正态性。实验证明了我们方法的有效性和卓越性能。该代码可在线提供:https://github.com/myzhao1999/lgn-net。
Video anomaly detection (VAD) has been intensively studied for years because of its potential applications in intelligent video systems. Existing unsupervised VAD methods tend to learn normality from training sets consisting of only normal videos and regard instances deviating from such normality as anomalies. However, they often consider only local or global normality in the temporal dimension. Some of them focus on learning local spatiotemporal representations from consecutive frames to enhance the representation for normal events. But powerful representation allows these methods to represent some anomalies and causes miss detection. In contrast, the other methods are devoted to memorizing prototypical normal patterns of whole training videos to weaken the generalization for anomalies, which also restricts them from representing diverse normal patterns and causes false alarm. To this end, we propose a two-branch model, Local-Global Normality Network (LGN-Net), to simultaneously learn local and global normality. Specifically, one branch learns the evolution regularities of appearance and motion from consecutive frames as local normality utilizing a spatiotemporal prediction network, while the other branch memorizes prototype features of the whole videos as global normality by a memory module. LGN-Net achieves a balance of representing normal and abnormal instances by fusing local and global normality. In addition, the fused normality enables LGN-Net to generalize to various scenes more than exploiting single normality. Experiments demonstrate the effectiveness and superior performance of our method. The code is available online: https://github.com/Myzhao1999/LGN-Net.