Abnormal Event Detection in Videos using Generative Adversarial Nets
About
In this paper we address the abnormality detection problem in crowded scenes. We propose to use Generative Adversarial Nets (GANs), which are trained using normal frames and corresponding optical-flow images in order to learn an internal representation of the scene normality. Since our GANs are trained with only normal data, they are not able to generate abnormal events. At testing time the real data are compared with both the appearance and the motion representations reconstructed by our GANs and abnormal areas are detected by computing local differences. Experimental results on challenging abnormality detection datasets show the superiority of the proposed method compared to the state of the art in both frame-level and pixel-level abnormality detection tasks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Abnormal Event Detection | UCSD Ped2 (test) | AUC93.5 | 146 | |
| Abnormal Event Detection | UCSD Ped2 | AUC93.5 | 132 | |
| Abnormal Event Detection | UCSD Ped1 (test) | Frame AUC97.4 | 33 | |
| Abnormal Event Detection | UCSD Ped1 | AUC0.974 | 28 | |
| Abnormal Event Detection | UMN dataset | Frame AUC (All Scenes)99 | 25 | |
| Video Novelty Detection | UCSD (test) | AUCROC0.935 | 14 | |
| Abnormal Event Detection | UMN Unusual Crowd Activity (test) | Frame-level AUC99 | 13 | |
| Abnormal Event Detection | UCSD Ped1 v1 (test) | AUC97.4 | 12 | |
| Abnormal Event Detection | UCSD Ped2 v1 (test) | AUC93.5 | 12 | |
| Abnormality Detection | UCSD Ped1 pixel-level | EER35 | 9 |