The recent developments in Large Multi-modal Video Models (Video-LMMs) have significantly enhanced our ability to interpret and analyze video data. Despite their impressive capabilities, current Video-LMMs have not been evaluated for anomaly detection tasks, which is critical to their deployment in practical scenarios e.g., towards identifying deepfakes, manipulated video content, traffic accidents and crimes. In this paper, we introduce VANE-Bench, a benchmark designed to assess the proficiency of Video-LMMs in detecting and localizing anomalies and inconsistencies in videos. Our dataset comprises an array of videos synthetically generated using existing state-of-the-art text-to-video generation models, encompassing a variety of subtle anomalies and inconsistencies grouped into five categories: unnatural transformations, unnatural appearance, pass-through, disappearance and sudden appearance. Additionally, our benchmark features real-world samples from existing anomaly detection datasets, focusing on crime-related irregularities, atypical pedestrian behavior, and unusual events. The task is structured as a visual question-answering challenge to gauge the models' ability to accurately detect and localize the anomalies within the videos. We evaluate nine existing Video-LMMs, both open and closed sources, on this benchmarking task and find that most of the models encounter difficulties in effectively identifying the subtle anomalies. In conclusion, our research offers significant insights into the current capabilities of Video-LMMs in the realm of anomaly detection, highlighting the importance of our work in evaluating and improving these models for real-world applications.
VANE-Bench is a challenging benchmark created to assess video anomaly detection within Large Multimodal Models (LMMs). It consists of 325 video clips, and 559 QA pairs from over nine different AI-generated and real-world video samples, each showcasing different types of anomalies. The subtle, and hard to detect anomalies present within VANE-Bench makes our benchmark challenging for all the SOTA LMMs, and even for few humans. Our benchmark enables the development and evaluation of stronger Video-LMMs for real-world applications like deepfake detection, crime detection, and traffic accidents identification.
Benchmark Category | Video-LLaMA | VideoChat | Video-ChatGPT | Video-LLaVA | MovieChat | LLaMA-VID | TimeChat | Gemini-1.5 Pro | GPT4o |
---|---|---|---|---|---|---|---|---|---|
SORA | 11.59 | 10.74 | 26.47 | 10.86 | 8.69 | 7.97 | 21.73 | 51.45 | 55.80 |
OpenSORA | 18.00 | 28.00 | 22.00 | 18.00 | 10.00 | 14.00 | 26.00 | 84.00 | 68.00 |
Runway Gen2 | 16.00 | 4.00 | 12.00 | 16.00 | 16.00 | 20.00 | 28.00 | 28.00 | 40.00 |
VideoLCM | 10.57 | 17.64 | 18.26 | 19.23 | 14.42 | 19.23 | 22.11 | 49.04 | 50.96 |
Modelscope-T2V | 10.41 | 20.83 | 16.66 | 16.66 | 6.25 | 14.58 | 20.83 | 75.00 | 64.58 |
Avenue | 30.00 | 32.25 | 39.39 | 3.03 | 18.18 | 27.27 | 24.20 | 100.00 | 84.85 |
UCFCrime | 9.47 | 11.57 | 31.57 | 10.52 | 18.51 | 15.78 | 7.30 | 76.84 | 83.16 |
UCSD-Ped1 | 16.66 | 13.33 | 40.00 | 2.77 | 6.66 | 6.66 | 27.58 | 96.67 | 93.33 |
UCSD-Ped2 | 5.55 | 13.88 | 19.44 | 6.06 | 11.11 | 19.44 | 11.11 | 94.44 | 86.11 |
@misc{bharadwaj2024vanebench,
title={VANE-Bench: Video Anomaly Evaluation Benchmark for Conversational LMMs},
author={Rohit Bharadwaj and Hanan Gani and Muzammal Naseer and Fahad Shahbaz Khan and Salman Khan},
year={2024},
eprint={2406.10326},
archivePrefix={arXiv},
primaryClass={id='cs.CV' full_name='Computer Vision and Pattern Recognition' is_active=True alt_name=None in_archive='cs' is_general=False description='Covers image processing, computer vision, pattern recognition, and scene understanding. Roughly includes material in ACM Subject Classes I.2.10, I.4, and I.5.'}
}