Skip to content
GCC AI Research

Search

Results for "video anomaly detection"

Detecting the undetectable: Transforming policing with AI

MBZUAI ·

Salem AlMarri, the first Emirati Ph.D. graduate from MBZUAI, developed a video anomaly detection (VAD) system for his thesis. The VAD system can detect subtle anomalies in video, such as suspicious interactions, to help police prevent crimes and save lives. AlMarri's work was carried out under the guidance of Karthik Nandakumar, Affiliated Associate Professor of Computer Vision at MBZUAI. Why it matters: This research showcases the potential of AI in enhancing public safety and security in the UAE, demonstrating practical applications of computer vision in law enforcement.

Making sense of space and time in video

MBZUAI ·

MBZUAI researchers presented a new approach to video analysis at ICCV in Paris, led by Syed Talal Wasim. The approach builds on still image processing techniques like focal modulation to analyze spatial and temporal information in video separately. It aims to improve temporal aggregation while avoiding the computational complexity of transformers. Why it matters: This research advances video understanding in computer vision by offering a more efficient method for temporal modeling, crucial for applications like activity recognition and video surveillance.

Teaching algorithms to see

KAUST ·

KAUST's Image and Video Understanding Lab is developing machine learning algorithms for computer vision and object tracking, with applications in video content search and UAV navigation. Their algorithms can detect specific activities in videos, helping platforms detect unwanted content and deliver relevant ads. The object tracking algorithm is also used to empower UAVs, enabling them to follow objects autonomously. Why it matters: This research enhances video content analysis and UAV capabilities, positioning KAUST as a leader in computer vision and AI applications within the region.

Unveiling Hidden Energy Anomalies: Harnessing Deep Learning to Optimize Energy Management in Sports Facilities

arXiv ·

This paper explores the use of deep learning for anomaly detection in sports facilities, with the goal of optimizing energy management. The researchers propose a method using Deep Feedforward Neural Networks (DFNN) and threshold estimation techniques to identify anomalies and reduce false alarms. They tested their approach on an aquatic center dataset at Qatar University, achieving 94.33% accuracy and 92.92% F1-score. Why it matters: The research demonstrates the potential of AI to improve energy efficiency and operational effectiveness in sports facilities within the GCC region.

Multimodal pretraining for objectionable content detection in videos

MBZUAI ·

Thamar Solorio from the University of Houston presented preliminary work on multimodal representation learning for detecting objectionable content in videos at MBZUAI. The research investigates two multimodal pretraining mechanisms, finding contrastive learning more effective than unimodal representation prediction. The study also assesses the value of common multimodal corpora for this task. Why it matters: This research contributes to the development of AI techniques for content moderation, an important issue for online platforms in the Middle East and globally.

FancyVideo: Towards Dynamic and Consistent Video Generation via Cross-frame Textual Guidance

arXiv ·

FancyVideo, a new video generator, introduces a Cross-frame Textual Guidance Module (CTGM) to enhance text-to-video models. CTGM uses a Temporal Information Injector and Temporal Affinity Refiner to achieve frame-specific textual guidance, improving comprehension of temporal logic. Experiments on the EvalCrafter benchmark demonstrate FancyVideo's state-of-the-art performance in generating dynamic and consistent videos, also supporting image-to-video tasks.

Video-ChatGPT: Towards Detailed Video Understanding via Large Vision and Language Models

arXiv ·

Video-ChatGPT is a new multimodal model that combines a video-adapted visual encoder with a large language model (LLM) to enable detailed video understanding and conversation. The authors introduce a new dataset of 100,000 video-instruction pairs for training the model. They also develop a quantitative evaluation framework for video-based dialogue models.

Old images to anticipate the future

MBZUAI ·

MBZUAI researchers presented a new approach to video question answering at ICCV 2023. The method leverages insights from analyzing still images to understand video content, potentially reducing the computational resources needed for training video question answering models. Guangyi Chen, Kun Zhang, and colleagues aim to apply pre-trained image models to understand video concepts. Why it matters: This research could lead to more efficient and accessible video analysis tools, benefiting fields like healthcare and security where video data is abundant.