Skip to content
GCC AI Research

Search

Results for "animation"

Cross-modal understanding and generation of multimodal content

MBZUAI ·

Nicu Sebe from the University of Trento presented recent work on video generation, focusing on animating objects in a source image using external information like labels, driving videos, or text. He introduced a Learnable Game Engine (LGE) trained from monocular annotated videos, which maintains states of scenes, objects, and agents to render controllable viewpoints. Why it matters: This talk highlights advancements in cross-modal AI, potentially enabling new applications in gaming, simulation, and content creation within the region.

Reconstruction and Animation of Realistic Head Avatars

MBZUAI ·

Egor Zakharov from ETH Zurich AIT lab will present research on creating controllable and detailed 3D head avatars using data from consumer-grade devices. The presentation will cover high-fidelity image-based facial reconstruction/animation and video-based reconstruction of detailed structures like hairstyles. He will showcase integrating human-centric assets into virtual environments for real-time telepresence and entertainment. Why it matters: This research contributes to advancements in digital human modeling and telepresence, with applications in communication and gaming within the region.

Visualizing and experiencing science at WEP 2016

KAUST ·

KAUST's Winter Enrichment Program (WEP) 2016 featured a poster competition highlighting research by graduate students, postdocs, and international undergraduates. A science fair included shows by science podcast host Dr. Chris Smith, art exhibits, and a visualization lab. Exhibits included "On the Trail of the Glaciers: An Interactive Experience" and short films produced by KAUST students. Why it matters: Such programs foster scientific engagement and communication within the KAUST community and beyond.

Art exhibits at WEP 2015

KAUST ·

KAUST will host a Modern Saudi Art Exhibit from Arabian Wings (Jan 11-15), an Al-Balad 24 Photography Exhibition featuring work by Marina Kochetyga and Andrea Bachofen (Jan 11-16), and an East African Tingatinga art exhibition (Jan 18-24). The Al-Balad exhibit includes a video by Dr. Lorenzo Pareschi documenting a fire in the historic district. Why it matters: These art exhibits expose the KAUST community to diverse artistic styles and cultural perspectives, fostering cross-cultural understanding.

An exhilarating experience

KAUST ·

KAUST's Discovery Week featured a gala and awards ceremony. Professor Gilles Lubineau opened the proceedings at the 2017 WEP Final Gala. A Javanese shadow puppet performance of the “Ramayana Epic” was also part of the event. Why it matters: Showcases KAUST's commitment to cultural exchange alongside its research activities.

Biweekly research update

KAUST ·

Professor Arnab Pain's group at KAUST discovered new insights on how a malaria protein enables parasites to spread malaria in human cells. Professor Haavard Rue's group upgraded the Integrated and Nested Laplace Approximation (INLA) for faster real-time modeling of large datasets. A KAUST-led study examined the stability of Y-series nonfullerene acceptors for organic solar cells. Why it matters: KAUST continues producing impactful research across diverse fields from medicine to climate change, advancing scientific knowledge and potential applications.

Is Human Motion a Language without Words?

MBZUAI ·

This article previews a talk by Gül Varol from Ecole des Ponts ParisTech on bridging natural language and 3D human motions. The talk will cover text-to-motion synthesis using generative models and text-to-motion retrieval models based on the ACTOR, TEMOS, TMR, TEACH, and SINC papers. Varol's research interests include video representation learning, human motion synthesis, and sign languages. Why it matters: Research in this area could enable more intuitive human-computer interaction and new applications in areas like virtual reality and robotics.

FancyVideo: Towards Dynamic and Consistent Video Generation via Cross-frame Textual Guidance

arXiv ·

FancyVideo, a new video generator, introduces a Cross-frame Textual Guidance Module (CTGM) to enhance text-to-video models. CTGM uses a Temporal Information Injector and Temporal Affinity Refiner to achieve frame-specific textual guidance, improving comprehension of temporal logic. Experiments on the EvalCrafter benchmark demonstrate FancyVideo's state-of-the-art performance in generating dynamic and consistent videos, also supporting image-to-video tasks.