Skip to content
GCC AI Research

Search

Results for "VinAI"

Image generation and manipulation research at VinAI

MBZUAI ·

VinAI Research presented research projects focused on advancing image generation and manipulation using GANs and Diffusion Models. The research aims to improve GANs regarding utility, coverage, and output consistency. For Diffusion Models, the work focuses on improving the models’ speed to approach real-time performance and prevent negative social impact of diffusion-based personalized text-to-image generation. Why it matters: This talk indicates ongoing research and development in generative AI in Southeast Asia, an area of growing interest globally.

G42 and Vietnamese Consortium Commit to Build National AI Infrastructure and Develop Southeast Asia’s Intelligence Capacity

G42 ·

G42 and a Vietnamese consortium (FPT Corporation and Viet Thai Group) have signed an agreement to develop AI capabilities and cloud infrastructure across Vietnam. The initiative is backed by consumption commitments of up to $1 billion and aims to support Vietnam's ambition to become a leading AI hub in Southeast Asia. The partnership will deploy cloud capacity across three data center locations in Vietnam to support public and private sector workloads. Why it matters: This marks a significant expansion of G42's footprint in Southeast Asia, fostering AI infrastructure development while respecting national data sovereignty.

To Make Just-Noticeable Difference (JND) Computable toward Visual Intelligence

MBZUAI ·

A professor from Nanyang Technological University (NTU), Singapore gave a talk at MBZUAI about "Just-Noticeable Difference (JND)" models in visual intelligence. The talk covered visual JND models, research and applications, and future opportunities for JND modeling. JND can help tackle big data challenges with limited resources by focusing on user-centric and green systems. Why it matters: Exploring JND could lead to advancements in AI applications related to visual signal processing, image synthesis, and generative AI in the region.

Multimodality for story-level understanding and generation of visual data

MBZUAI ·

Vicky Kalogeiton from École Polytechnique discussed the importance of multimodality for story-level recognition and generation using video, audio, text, masks and clinical data. She presented on multimodal video understanding using FunnyNet-W and Short Film Dataset. She further showed examples of visual generation from text and other modalities (ET, CAD, DynamicGuidance). Why it matters: Multimodal AI research is growing globally, and this talk highlights the potential of combining different data types for enhanced understanding and generation, which could have implications for various applications, including those relevant to the Middle East.

Super-aligned Machine Intelligence via a Soft Touch

MBZUAI ·

Song Chaoyang from the Southern University of Science and Technology (SUSTech) presented research on Vision-Based Tactile Sensing (VBTS) for robot learning, combining soft robotic design with learning algorithms to achieve state-of-the-art performance in tactile perception. Their VBTS solution demonstrates robustness up to 1 million test cycles and enables multi-modal outputs from a single, vision-based input, facilitating applications such as amphibious tactile grasping and industrial welding. The talk also highlighted the DeepClaw system for capturing human demonstration actions, aiming for a universal interaction interface. Why it matters: This research advances embodied intelligence by improving robot dexterity and adaptability through enhanced tactile sensing, which is crucial for complex manipulation tasks in various sectors such as manufacturing and healthcare within the region.

Unlocking the Potential of Large Models for Vision Related Tasks

MBZUAI ·

Yanwei Fu from Fudan University will present research on multimodal models, robotic grasping, and fMRI neural decoding. Topics include few-shot learning, object-centered self-supervised learning, image manipulation, and visual-language alignment. The research also covers Transformer compression and applications of large models with MVS 3D modeling in robotic arm grasping. Why it matters: While the talk is not directly about Middle East AI, the topics covered are core to advancing AI research and applications in the region.

AI that's built to save lives

KAUST ·

A KAUST team led by Xin Gao developed an AI model for COVID-19 detection from CT scans, addressing limitations of existing methods. The model incorporates a novel embedding strategy, a CT scan simulator, and a 2.5D deep-learning algorithm. Tested at King Faisal Specialist Hospital, the model demonstrated high accuracy in detecting COVID-19 cases. Why it matters: This research provides a valuable tool for rapid and accurate COVID-19 diagnosis in the region, especially in early-stage infections, improving healthcare outcomes.