Skip to content
GCC AI Research

Enhancing Human Touch in Healthcare: The Role of Generative AI and Multimodal Technologies

MBZUAI · Notable

Summary

Ehsan Hoque from the University of Rochester gave a talk at MBZUAI discussing how to integrate AI into healthcare to improve access and equity. He emphasized that technology should align with values and infrastructure, advocating for AI solutions developed through collaboration between computer scientists and healthcare professionals. Hoque presented examples like using AI to quantify movement disorders and improve empathy skills. Why it matters: This highlights the importance of human-centered AI development in the GCC region, particularly in sensitive sectors like healthcare, and MBZUAI's role in fostering such discussions.

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

Humanizing Technology with Assistive Augmentations

MBZUAI ·

This article discusses a talk on "Assistive Augmentation," designing human-computer interfaces to augment human abilities. Examples include 'AiSee' for blind users, 'Prospero' for memory training, and 'MuSS-Bits' for deaf users to feel music. Suranga Nanayakkara from the National University of Singapore will present the talk, highlighting insights from psychology, human-centered machine learning, and design thinking. Why it matters: Such assistive technologies can significantly improve the quality of life for individuals with disabilities and extend human capabilities.

Multimodal machine intelligence and its human-centered possibilities

MBZUAI ·

A panel discussion was hosted at MBZUAI in collaboration with the Manara Center for Coexistence and Dialogue. The discussion centered on the potential of multimodal machine intelligence for human-centered applications, particularly in health and wellbeing. USC Professor Shrikanth Narayanan spoke on creating trustworthy and inclusive AI that considers protected variables. Why it matters: This signals MBZUAI's interest in exploring ethical AI development and its applications for societal good, potentially driving research and policy initiatives in the region.

Foundations of Multisensory Artificial Intelligence

MBZUAI ·

Paul Liang from CMU presented on machine learning foundations for multisensory AI, discussing a theoretical framework for modality interactions. The talk covered cross-modal attention and multimodal transformer architectures, and applications in mental health, pathology, and robotics. Liang's research aims to enable AI systems to integrate and learn from diverse real-world sensory modalities. Why it matters: This highlights the growing importance of multimodal AI research and its potential for advancements across various sectors in the region, including healthcare and robotics.

MOTOR: Multimodal Optimal Transport via Grounded Retrieval in Medical Visual Question Answering

arXiv ·

This paper introduces MOTOR, a multimodal retrieval and re-ranking approach for medical visual question answering (MedVQA) that uses grounded captions and optimal transport to capture relationships between queries and retrieved context, leveraging both textual and visual information. MOTOR identifies clinically relevant contexts to augment VLM input, achieving higher accuracy on MedVQA datasets. Empirical analysis shows MOTOR outperforms state-of-the-art methods by an average of 6.45%.