Skip to content
GCC AI Research

Search

Results for "cultural diversity"

Advancing cultural diversity through AI

MBZUAI ·

MBZUAI is conducting research to improve cross-cultural understanding using AI, including studying LLM limitations in recognizing cultural references. They developed "Culturally Yours," a tool that helps users comprehend cultural references in text, and the "All Languages Matter Benchmark" (ALM Bench) to evaluate multimodal LLMs across 100 languages. MBZUAI has also developed LLMs tailored to low-resource languages like Jais (Arabic), Nanda (Hindi), and Sherkala (Kazakh). Why it matters: These initiatives promote inclusivity and ensure AI systems are culturally aware and can serve diverse populations effectively, particularly in the Middle East's multicultural context.

Cultural inclusivity in AI: A new benchmark dataset on 100 languages

MBZUAI ·

MBZUAI researchers have released ALM Bench, a new benchmark dataset for evaluating the performance of multimodal LLMs on cultural visual question-answer tasks across 100 languages. The dataset includes over 22,000 question-answer pairs across 19 categories, with a focus on low-resource languages and cultural nuances, including three Arabic dialects. They tested 16 open- and closed-source multimodal LLMs on it, revealing a significant need for greater cultural and linguistic inclusivity. Why it matters: The benchmark aims to improve the inclusivity of multimodal AI systems by addressing the underrepresentation of low-resource languages and cultural contexts.

Teaching language models about Arab culture through cross-cultural transfer

MBZUAI ·

MBZUAI researchers presented a method for cross-cultural transfer learning to improve language models' understanding of diverse Arab cultures. They used in-context learning and demonstration-based reinforcement (DITTO) to transfer cultural knowledge between countries. Experiments showed up to 34% improvement in performance on cultural understanding benchmarks using only a few demonstrations. Why it matters: This research addresses the gap in cultural understanding of Arabic language models, especially for smaller Arab countries, and provides a novel transfer learning approach.

A magical place

KAUST ·

Todd Nims, a filmmaker born in Saudi Arabia, premiered his film "Joud" at KAUST's 2018 Winter Enrichment Program. The film, set in Saudi Arabia, explores the cycle of life in reverse and the meaning of "Joud" (generosity in the face of scarcity). Nims describes Saudi Arabia as a "magical place" due to its rich storytelling tradition. Why it matters: The article highlights KAUST's role in showcasing cultural works and supporting Saudi artists, though the AI relevance is limited.

Culture and bias in LLMs: Defining the challenge and mitigating risks

MBZUAI ·

Researchers from MBZUAI, University of Washington, and other institutions presented studies at EMNLP 2024 exploring how LLMs represent cultures. A survey analyzed dozens of recent studies on LLMs and culture and proposes a new framework for future research. The survey found that there is no widely accepted definition of 'culture' in NLP, making it challenging to interpret how models represent culture through language. Why it matters: This highlights a key gap in the field and emphasizes the need for a more rigorous and consistent understanding of culture in AI, especially as LLMs become more globally integrated.

MBZUAI celebrates diversity with music, dance, and food – International Day 2021

MBZUAI ·

MBZUAI held its inaugural International Day celebrating the university's multicultural community with music, dance, food, and country pavilions. Over 300 staff, faculty, researchers, and students attended the event, which featured 17 student teams representing over 25 nations. The event included the International Day Booth Awards, with India winning first place. Why it matters: This event highlights MBZUAI's commitment to diversity and inclusion, fostering a global environment for AI research and development in the UAE.

Cultural awareness in AI: New visual question answering benchmark shared in oral presentation at NeurIPS

MBZUAI ·

MBZUAI researchers, in collaboration with over 70 researchers, have created the Culturally diverse Visual Question Answering (CVQA) benchmark to evaluate cultural understanding in multimodal LLMs. The CVQA dataset includes over 10,000 questions in 31 languages and 13 scripts, testing models on images of local dishes, personalities, and monuments. Testing of several multimodal LLMs on the CVQA benchmark revealed significant challenges, even for top models. Why it matters: This benchmark highlights the need for AI models to better understand diverse cultures, promoting fairness and relevance across different languages and regions.

SaudiCulture: A Benchmark for Evaluating Large Language Models Cultural Competence within Saudi Arabia

arXiv ·

The paper introduces SaudiCulture, a new benchmark for evaluating the cultural competence of LLMs within Saudi Arabia, covering five major geographical regions and diverse cultural domains. The benchmark includes questions of varying complexity and distinguishes between common and specialized regional knowledge. Evaluations of five LLMs (GPT-4, Llama 3.3, FANAR, Jais, and AceGPT) revealed performance declines on region-specific questions, highlighting the need for region-specific knowledge in LLM training.