Skip to content
GCC AI Research

Search

Results for "liberal arts"

chatGPT for generating questions and assessments based on accreditations

arXiv ·

This research explores the use of generative AI, specifically ChatGPT, to create student assessments that align with academic accreditation standards, such as those of the National Center for Academic Accreditation in Saudi Arabia and ABET. The study introduces a method for mapping verbs used in questions to educational outcomes, enabling AI to produce and validate accreditation-compliant questions. A survey of faculty members in Saudi universities showed high acceptance rates for AI-generated exam questions and AI assistance in editing existing questions.

KAUST developing AI education for personalized learning

KAUST ·

KAUST is developing AI-driven personalized learning and testing platforms to address STEM education resource gaps in Saudi Arabia. The project involves building an intelligent tutoring system in collaboration with Saudi high schools, the Ministry of Education, and SDAIA. The AI tutor, designed in a Socratic style, enhances learning through GenAI tutoring, including in Arabic, and supports teachers by generating test and homework problems. Why it matters: This initiative aims to prepare Saudi youth for future workforce demands and enhance educational outcomes, aligning with Saudi Vision 2030's goals for human capital development.

User-Centric Gender Rewriting

MBZUAI ·

NYU and NYU Abu Dhabi researchers are working on user-centric gender rewriting in NLP, especially for Arabic. They are building an Arabic Parallel Gender Corpus and developing models for gender rewriting tasks. The work aims to address representational harms caused by NLP systems that don't account for user preferences regarding grammatical gender. Why it matters: This research promotes fairness and inclusivity in Arabic NLP by enabling systems to generate gender-specific outputs based on user preferences, mitigating biases present in training data.

Evaluating Arabic Large Language Models: A Survey of Benchmarks, Methods, and Gaps

arXiv ·

This survey paper analyzes over 40 benchmarks used to evaluate Arabic large language models, categorizing them into Knowledge, NLP Tasks, Culture and Dialects, and Target-Specific evaluations. It identifies progress in benchmark diversity but also highlights gaps like limited temporal evaluation and cultural misalignment. The paper also examines methods for creating benchmarks, including native collection, translation, and synthetic generation. Why it matters: The survey provides a comprehensive reference for Arabic NLP research and offers recommendations for future benchmark development to better align with cultural contexts.