Skip to content
GCC AI Research

Search

Results for "FairLib"

AI Startup Spotlight Series: LibrAI

MBZUAI ·

LibrAI is an AI startup founded by Xudong Han, a University of Melbourne PhD graduate, in December 2023 after the release of ChatGPT. The company focuses on advancing AI safety and responsible AI practices, building on Han's prior work creating FairLib, an open-source toolkit for fairness in deep neural networks. LibrAI's team of seven aims to create practical solutions ensuring AI is both responsible and revolutionary. Why it matters: The establishment of a startup focused on AI safety highlights growing awareness of ethical considerations in AI development within the region.

Highlighting LLM safety: How the Libra-Leaderboard is making AI more responsible

MBZUAI ·

MBZUAI-based startup LibrAI has launched the Libra-Leaderboard, an evaluation framework for LLMs that assesses both capability and safety. The leaderboard evaluates 26 mainstream LLMs using 57 datasets, assigning scores based on bias, misinformation, and oversensitivity. LibrAI also launched the Interactive Safety Arena to engage the public and educate them on AI safety through adversarial prompt testing. Why it matters: The Libra-Leaderboard provides a benchmark for responsible AI development, emphasizing the importance of aligning AI capabilities with safety considerations in the rapidly evolving LLM landscape.

OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs

arXiv ·

MBZUAI researchers release OpenFactCheck, a unified framework to evaluate the factual accuracy of large language models. The framework includes modules for response evaluation, LLM evaluation, and fact-checker evaluation. OpenFactCheck is available as an open-source Python library, a web service, and via GitHub.

ArabJobs: A Multinational Corpus of Arabic Job Ads

arXiv ·

The ArabJobs dataset is a new corpus of over 8,500 Arabic job advertisements collected from Egypt, Jordan, Saudi Arabia, and the UAE. The dataset contains over 550,000 words and captures linguistic, regional, and socio-economic variation in the Arab labor market. It is available on GitHub and can be used for fairness-aware Arabic NLP and labor market research.

DaringFed: A Dynamic Bayesian Persuasion Pricing for Online Federated Learning under Two-sided Incomplete Information

arXiv ·

This paper introduces DaringFed, a novel dynamic Bayesian persuasion pricing mechanism for online federated learning (OFL) that addresses the challenge of two-sided incomplete information (TII) regarding resources. It formulates the interaction between the server and clients as a dynamic signaling and pricing allocation problem within a Bayesian persuasion game, demonstrating the existence of a unique Bayesian persuasion Nash equilibrium. Evaluations on real and synthetic datasets demonstrate that DaringFed optimizes accuracy and convergence speed and improves the server's utility.

AI Safety Research

MBZUAI ·

Adel Bibi, a KAUST alumnus and researcher at the University of Oxford, presented his research on AI safety, covering robustness, alignment, and fairness of LLMs. The research addresses challenges in AI systems, alignment issues, and fairness across languages in common tokenizers. Bibi's work includes instruction prefix tuning and its theoretical limitations towards alignment. Why it matters: This research from a leading researcher highlights the importance of addressing safety concerns in LLMs, particularly regarding alignment and fairness in the Arabic language.

Making LLM accuracy a matter of fact

MBZUAI ·

MBZUAI NLP master's graduate Hasan Iqbal developed OpenFactCheck, a framework for fact-checking and evaluating the factual accuracy of large language models. The framework consists of three modules: ResponseEvaluator, LLMEvaluator, and CheckerEvaluator. OpenFactCheck was published at EMNLP 2024 and accepted at NAACL 2025 and COLING 2025, with Iqbal playing an active role at COLING in Abu Dhabi. Why it matters: The development of automated fact-checking frameworks is crucial for ensuring the reliability and trustworthiness of information generated by increasingly prevalent LLMs, especially in the Arabic-speaking world.

FAID: Fine-Grained AI-Generated Text Detection Using Multi-Task Auxiliary and Multi-Level Contrastive Learning

arXiv ·

MBZUAI researchers introduce FAID, a fine-grained AI-generated text detection framework capable of classifying text as human-written, LLM-generated, or collaboratively written. FAID utilizes multi-level contrastive learning and multi-task auxiliary classification to capture authorship and model-specific characteristics, and can identify the underlying LLM family. The framework outperforms existing baselines, especially in generalizing to unseen domains and new LLMs, and includes a multilingual, multi-domain dataset called FAIDSet.