Middle East AI

This Week arXiv

A Case Study for Compliance as Code with Graphs and Language Models: Public release of the Regulatory Knowledge Graph

arXiv · · Significant research

Summary

This paper introduces a Regulatory Knowledge Graph (RKG) for the Abu Dhabi Global Market (ADGM) regulations, constructed using language models and graph technologies. A portion of the regulations was manually tagged to train BERT-based models, which were then applied to the rest of the corpus. The resulting knowledge graph, stored in Neo4j, and code are open-sourced on GitHub to promote advancements in compliance automation.

Keywords

Knowledge Graph · Compliance · BERT · ADGM · Regulations

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

RIRAG: Regulatory Information Retrieval and Answer Generation

arXiv ·

Researchers introduce a new task for generating question-passage pairs to aid in developing regulatory question-answering (QA) systems. The ObliQA dataset, comprising 27,869 questions from Abu Dhabi Global Markets (ADGM) financial regulations, is presented. A baseline Regulatory Information Retrieval and Answer Generation (RIRAG) system is designed and evaluated using the RePASs metric.

Fact-Checking Complex Claims with Program-Guided Reasoning

arXiv ·

This paper introduces ProgramFC, a fact-checking model that decomposes complex claims into simpler sub-tasks using a library of functions. The model uses LLMs to generate reasoning programs and executes them by delegating sub-tasks, enhancing explainability and data efficiency. Experiments on fact-checking datasets demonstrate ProgramFC's superior performance compared to baseline methods, with publicly available code and data.

ILION: Deterministic Pre-Execution Safety Gates for Agentic AI Systems

arXiv ·

The paper introduces ILION, a deterministic execution gate designed to ensure the safety of autonomous AI agents by classifying proposed actions as either BLOCK or ALLOW. ILION uses a five-component cascade architecture that operates without statistical training, API dependencies, or labeled data. Evaluation against existing text-safety infrastructures demonstrates ILION's superior performance in preventing unauthorized actions, achieving an F1 score of 0.8515 with sub-millisecond latency.

How secure is AI-generated Code: A Large-Scale Comparison of Large Language Models

arXiv ·

A study compared the vulnerability of C programs generated by nine state-of-the-art Large Language Models (LLMs) using a zero-shot prompt. The researchers introduced FormAI-v2, a dataset of 331,000 C programs generated by these LLMs, and found that at least 62.07% of the generated programs contained vulnerabilities, detected via formal verification. The research highlights the need for risk assessment and validation when deploying LLM-generated code in production environments.