Skip to content
GCC AI Research

Search

Results for "Reasoning System"

Empowering Large Language Models with Reliable Reasoning

MBZUAI ·

Liangming Pan from UCSB presented research on building reliable generative AI agents by integrating symbolic representations with LLMs. The neuro-symbolic strategy combines the flexibility of language models with precise knowledge representation and verifiable reasoning. The work covers Logic-LM, ProgramFC, and learning from automated feedback, aiming to address LLM limitations in complex reasoning tasks. Why it matters: Improving the reliability of LLMs is crucial for high-stakes applications in finance, medicine, and law within the region and globally.

Intelligent, sovereign, explainable energy decisions: powered by open-source AI reasoning

MBZUAI ·

MBZUAI researchers have developed K2 Think, an open-source AI reasoning system for interpretable energy decisions. K2 Think uses long chain-of-thought supervised fine-tuning and reinforcement learning to improve accuracy on multi-step reasoning in complex energy problems. The system breaks down challenges into smaller, auditable steps and uses test-time scaling for real-time adaptation. Why it matters: The open-source nature of K2 Think promotes transparency, trust, and compliance in critical energy environments while allowing secure deployment on sovereign infrastructure.

What reinforcement learning can teach language models about reasoning

MBZUAI ·

MBZUAI researchers at the Institute of Foundation Models (IFM) investigated the role of reinforcement learning (RL) in improving reasoning abilities of language models. Their study found that RL acts as an 'elicitor' for reasoning in domains frequently encountered during pre-training (e.g., math, coding), while genuinely teaching new reasoning skills in underrepresented domains (e.g., logic, simulations). To support their analysis, they created a new dataset called GURU containing 92,000 examples across six domains. Why it matters: This research clarifies the impact of reinforcement learning on language model reasoning, paving the way for developing models with more generalizable reasoning abilities across diverse domains, an important direction for more capable AI systems.

MBZUAI and G42 Launch K2 Think: A Leading Open-Source System for Advanced AI Reasoning

MBZUAI ·

MBZUAI and G42 have launched K2 Think, an open-source AI system for advanced reasoning with 32 billion parameters. It outperforms reasoning models 20 times larger, employing techniques like long chain-of-thought fine-tuning and reinforcement learning. K2 Think will be available on Cerebras' platform, achieving 2,000 tokens per second, and ranks highly in math performance. Why it matters: This launch positions the UAE as a leader in AI innovation through public-private partnerships and open-source contributions, demonstrating that efficient AI design can rival larger models.

Web-Based Expert System for Civil Service Regulations: RCSES

arXiv ·

The paper introduces a web-based expert system called RCSES for civil service regulations in Saudi Arabia. The system covers 17 regulations and utilizes XML for knowledge representation and ASP.net for rule-based inference. RCSES was validated by domain experts and technical users, and compared favorably to other web-based expert systems.

MBZUAI launches K2 Think V2: UAE’s fully sovereign, next-generation reasoning system

MBZUAI ·

MBZUAI, G42, and Cerebras Systems have launched K2 Think V2, a 70-billion parameter reasoning system built on the K2-V2 base model. K2 Think V2 is fully open-source, from pre-training data to post-training alignment, ensuring transparency and reproducibility. It achieves leading results on complex reasoning benchmarks like AIME2025 and GPQA-Diamond. Why it matters: This release marks a significant advancement in the UAE's AI capabilities, demonstrating leadership in building globally accessible and fully sovereign AI systems focused on reasoning.

Fact-Checking Complex Claims with Program-Guided Reasoning

arXiv ·

This paper introduces ProgramFC, a fact-checking model that decomposes complex claims into simpler sub-tasks using a library of functions. The model uses LLMs to generate reasoning programs and executes them by delegating sub-tasks, enhancing explainability and data efficiency. Experiments on fact-checking datasets demonstrate ProgramFC's superior performance compared to baseline methods, with publicly available code and data.

Developing an AI system that thinks like a scientist

KAUST ·

KAUST researchers developed a new algorithm for detecting cause and effect in large datasets. The algorithm aims to find underlying models that generate data, helping uncover cause-and-effect dynamics. It could aid researchers across fields like cell biology and genetics by answering questions that typical machine learning cannot. Why it matters: This advancement could equip current machine learning methods with abilities to better deal with abstraction, inference, and concepts such as cause and effect.