Skip to content
GCC AI Research

The cost of truth: An efficient fact-checking framework | NAACL

MBZUAI · Significant research

Summary

MBZUAI researchers presented FIRE, a new fact-checking framework for LLM outputs, at NAACL 2025. FIRE first assesses the LLM's confidence in its claims before searching the web, reducing computational cost. It also stores knowledge gained from web searches to aid in classifying other claims. Why it matters: This approach improves the efficiency and cost-effectiveness of automatically verifying the accuracy of LLMs, addressing a key limitation in their reliability.

Keywords

fact-checking · LLM · MBZUAI · NAACL · FIRE

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

FIRE: Fact-checking with Iterative Retrieval and Verification

arXiv ·

A novel agent-based framework called FIRE is introduced for fact-checking long-form text. FIRE iteratively integrates evidence retrieval and claim verification, deciding whether to provide a final answer or generate a subsequent search query. Experiments show FIRE achieves comparable performance to existing methods while reducing LLM costs by 7.6x and search costs by 16.5x.

Fact-Checking Complex Claims with Program-Guided Reasoning

arXiv ·

This paper introduces ProgramFC, a fact-checking model that decomposes complex claims into simpler sub-tasks using a library of functions. The model uses LLMs to generate reasoning programs and executes them by delegating sub-tasks, enhancing explainability and data efficiency. Experiments on fact-checking datasets demonstrate ProgramFC's superior performance compared to baseline methods, with publicly available code and data.

OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs

arXiv ·

MBZUAI researchers release OpenFactCheck, a unified framework to evaluate the factual accuracy of large language models. The framework includes modules for response evaluation, LLM evaluation, and fact-checker evaluation. OpenFactCheck is available as an open-source Python library, a web service, and via GitHub.

UrduFactCheck: An Agentic Fact-Checking Framework for Urdu with Evidence Boosting and Benchmarking

arXiv ·

Researchers from MBZUAI have introduced UrduFactCheck, a new framework for fact-checking in Urdu, along with two datasets: UrduFactBench and UrduFactQA. The framework uses monolingual and translation-based evidence retrieval to address the lack of Urdu resources. Evaluations using twelve LLMs showed that translation-augmented methods improve performance, highlighting challenges for open-source LLMs in Urdu.