TLDR: ClaimCheck is an LLM-guided fact-checking system that uses small language models (like Qwen3-4B) and live web evidence to verify claims. It achieves state-of-the-art accuracy by breaking down fact-checking into modular, interpretable steps, making it more accessible and computationally efficient than systems relying on large, closed-source models. The system’s transparent design and strong performance with smaller models offer a promising direction for democratizing fact verification.
In an era where misinformation spreads rapidly across digital platforms, the need for accessible and reliable fact-checking tools has never been more critical. While human fact-checking remains the gold standard, its time-intensive nature limits its scalability. Recent advancements in automated fact-checking often rely on large, expensive, and closed-source language models, posing significant barriers to widespread adoption.
Introducing ClaimCheck: Fact-Checking for Everyone
A new system called ClaimCheck emerges as a promising solution, offering an LLM-guided automatic fact-checking system designed to verify real-world claims using live web evidence and, notably, small language models. Developed by Akshith Reddy Putta, Jacob Devasier, and Chengkai Li from the University of Texas at Arlington, ClaimCheck aims to democratize access to trustworthy fact-checking by demonstrating that effective verification can be achieved with substantially smaller and more accessible models than previously thought.
Unlike prior systems that depend on static knowledge stores, ClaimCheck employs a transparent, stepwise verification pipeline that closely mirrors how humans fact-check. This process includes planning web search queries, retrieving and summarizing web-based evidence, synthesizing evidence, and finally evaluating the claim to deliver a verdict. Each module within ClaimCheck is optimized for small LLLMs, allowing the system to provide accurate and interpretable fact-checking with significantly lower computational requirements.
How ClaimCheck Works: A Step-by-Step Approach
ClaimCheck operates through five core stages, making its reasoning process clear and understandable:
- Query Planning: The system first analyzes the claim and constructs targeted web search queries to gather relevant evidence.
- Evidence Retrieval: Using these queries, ClaimCheck searches the web (via Google search engine, limited to top-3 results for speed) to collect relevant articles and snippets.
- Evidence Summarization: A small language model then individually extracts and summarizes the key information from each collected article, discarding irrelevant content.
- Evidence Synthesis: All summarized evidence is then combined into a single, coherent analysis. This stage identifies patterns, highlights key facts, and assesses the overall reliability of the information. If gaps in evidence are found, the system can perform additional web searches.
- Claim Evaluation: Finally, the system uses the synthesized evidence to determine the veracity of the claim, assigning a verdict (Supported, Refuted, Conflicting Evidence/Cherrypicking, or Not Enough Evidence) along with a clear explanation.
This modular design not only facilitates internal coordination but also generates a comprehensive fact-checking report for users, detailing every step of the reasoning process. This transparency is crucial for building trust and allowing users to understand how a verdict was reached.
Performance and Efficiency
Despite utilizing a much smaller Qwen3-4B model, ClaimCheck achieves a state-of-the-art accuracy of 76.4% on the A VeriTeC dataset. This performance surpasses previous approaches that relied on significantly larger models like LLaMA3.1 70B and GPT-4o, which often used pre-fetched knowledge stores rather than live web evidence. The research highlights that careful modular design and prompting strategies can effectively overcome the limitations typically associated with smaller LLMs.
The system also boasts impressive computational efficiency. ClaimCheck with Qwen3-4B can process each claim in approximately 66 seconds on an H100 GPU, making it a practical solution for real-time applications. The ‘thinking’ mechanism, which enables comprehensive reasoning at each step, adds minimal overhead.
Also Read:
- Fact Grounded Attention: A New Approach to Reliable LLMs
- Generalized Models for Accurate LLM Correctness Prediction
Ethical Considerations and Future Directions
The authors acknowledge potential risks, such as the system inadvertently retrieving and summarizing low-quality or misleading sources from the live web. To mitigate this, ClaimCheck provides transparent reporting, allowing users to trace each verdict back to its evidence. They also caution against over-reliance, emphasizing that the system is experimental and should not replace human judgment in high-stakes domains.
ClaimCheck represents a significant step towards making automated fact-checking more accessible, transparent, and computationally efficient. By proving that powerful fact-checking can be done with smaller, more manageable AI models, it opens new avenues for combating misinformation on a broader scale. For more details, you can read the full research paper here.


