TLDR: METER is a new framework and benchmark addressing the urgent need for interpretable forgery detection across images, videos, and audio. Unlike previous methods that offer only binary classification, METER provides detailed explanations, including spatio-temporal localization and forgery type tracing. It introduces a comprehensive dataset covering both digital and physical attacks, along with novel evaluation metrics. The framework also proposes a three-stage, human-aligned training strategy (SFT, DPO, GRPO) to cultivate trustworthy and explainable reasoning, aiming to advance reliable media forensics in the age of generative AI.
In an era where generative artificial intelligence (AI) is rapidly advancing, creating incredibly realistic synthetic content across images, videos, and audio, the risk of misinformation has escalated dramatically. While existing methods can often tell if something is fake, they usually fall short in explaining *how* or *why* it’s fake. This lack of detailed, interpretable explanations limits their use in critical areas like journalism, law enforcement, and finance, where trust and actionable evidence are paramount.
Furthermore, most current detection techniques focus on one type of media at a time, without a unified way to detect and explain forgeries across different modalities. To tackle these significant challenges, a new framework called METER has been introduced. METER stands for Multi-modal Evidence-based Thinking and Explainable Reasoning, and it offers a unified benchmark and a novel approach for interpretable forgery detection across images, videos, audio, and even combined audio-visual content.
What METER Brings to the Table
METER is designed to provide a comprehensive solution that goes beyond simple ‘real vs. fake’ classification. It aims to answer three fundamental questions for any piece of media: 1) Where is the forgery? (Localization), 2) Why is it a forgery? (Explanation), and 3) How was it forged? (Traceability).
The METER dataset is unique because it’s the first to bring together image, video, audio, and audio-visual modalities under a single framework for explainable forgery detection. It also uniquely covers both digital attacks (like AI-generated images or voice cloning) and physical attacks (such as re-capturing a printed photo or replaying audio/video on a screen). This broad coverage ensures that the system can handle a wide range of real-world forgery scenarios.
For evaluation, METER introduces a robust set of metrics. These include spatio-temporal Intersection over Union (IoU) to precisely locate the forged regions in space and time, multi-class traceability accuracy to identify the specific method used to create the forgery, and a novel evidence rationality score to assess how logical and convincing the generated explanations are to humans.
Also Read:
- HICOM: A New AI Framework Detects Multi-Face Deepfakes by Learning from Human Perception
- AdViT: Unmasking Vulnerabilities in Interpretable Vision Transformer Systems
A Human-Aligned Training Approach
To ensure that the explanations generated by the system are not only accurate but also understandable and trustworthy to humans, METER employs an innovative three-stage training strategy. This strategy is designed to systematically build the model’s ability to reason and explain its findings:
- Stage 1: Supervised Fine-Tuning (SFT): This initial stage teaches the model the basics, allowing it to learn from high-quality, human-annotated data to generate structured outputs, including classification labels, localization details, and evidence chains.
- Stage 2: Direct Preference Optimization (DPO): Here, the model learns to align its explanations with human judgment. By training on pairs of preferred and less-preferred explanations, it refines its ability to generate responses that are more intuitive and convincing to a human user.
- Stage 3: Group Relative Policy Optimization (GRPO) with Chain-of-Thought: This final stage significantly enhances the model’s reasoning capabilities. It uses a specially built evaluation model to provide feedback on the rationality of the generated evidence. The model learns by comparing multiple generated explanations for the same input, ensuring it produces high-quality, forensically accurate explanations.
The METER framework represents a significant leap forward in the fight against misinformation. By focusing on explainability, multi-modality, and human-aligned reasoning, it sets a new standard for trustworthy media forensics. This work provides a standardized foundation for developing more generalizable and interpretable forgery detection technologies, which is crucial in our increasingly digital world. For more in-depth information, you can read the full research paper here.


