TLDR: MIND is a novel multi-agent AI framework designed for zero-shot harmful meme detection, eliminating the need for annotated training data. It employs three key strategies: Similar Sample Retrieval to find contextual memes, a bi-directional Relevant Insight Derivation mechanism for comprehensive understanding, and an Insight-Augmented Inference stage with a multi-agent debate for robust decision-making. Experiments demonstrate that MIND significantly outperforms existing zero-shot methods and shows strong generalization across various Large Multimodal Models, offering a scalable and adaptable solution for identifying evolving harmful content on social media.
The rapid spread of memes across social media platforms has brought to light a pressing need for effective ways to identify harmful content. Traditional methods, which rely heavily on large amounts of pre-labeled data, often struggle to keep up with the ever-changing nature of memes and the constant emergence of new ones. This challenge makes it difficult to detect harmful memes quickly and efficiently.
To address this issue, researchers have introduced a new framework called MIND, which stands for Multi-agent Insight Derivation for harmful meme Detection. MIND is a groundbreaking multi-agent system designed for zero-shot harmful meme detection, meaning it doesn’t require any pre-annotated data to learn what’s harmful. This makes it particularly adaptable to the fast-evolving landscape of online content.
How MIND Works: A Collaborative Approach
MIND operates through three core strategies, mimicking how humans might collaboratively analyze content:
1. Similar Sample Retrieval (SSR): When faced with a new meme, MIND first searches through a collection of unannotated memes to find others that are visually and textually similar. This step provides crucial context, as memes often share underlying patterns even when they evolve into new formats. By combining visual and text features, MIND identifies the most relevant reference memes.
2. Relevant Insight Derivation (RID): Once similar memes are retrieved, MIND employs a unique bi-directional insight derivation mechanism. Two Large Multimodal Model (LMM) agents work together to process these similar memes. They analyze the memes in both a forward and backward sequence, ensuring that all retrieved examples contribute comprehensively to understanding potential harm. This dual-directional approach helps to capture a complete picture and prevents biases that might arise from a single processing order.
3. Insight-Augmented Inference (IAI): Finally, MIND uses a multi-agent debate mechanism to make a robust decision. Two ‘debater’ agents, each leveraging insights from the forward and backward passes, generate their judgments on the target meme’s harmfulness. If they agree, that’s the final decision. If they disagree, a ‘judge’ agent steps in to arbitrate, carefully analyzing both debaters’ reasoning to reach a well-reasoned conclusion. This debate process enhances reliability and reduces potential biases.
Impressive Results and Generalizability
Extensive experiments were conducted on three different meme datasets: HarM, FHM, and MAMI. The results show that MIND not only significantly outperforms existing zero-shot approaches but also demonstrates strong generalization across various Large Multimodal Model architectures and sizes, including powerful proprietary models like Gemini-1.5-Flash and GPT-4o. For instance, MIND, built on a smaller open-source model (LLaVA-1.5-13B), managed to surpass GPT-4o on one dataset and achieve comparable performance with Gemini-1.5-Flash on another.
Ablation studies further confirmed the importance of each component within the MIND framework. Removing any of the three core strategies (Similar Sample Retrieval, Relevant Insight Derivation, or Insight-Augmented Inference) led to a noticeable drop in performance, highlighting their complementary roles in achieving accurate harmful meme detection.
The research also explored the optimal number of similar memes to retrieve, finding that a smaller number (around K=3) generally yielded the best results, balancing performance and efficiency. This indicates that quality over quantity is key when providing contextual information.
Also Read:
- Combating Misinformation: A Multi-Agent System for Verifying Multimedia Content
- Enhancing Video Understanding Through Multimodal Fusion and Cross-Attention
Looking Ahead
MIND represents a significant step forward in combating harmful content online without the need for constant data annotation. While the framework is powerful, the researchers acknowledge areas for future improvement, such as refining the quality of retrieved similar memes, implementing more nuanced weighting for insights, and quantitatively evaluating the reliability of derived insights. Despite its computational overhead compared to simpler baselines, MIND offers a scalable and adaptable solution for maintaining safer online spaces.
The code for MIND is available on GitHub, demonstrating the researchers’ commitment to open science and further development. You can find the research paper here: MIND: A Multi-agent Framework for Zero-shot Harmful Meme Detection.


