Skip to the content.

FNDetect: Multimodal Fake News Detection & Conflict-Aware AI Benchmark

1. Motivation and Introduction

Social media has emerged as a key arena for shaping narratives during geopolitical crises. The discussions around the Nakba and the aftermath of the October 7th conflict in Gaza illustrate a complex mix of historical trauma, live conflict reporting, and highly polarized public opinion.

Analyzing what drives a post to “go viral” in this context is essential for understanding information spread, propaganda dynamics, and public sentiment. This shared task introduces a novel challenge: predicting the reach (virality) and engagement (interactions) of posts that are emotionally charged, historically significant, and often multimodal—combining text with graphic or symbolic imagery.

By concentrating on this specialized domain, the task aims to advance the capabilities of NLP and computer vision models in processing context-rich, sensitive, and polarizing content.

2. Data Description

We present a curated, multi-platform dataset specifically designed for this task.

Note on Ethics: All data will be anonymized to protect user privacy, given the sensitive nature of the topic. IDs and handles will be hashed.

3. Task Definitions

We propose three distinct tasks to evaluate model performance on different modalities and prediction objectives.

Task 1: Binary Multilingual Fake News Detection

The first subtask focuses on detecting misinformation in Arabic and English posts using textual information only. The dataset includes posts collected from Facebook and Twitter, containing fields such as author information, post URL, textual content. Each post is labeled as real or fake based on the Text Target field. The goal is to build multilingual classification models that can accurately identify misinformation across different languages and contexts.

Task 2: Arabic Cross-Modal Inconsistency Detection

The second subtask targets Arabic posts containing both text and images. It aims to identify inconsistencies between the textual and visual content. Based on the Text Target and Img Target annotations, new labels—consistent, inconsistent, and manipulated—are generated. This enables the detection of cases where textual and visual information either align or contradict each other, supporting the development of models capable of identifying complex multimodal misinformation.

Task 3: Image Captioning for Conflict Awareness: Gaza–Israel War 2023–2025

Participants must build an image captioning model that generates accurate, descriptive, and unbiased captions for images related to the Gaza–Israel conflict. The images, collected from Facebook and Twitter between October 2023 and August 2025, cover events such as protests, destruction, aid efforts, and media scenes. The system should produce empathetic captions that factually describe each image’s content.

4. Baseline Systems

To assist participants, the organizers will provide the following baselines:

5. Significance and Impact

This shared task contributes to the NLP and multimodal AI communities by:

6. Tentative Schedule

7. Organizers

Participation Guidelines