Track 2: Event-Based Image Retrieval
EVENTA Grand Challenge - ACM Multimedia 2025

To participate in the EVENTA 2025 Grand Challenge, please first register by submitting the form.

Overview

Given a realistic caption, participants are required to retrieve corresponding images from a provided database.This retrieval task is a fundamental task in computer vision and natural language processing that requires learning a joint representation space where visual and textual modalities can be meaningfully compared. Image retrieval with textual queries is widely used in search engines, medical imaging, e-commerce, and digital asset management. However, challenges remain, such as handling abstract or ambiguous queries, improving retrieval efficiency for large-scale datasets, and ensuring robustness to linguistic variations and biases in training data. This image retrieval track aims to tackle issues of realistic information from events in real life.

Submission Instructions

Participants must submit a CSV file named using the following format: submission.csv. This file must be compressed into a ZIP archive named submission.zip before uploading it to the CodaBench platform.

The CSV file should contain predictions for all queries in the provided query set. It must include 11 columns, separated by commas (,), with the following structure:

  • Column 1: query_id — the unique ID of the query text
  • Columns 2–11: Top-10 retrieved image_id values, ordered from most to least relevant (top-1 to top-10). If an image cannot be retrieved, use # as a placeholder.

CSV Row Format Template:

query_id,article_id_1,article_id_2,...,article_id_10
<query_id>,<article_id_1>,<article_id_2>,...,<article_id_10>
<query_id>,<article_id_1>,#,...,#

We provide a submission example:

query_id,article_id_1,article_id_2,...,article_id_10
12312,56712,56723,56734,56745,56756,56767,56778,56789,56790,56701
12334,56712,#,#,#,#,#,#,#,#,#
12345,56712,56723,56734,56745,56756,#,#,#,#,#

Submission Requirements:

  • Each query_id in your submission must be unique and must come from the official provided list.
  • All image_id values must correspond to valid entries in our image database.
  • There is no requirement to sort the rows by query_id — this will be automatically handled during evaluation.
  • Your raw submission.csv file must not contain any blank lines.
  • Submit the submission.csv file as a single ZIP file named submission.zip — do not include any containing folder.
  • The submission.csv must include all required headers and columns, in the exact order as shown in the example file.

Submission Limits:

  • For public test submissions, each team is allowed a maximum of 10 submissions per day and a total of 150 submissions.
  • For private test submissions, each team is allowed a maximum of 10 submissions per day and a total of 150 submissions.

Participants are also required to submit a detailed technical paper via the CMT platform to validate their approach and results.


Evaluation

To fairly and comprehensively evaluate submissions in this challenge task, we define an Overall Score that integrates key evaluation criteria using a principled and balanced approach.

The Overall Score combines the following five metrics:

MetricDescription
mAPMean Average Precision – measures overall retrieval precision
MRRMean Reciprocal Rank – measures how early the first correct item is retrieved
Recall@1Whether the correct item is ranked first
Recall@5Whether the correct item is within the top 5 results
Recall@10Whether the correct item is within the top 10 results

These metrics are combined using a weighted harmonic mean, which penalizes poor-performing metrics more strongly than an arithmetic mean and encourages balanced improvements across all criteria. The formula is:

\[ \text{Overall Score} = r \cdot \left( \frac{\sum w_i}{\sum \left( \frac{w_i}{m_i + \varepsilon} \right)} \right) \]

Where:

  • mi is the value of the i-th retrieval metric
  • wi is the weight for the i-th metric
  • ε is a small constant (e.g., 10-8) to avoid division by zero
  • r = (#valid queries) / (#total ground-truth queries)

Default Weights: [0.3, 0.2, 0.2, 0.15, 0.15] for [mAP, MRR, R@1, R@5, R@10]

This Overall Score will be used as the official ranking metric for evaluating all participant submissions on the retrieval leaderboard.

Leaderboard
TBA