Poison Sample Detection and Trigger Retrieval in Multimodal VLMs - ICIP2025 Grand Challenge
This grand challenge is supported by National Intelligence and Security Discovery Research Grants (project# NS220100007), funded by the Department of Defence Australia.
Multimodal Vision Language Models (VLMs) systems integrate both visual and textual inputs to provide context-aware responses. These models are rapidly becoming foundational tools across domains like autonomous vehicles, defense systems, medical imaging, and assistive technologies.
However, their deployment introduces a significant threat surface for backdoor attacks โ subtle, malicious manipulations of training data that embed hidden triggers. When activated, these triggers can force the model to output incorrect or adversarial responses. Such vulnerabilities raise concerns about trust, transparency, and safety in AI systems.
This ICIP 2025 Grand Challenge seeks to address this problem head-on by inviting the research community to develop methods for detecting poisoned samples and retrieving hidden triggers in multimodal VLMs.
The registration document, dataset and results template are now avaliable here
๐ฏ Challenge Overview
The organizers have trained poisoned VLMs using samples from the released dataset. Participants will analyze the image dataset, which includes both clean and poisoned samples. The task is to detect poisoned samples and classify the severity of the poisoning trigger.
Securing multimodal VLMs is more than a technical problem โ itโs a societal necessity, with implications ranging from public safety to the ethical deployment of AI in the real world!
๐ค Submission Details
Participants must submit:
- The completed โICIP25_poison_sample_detection_results.csvโ results file
- A 4-page paper (ICIP format) detailing methods and results. References are not included in the page limit.
๐งช Evaluation Criteria:
- Detection Accuracy (40%): Poisoned vs. clean
- Poisoning Severity Classification (40%):
- Level 1 (Weak) โ 10%
- Level 2 (Moderate) โ 15%
- Level 3 (Strong) โ 15%
- Paper Quality (20%): Clarity, methodology, and insight
Participants are also welcome to include appendices/supplementary material to support their submission. Paper submissions are single-blind reviewed by at least three reviewers.
All submissions are to be submitted via our ICIP grand challenge submission portal. The paper format is the same as for the main conference, and all relevant information can be found here. After the final challenge results notification, the top 3 teams will be encouraged to submit a final version of their paper, incorporating reviewer feedback. Accepted papers will then be part of the ICIP workshop proceedings.
The winner of the challenge will be announced at ICIP during the grand challenge session and the winning team members will be awarded a certificate on the day. The top 3 placed teams are also invited to do a 15 minute presentation of their approach if they are attending ICIP.
๐ Timeline
Milestone | Date |
---|---|
๐ Dataset Release | April 22, 2025 |
๐ Registration Deadline | |
๐ค Challenge + Paper Submission | |
โ Final Challenge Results Notification | June 25, 2025 |
๐ Top 3 Final Paper Submission Due | July 2, 2025 |
๐ค Grand Challenge Session (@ ICIP) | September 2025 |
๐ Rules
- Eligibility: Open to all research, academic, and industry participants
- Submissions: A short manuscript detailing approach and results (max 4 pages) + CSV results file.
- Dataset Use: Only the provided dataset is permitted
- Evaluation: Two-stage scoring (accuracy + type)
- Ethics: No plagiarism; original work only
- Team Submissions: Allowed; list all team members
- Deadlines: Late submissons will not be reviewed. Revised submissions after the deadline will also not be reveiwed.
- Disqualification: For any rule violations. All submissions must include a discussion of the methods used.
๐ฅ Organizers
- Jordan Vice โ University of Western Australia
jordan.vice@uwa.edu.au - Ajmal Mian โ University of Western Australia
ajmal.mian@uwa.edu.au - Richard Hartley โ Australian National University
- Naveed Akhtar โ University of Melbourne
Registration and Contact Information
Please download and send the completed registration form to jordan.vice@uwa.edu.au by the registration date (May 28 June 10, 2025).
All submissions, questions, clarifications or issues must be directed to:
๐ง jordan.vice@uwa.edu.au
if you have trouble accessing any of the resources for the challenge, please contact an organizer ASAP.
Dataset Information
- ๐ผ๏ธ Number of Samples: 32,648 .png images
- ๐ฆ Compressed .zip dataset size: 18.7GB
- ๐งช Data Composition: A mix of clean and poisoned samples
- ๐ Image Labeling: Unique IDs only (no descriptive metadata)
- ๐ Results File Format (.csv):
image_path | poisoned (1 or 0) | poison_type (0-3) |
---|---|---|
ICIP_CHALLENGE_DATASET/img_0001.png | 1 | 3 |
ICIP_CHALLENGE_DATASET/img_0002.png | 0 | 0 |
Evaluation results must be shared in this format only. Submissions that do not conform to this template will be disqualified.
๐ Challenge Leaderboard
Rank | Team | Score (not inc. paper) |
---|---|---|
1 | ย | ย |
2 | ย | ย |
3 | ย | ย |
4 | ย | ย |
5 | ย | ย |
โ FAQ
-
Will there be any labeled evaluation set released before the submission deadline, or should we only use the provided unlabeled dataset?
A: All of the provided, unlabeled dataset is evaluated in the challenge. There will be no evaluation set released closer to the date.
-
For poisoning severity classification, is severity defined based on any specific metric (e.g., classification impact, image quality degradation, etc.)?
A: This is a tricky one as I canโt give too much away/any competetive edge! Severity levels reflect the increasing subtlety and complexity of the poisoning. While I canโt disclose the exact criteria to maintain the challenge integrity, higher severity indicates perturbations that are designed to be harder to detect, both visually and algorithmically.
-
Are we allowed to use pre-trained models trained on external datasets for our methods?
A: Any method or pre-trained model is allowed, noting that part of the evaluation criteria is the quality of the written work. Thus, novelty and innovation in your approach is something thst should be considered. But, we do not have any constraints on this.
-
Should images with watermarks like ones with โalamyโ and โshutterstockโ be considered as clean or poisoned?. The definition of clean and poisoned should be more clarified in this challenge.
A: Watermarked images are clean as these features are intrinsic to the images. Apologies about the lack of descriptions for the data, but in order for it to be an effective challenge, we opted to provide minimal details - if we gave too much details about the dataset away, we may risk invalidating the challenge itself or giving participants advantages over other who do not ask questions. The blind labelling of data alludes to this purpose.
-
Is it possible to provide us with the percentage of poisoning? if it is 0.1% or 1% or 10%?
A: Answering this is difficult as one of criteria is poison type. We cannot give specifics about poisoning, apologies.