Building on the success of previous iterations, the 2026 edition shifts focus toward a updated high-quality dataset from 2025 academic questions
and fully transparent open-source development.
Evaluate the next generation of frontier multimodal models on challenging vision-based exam
problems.
Updated reasoning-intensive data from 2025 academic questions, focusing on diagrams, charts, and technical schematics.
Promoting transparency with fully open-source architectures and reproducible systems.
Official announcement and website launch for the 2026 edition.
Registration opens for all ImageCLEF tasks
(OpenQA) Train & Development dataset release
(OpenQA+MCQ) Test dataset release
Registration closes for all ImageCLEF tasks
Deadline for submitting participant runs
The competition tasks are designed to evaluate the performance of multimodal models on challenging vision-based exam problems.
Given an image of a question with three to five possible answer options, the solution must select the single correct answer.
ClassificationGiven an image of a question without predefined answer options, the solution must generate a free-form textual answer.
GenerativeSubmissions must be reproducible and run within these constraints for the private test set evaluation.
To promote and encourage efficient solutions, the competition has two categories:
Submit JSON predictions only.
[{ "question_id": "...", "predicted_answer": "..." }]
Max 20 submissions/day, 200 total.
Sofia University "St. Kliment Ohridski", Bulgaria
MBZUAI, UAE
MBZUAI, UAE
MBZUAI, UAE
MBZUAI, UAE
Sofia University "St. Kliment Ohridski", Bulgaria
Sofia University "St. Kliment Ohridski", Bulgaria
Sofia University "St. Kliment Ohridski", Bulgaria
MBZUAI, UAE