CLEF 2026 FinMMEval Lab - Call for Participation Register Now
CLEF 2026 - Call for Participation

FinMMEval Lab 2026

Multilingual and Multimodal Evaluation of Financial AI Systems

Training Data Released Released: 15 December 2025

Ready-to-use splits for all tasks

Get calibrated training splits for exam-style Q&A, multilingual financial reasoning, and trading decision-making. Each dataset card includes format, licensing, and citation guidance.

1
Task 1
Multilingual exam-style multiple choice.
2
Task 2
PolyFiQA (Easy + Expert) filings with multilingual news Q&A.
3
Task 3
BTC and TSLA daily contexts for Buy/Hold/Sell reasoning.
New
Browse the Hugging Face collection to download splits and review licensing.
Open Collection

Check each dataset card for citation, licensing, and format details.

About the Lab

FinMMEval Lab integrates financial reasoning, multilingual understanding, and decision-making into a unified evaluation suite designed to promote robust, transparent, and globally competent financial AI. The 2026 edition introduces three interconnected tasks spanning five languages.

Multi-modal inputs: news, filings, macro indicators, tests.

Multiple languages with low-resource representations: English, Chinese, Arabic, Hindi, Greek, Japanese, Spanish.

Tasks spanning Q&A, and decision making.

Metrics centered on Accuracy, ROUGE-1, BLEURT and performance quantitative metrics (e.g. CR, SR, MD).

Financial AI Framework

"How can I tailor my setup to make an LLM exceptionally good at finance?"

Tasks

Choose one or more tasks. Each submission must provide calibrated confidence scores and an evidence trace.

Task 1 - Financial Exam Q&A

Given a stand-alone multiple-choice question Q with four candidate options { A1, A2, A3, A4 }, the system must select the correct answer A. Questions cover valuation, accounting, ethics, corporate finance, and regulatory knowledge.

Motivation

Professional financial qualification exams (e.g., CFA, EFPA) require the integration of theoretical and regulatory knowledge with applied reasoning. Existing LLMs often rely on factual recall without demonstrating the analytical rigor expected from human candidates.

Data

  • EFPA (Spanish): 50 exam-style financial questions on investment and regulation.
  • GRFinQA (Greek): 225 multiple-choice finance questions from university-level exams.
  • CFA (English): 600 exam-style multiple-choice questions covering nine core domains.
  • CPA (Chinese): 300 exam-style financial questions focusing on major modules.
  • BBF (Hindi): 500-1000 exam-style financial multiple-choice questions covering over 30 domains.

Evaluation

Models are required to output the correct answer label. Performance is measured by accuracy, defined as the proportion of correctly identified options in the test set.

Important Dates

Track key milestones for participation, training data, submissions, and workshop deadlines.

17 Nov 2025

Lab registration opens

Sign up for FinMMEval tasks.

15 Dec 2025

Training data released

Available now via the Hugging Face collection.

23 Apr 2026

Lab registration closes

Last day to register teams.

May 2026

Evaluation cycle begins

Test sets released to participants.

07 May 2026

Run submission

Deadline for submitting system runs.

28 May 2026

Working notes due [CEUR-WS]

Submit your workshop paper.

28 May – 30 Jun 2026

Participant paper review

Review phase for submitted notes.

08 Jun 2026

Condensed Lab Overviews due [LNCS]

Submit condensed lab overview.

15 Jun 2026

Overview acceptance (LNCS)

Notifications for condensed overviews.

22 Jun 2026

Camera ready (LNCS)

Camera-ready condensed lab overview due.

30 Jun 2026

Participant papers acceptance

Notifications for CEUR-WS participant papers.

06 Jul 2026

Camera ready (CEUR-WS)

Camera-ready participant papers and extended overviews.

21–24 Sep 2026

CLEF 2026 Conference

Jena, Germany.

How to Participate

Engage with the challenges in a way that suits you - from a quick, one-time experiment to a detailed research project. While we invite you to share your findings in our workshop notes, you are also free to develop promising results into a full paper for an archival journal.

The workshop itself is a perfect opportunity to refine your ideas through discussion with peers.

Ready to join?

Sign Up

Sign up via the CLEF registration form (FinMMEval section)

Packaging Checklist


  • Results JSONL (per task)
  • System Card (architecture, data usage, risks)
  • Reproducibility (seed, versions, hardware)
  • License compliance acknowledgements (if applicable)

Organizers

Organizing committee and partner institutions.

Zhuohan Xie
Zhuohan Xie
MBZUAI (UAE)
Rania Elbadry
Rania Elbadry
MBZUAI (UAE)
Fan Zhang
Fan Zhang
University of Tokyo (Japan)
Georgi Georgiev
Georgi Georgiev
Sofia University "St. Kliment Ohridski" (Bulgaria)
Xueqing Peng
Xueqing Peng
The Fin AI (USA)
Lingfei Qian
Lingfei Qian
The Fin AI (USA)
Jimin Huang
Jimin Huang
The Fin AI (USA)
Dimitar Dimitrov
Dimitar Dimitrov
Sofia University "St. Kliment Ohridski" (Bulgaria)
Vanshikaa Jani
Vanshikaa Jani
University of Arizona (USA)
Yuyang Dai
Yuyang Dai
INSAIT (Bulgaria)
Jiahui Geng
Jiahui Geng
MBZUAI (UAE)
Yuxia Wang
Yuxia Wang
INSAIT (Bulgaria)
Ivan Koychev
Ivan Koychev
Sofia University "St. Kliment Ohridski" (Bulgaria)
Veselin Stoyanov
Veselin Stoyanov
MBZUAI (UAE)
Preslav Nakov
Preslav Nakov
MBZUAI (UAE)

Frequently Asked Questions

Who can participate?

Researchers and practitioners from academia and industry. Student teams are particularly welcome.

How is data licensed?

Research-only license; redistribution of raw sources may be restricted.

Can we submit to multiple tasks?

Yes. Submit independent result bundles per task.

Are ensembles allowed?

Yes, but disclose all components in the system card.