SciVQA: Scientific Visual Question Answering

Hosted as part of the SDP 2025 Workshop at ACL 2025 at Vienna, Austria.
Scholarly articles convey valuable information not only through unstructured text but also via (semi-)structured figures such as charts and diagrams. Automatically interpreting the semantics of knowledge encoded in these figures can be beneficial for downstream tasks such as question answering (QA).
In the SciVQA challenge, participants will develop multimodal QA systems using a dataset of scientific figures from ACL Anthology and arXiv papers. Each figure is annotated with seven QA pairs and includes metadata such as caption, figure ID, figure type (e.g., compound, line graph, bar chart, scatter plot), QA pair type. This shared task specifically focuses on closed-ended visual (i.e., addressing visual attributes such as colour, shape, size, height, etc.) and non-visual (not addressing figure visual attributes) questions.
Competition Platform
The shared task is available on Codabench: https://www.codabench.org/competitions/5904/#/pages-tab.
All updates and details on the competition will be published there.
Dataset
The dataset is available for download on Hugging Face: https://huggingface.co/datasets/katebor/SciVQA.
Evaluation
Systems will be evaluated using BERTscore, ROUGE-L, and ROUGE-1 metrics. Automated evaluations of submitted systems will be done through the Codabench platform.
Important Dates
- Release of training data: April 1, 2025
- Release of testing data: April 15, 2025
- Deadline for system submissions: May 16, 2025
- Paper submission deadline: May 23, 2025
- Notification of acceptance: June 13, 2025
- Camera-ready paper due: June 20, 2025
- Workshop: July 31, 2025
Organizers
Funding
This work has received funding through the DFG project NFDI4DS (no. 460234259).
Learn more here.