Skip to content

esborisova/SciVQA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

drawing

Scientific Visual Question Answering (SciVQA) shared task was orginised as part of the Scholarly Document Processing workshop (SDP) at ALC 2025. In this challenge, participants developed multimodal QA systems using images of scientific figures, their captions, associated natural language QA pairs, and optionally additional metadata. The competition was hosted on the Codabench platform.

This repository stores the code used for constructing the SciVQA dataset and developing the competition baseline.

Data

The SciVQA dataset comprises 3000 images of real-world figures extracted from English scientific publications in Computational Linguistics available in arXiv and ACL Anthology. The images are collected from the two pre-existing datasets:

  • ACL-Fig

  • SciGraphQA

Each figure is availabe as PNG and associated with 7 QA pairs according to the custom schema (see below). All figures are automatically annotated using the Gemini 1.5-flash model and then manually validated by graduate students with Computational Linguistics background. SciVQA contains 21000 QA pairs in total. The language of all QA pairs is English.

The dataset is publicly available on 🤗Hugging Face.

QA pair types schema

drawing

  • Closed-ended - it is possible to answer a question based only on a given data source, i.e., an image or an image and a caption. No additional resources such as the main text of a publication, other documents/figures/tables, etc. are required.
  • Unanswerable - it is not possible to infer an answer based solely on a given data source.
  • Infinite answer set - there are no predefined answer options., e.g., "What is the sum of Y and Z?".
  • Finite answer set - associated with a limited range of answer options. Such QA pairs fall into two subcategories:
    • Binary - require a yes/no or true/false answer, e.g., "Is the percentage of positive tweets equal to 15%?".
    • Non-binary - require to choose from a set of four predefined answer options where one or more are correct, e.g., "What is the maximum value of the green bar at the threshold equal to 10?" Answer options: "A: 5, B: 10, C: 300, D: None of the above".
  • Visual - address or incorporate information on one or more of the six visual attributes of a figure, i.e., shape, size, position, height, direction or colour. E.g., "In the bottom left figure, what is the value of the blue line at an AL of 6?". Here the visual aspects are: position (bottom left), colour (blue), and shape (line).
  • Non-visual - do not involve any of the six visual aspects of a figure defined in our schema, e.g., "What is the minimum value of X?".

Repository structure

    ├── data               # qa prompts, annotation guidelines, csv files with human answers and predictions from the baseline
    ├── src               
    │   ├── data_pred      # data preparation scripts
    │   ├── baseline       # code for baseline
    │   ├── eval           # scoring script      
    └──         

Cite

@inproceedings{borisova-etal-2025-scivqa,
    title = "{S}ci{VQA} 2025: Overview of the First Scientific Visual Question Answering Shared Task",
    author = "Borisova, Ekaterina  and
      Rauscher, Nikolas  and
      Rehm, Georg",
    editor = "Ghosal, Tirthankar  and
      Mayr, Philipp  and
      Singh, Amanpreet  and
      Naik, Aakanksha  and
      Rehm, Georg  and
      Freitag, Dayne  and
      Li, Dan  and
      Schimmler, Sonja  and
      De Waard, Anita",
    booktitle = "Proceedings of the Fifth Workshop on Scholarly Document Processing (SDP 2025)",
    month = jul,
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2025.sdp-1.18/",
    pages = "182--210",
    ISBN = "979-8-89176-265-7",
    abstract = "This paper provides an overview of the First Scientific Visual Question Answering (SciVQA) shared task conducted as part of the Fifth Scholarly Document Processing workshop (SDP 2025). SciVQA aims to explore the capabilities of current multimodal large language models (MLLMs) in reasoning over figures from scholarly publications for question answering (QA). The main focus of the challenge is on closed-ended visual and non-visual QA pairs. We developed the novel SciVQA benchmark comprising 3,000 images of figures and a total of 21,000 QA pairs. The shared task received seven submissions, with the best performing system achieving an average F1 score of approx. 0.86 across ROUGE-1, ROUGE-L, and BertScore metrics. Participating teams explored various fine-tuning and prompting strategies, as well as augmenting the SciVQA dataset with out-of-domain data and incorporating relevant context from source publications. The findings indicate that while MLLMs demonstrate strong performance on SciVQA, they face challenges in visual reasoning and still fall behind human judgments."
}

About

SciVQA: Scientific Visual Question Answering shared task

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages