SCIDOCA 2025 Shared Task


SCIDOCA 2025 Shared Task Overview

This year’s shared task will focus on Citation Prediction, Discovery, and Placement within scientific documents. Participants will be challenged to develop models that can accurately predict relevant citations, discover masked citations, and identify the specific sentences in which citations should be inserted. The shared task is designed to assess models’ abilities to understand the intricate citation networks in scientific discourse while also exploring how well they handle domain-specific knowledge.

Motivation and Impact

The SCIDOCA 2025 Shared Task is designed to address the growing need for automated citation systems that assist researchers in managing the ever-expanding corpus of scientific literature. By improving citation discovery and placement, this task could lead to advancements in:


By focusing on these tasks, the shared task aims to advance the state of research in scientific document analysis and citation management.

Important Dates

Data Usage Rules Summary

Subtasks

Predict relevant citations for a given paragraph without specifying the exact sentence.

Predict the correct citation for each masked citation slot in a paragraph.

Identify the correct citation for each sentence in a paragraph that contains a citation.

Subtask 1: Citation Discovery

Objective:

Predict relevant citations for a paragraph without specifying the exact sentence where the citation belongs.


Input:

Example Input:

{

   "paragraph": "Recent advances in natural language processing have significantly improved the performance of models on various tasks such as machine translation and question answering.",

   "candidate_references": [

      "[Vaswani et al. 2017]",

      "[Devlin et al. 2019]",

      "[Brown et al. 2020]",

      "[Radford et al. 2018]"

   ]

}


Output:

Example Output:

{

   "predicted_citations": [

      "[Vaswani et al. 2017]",

      "[Devlin et al. 2019]"

   ]

}


Evaluation: 

For a given paragraph i, calculate Precision, Recall, and F1-Score using the equations:

Evaluation Across the Dataset: Weight the metrics by the number of ground-truth citations (GT_i) in each paragraph.


Subtask 2: Masked Citation Prediction

Objective: Participants will predict the correct citation for each masked citation slot within a paragraph where the citation has been removed.


Input:

Example Input:

{

   "paragraph": "Transformer models like BERT [MASK1a][MASK1b] and GPT-3 [MASK2] have revolutionized natural language processing tasks. These models [MASK3] continue to set benchmarks across various domains.",

   "candidate_references": [

      "[Vaswani et al. 2017]",

      "[Devlin et al. 2019]",

      "[Brown et al. 2020]",

      "[Radford et al. 2018]"

   ]

}


Output:

Example Output:

{

   "predicted_citations": {

      "[MASK1a][MASK1b]": ["[Devlin et al. 2019]"],

      "[MASK2]": ["[Brown et al. 2020]"],

      "[MASK3]": ["[Radford et al. 2018]"]

   }

}

Subtask 3: Citation Sentence Prediction

Objective: Given a paragraph, participants will predict the correct citation for each sentence that contains a citation.


Input:

Example Input:

{

   "paragraph": ["Transformer models have transformed the field of NLP.","One of the most influential models is BERT.", "We will investigate the results of BERT models." ,"GPT-3 has further pushed the boundaries of language modeling."],

   "candidate_references": [

      "[Vaswani et al. 2017]",

      "[Devlin et al. 2019]",

      "[Brown et al. 2020]",

      "[Radford et al. 2018]"

   ]

}


Output:

Sentence Citations: A mapping of sentences to the correct citation(s), if required.

Example Output:

{

   "sentence_citations": [

      {

         "sentence": "Transformer models have transformed the field of NLP.",

         "predicted_citation": ["[Vaswani et al. 2017]"]

      },

      {

         "sentence": "One of the most influential models is BERT.",

         "predicted_citation": ["[Devlin et al. 2019]"]

      },

      {

         "sentence": "We will investigate the results of BERT models.",

         "predicted_citation": [[empty], [Devlin et al. 2019]]

      },

      {

         "sentence": "GPT-3 has further pushed the boundaries of language modeling.",

         "predicted_citation": ["[Brown et al. 2020]"]

      }

   ]

}