SOTA?

Tracking the State-of-the-Art in Scholarly Publications



Background

In Artificial Intelligence (AI), a common research objective is the development of new models that can report state-of-the-art (SOTA) performance.  The reporting usually comprises four integral elements: Task, Dataset, Metric, and Score (TDMS). An illustrated example is shown in the Figure below.

These TDMS tuples coming from various AI research papers go on to power leaderboards in the community. Leaderboards, akin to scoreboards, traditionally curated by the community, are platforms displaying various AI model scores for specific tasks, datasets, and metrics. Examples of such platforms include the benchmarks feature on the Open Research Knowledge Graph and Papers with Code (PwC). Utilizing text mining techniques allows for a transition from the conventional community-based leaderboard curation to an automated text mining approach. Consequently, we introduce the "SOTA? Tracking the State-of-the-Art in Scholarly Publications" shared task as Task 4 in the SimpleText track of CLEF 2024. The goal of the SOTA? shared task is to develop systems which given the full text of an AI paper, are capable of recognizing whether an incoming AI paper indeed reports model scores on benchmark datasets, and if so, to extract all pertinent (Task, Dataset, Metric, Score) tuples presented within the paper. The Figure below shows the scholarly paper source and the downstream Leaderboard application powered by extracted TDMS tuples.

Figure: End-to-end workflow of extracting (Task, Dataset, Metric, Score) tuples from an AI scholarly publication to power the Leaderboards dashboard on the Open Research Knowledge Graph.

Task Overview

The SOTA? shared task is defined on a dataset of Artificial Intelligence scholarly articles. There are two kinds of articles: one reporting (Task, Dataset, Metric, Score) tuples and another kind that do not report the TDMS tuples. For the articles reporting TDMS tuples, all the reported TDMS annotations are provided in a separate file accompanying the scraped full-text of the articles. The extraction task is defined as follows. 

Develop a machine learning model that can distinguish whether a scholarly article provided as input to the model reports a TDMS or not. And for articles reporting TDMSs, extract all the relevant ones.

Given the recent upsurge in the developments in generative AI in the form of Large Language Models (LLMs), creative LLM-based solutions to the task are particularly invited. The task does not place any restrictions on the application of open-sourced versus closed-sourced LLMs. Nonetheless, development of open-sourced solutions are encouraged.

For more background information on this task, we recommend the following publications:

Contact

sota.task [at] gmail.com

Funding Statement

SimpleText's SOTA Task is jointly supported by the NFDI4DataScience initiative (DFG, German Research Foundation, Grant ID: 460234259) and the SCINEXT project (BMBF, German Federal Ministry of Education and Research, Grant ID: 01lS22070).