SUMEval-2: The 2nd Workshop on Scaling Up Multilingual & Multi-Cultural Evaluation


COLING 2025

Call for Papers

Submission Channel

Timeline

All deadlines are 11:59 PM (Anywhere on Earth)

Invitation for Paper Submissions

Massively Multilingual Language Models (MMLMs) like mBERT, XLMR and XY-LENT support around 100 languages of the world. Additionally, generative models like GPT-4 and BLOOM are getting attention from the NLP community and the public. However, most existing multilingual NLP benchmarks reflect a handful of cultures and languages. The languages present in evaluation benchmarks are usually high-resource and largely belong to the Indo-European language family. By extension, the cultures represented in evaluation benchmarks are also largely reflective of Western society. This makes current evaluation unreliable and does not provide a full picture of the performance of MMLMs across the linguistic and cultural landscape. Although efforts are being made to create benchmarks that cover a larger variety of tasks, cultures, languages, and language families, it is unlikely that we will be able to build benchmarks covering all languages and cultures. Due to this, there is recent interest in alternate strategies for evaluating MMLMs, including performance prediction and Machine Translation of test data. This workshop is an extension of the SumEval 2022 workshop, with a wide scope focusing on multicultural evaluation in addition to multilingual evaluation. Topics of interest include but are not limited to:

Submission types:

We would appreciate seeing various types of works on this (but not only) topic, like:

Speakers

to be announced

Organizing Committee

Program Committee

Contact