Accepted Papers
Distributionally Robust Recurrent Decoders with Random Network Distillation
Antonio Valerio Miceli Barone, Alexandra Birch, Rico Sennrich
Q-Learning Scheduler for Multi Task Learning Through the use of Histogram of Task Uncertainty
Kourosh Meshgi, Maryam Sadat Mirzaei, Satoshi Sekine
When does CLIP generalize better than unimodal models? When judging human-centric concepts
Romain Bielawski, Benjamin Devillers, Tim Van De Cruys, Rufin Vanrullen
From Hyperbolic Geometry Back to Word Embeddings
Zhenisbek Assylbekov, Sultan Nurmukhamedov, Arsen Sheverdin, Thomas Mach
A Comparative Study of Pre-trained Encoders for Low-Resource Named Entity Recognition
Yuxuan Chen, Jonas Mikkelsen, Arne Binder, Christoph Alt, Leonhard Hennig
Clozer: Adaptable Data Augmentation for Cloze-style Reading Comprehension
Holy Lovenia, Bryan Wilie, Willy Chung, Zeng Min, Samuel Cahyawijaya, Dan Su, Pascale Fung
Analyzing Gender Representation in Multilingual Models
Hila Gonen, Shauli Ravfogel, Yoav Goldberg
Na Liu, Mark Dras, Wei Emma Zhang
A Vocabulary-Free Multilingual Neural Tokenizer for End-to-End Task Learning
Md Mofijul Islam, Gustavo Aguilar, Pragaash Ponnusamy, Clint Solomon Mathialagan, Chengyuan Ma, Chenlei Guo
Identifying the Limits of Cross-Domain Knowledge Transfer for Pretrained Models
Zhengxuan Wu, Nelson F. Liu, Christopher Potts
Temporal Knowledge Graph Reasoning with Low-rank and Model-agnostic Representations
Ioannis Dikeoulias, Saadullah Amin, Günter Neumann
ANNA: Enhanced Language Representation for Question Answering
Changwook Jun, Hansol Jang, Myoseop Sim, Hyun Kim, Jooyoung Choi, Kyungkoo Min, Kyunghoon Bae
Video Language Co-Attention with Multimodal Fast-Learning Feature Fusion for VideoQA
Adnen Abdessaied, Ekta Sood, Andreas Bulling
Detecting Word-Level Adversarial Text Attacks via SHapley Additive exPlanations
Edoardo Mosca, Lukas Huber, Marc Alexander Kühn, Georg Groh
Binary Encoded Word Mover's Distance
Christian Johnson
Shaked Haim Meirom, Omer Bobrowski
A Study on Entity Linking Across Domains: Which Data is Best for Fine-Tuning?
Hassan Soliman, Heike Adel, Mohamed H. Gad-Elrab, Dragan Milchevski, Jannik Strötgen
TRAttack: Text Rewriting Attack Against Text Retrieval
Junshuai Song, Jiangshan Zhang, Jifeng Zhu, Mengyun Tang, Yong Yang
On the Geometry of Concreteness
Christian Wartena
Towards Improving Selective Prediction Ability of NLP Systems
Neeraj Varshney, Swaroop Mishra, Chitta Baral
On Target Representation in Continuous-output Neural Machine Translation
Evgeniia Tokarchuk, Vlad Niculae
Zero-shot Cross-lingual Transfer is Under-specified Optimization
Shijie Wu, Benjamin Van Durme, Mark Dredze
Same Author or Just Same Topic? Towards Content-Independent Style Representations
Anna Wegmann, Marijn Schraagen, Dong Nguyen
WeaNF: Weak Supervision with Normalizing Flows
Andreas Stephan, Benjamin Roth
Isomorphic Cross-lingual Embeddings for Low-Resource Languages [Non-archival]
Sonal Sannigrahi, Jesse Read
PARADISE: Exploiting Parallel Data for Multilingual Sequence-to-Sequence Pretraining [Non-archival]
Machel Reid, Mikel Artetxe
PALBERT: Teaching ALBERT to Ponder [Non-archival]
Daniil Gavrilov, Nikita Balagansky
Lacking the Embedding of a Word? Look it up into a Traditional Dictionary [Non-archival]
Elena Sofia Ruzzetti, Leonardo Ranaldi, Michele Mastromattei, Francesca Fallucchi, Noemi Scarpato, Fabio Massimo Zanzotto