Accepted Papers

Contrastive Loss is All You Need to Recover Analogies as Parallel Lines

Narutatsu Ri, Fei-Tzin Lee, Nakul Verma


Fine-grained Text Style Transfer with Diffusion-Based Language Models

Yiwei Lyu, Tiange Luo, Jiacheng Shi, Todd C Hollon, Honglak Lee


Friendly Neighbors: Contextualized Sequence-to-Sequence Link Prediction

Adrian Kochsiek, Apoorv Umang Saxena, Inderjeet Jayakumar Nair, Rainer Gemulla


Grammatical information in BERT sentence embeddings as two-dimensional arrays

Vivi Nastase, Paola Merlo


Effectiveness of Data Augmentation for Parameter Efficient Tuning with Limited Data

Stephen Obadinma, Hongyu Guo, Xiaodan Zhu


Improving Zero-shot Relation Classification via Automatically-acquired Entailment Templates

Mahdi Rahimi, Mihai Surdeanu


One does not fit all! On the Complementarity of Vision Encoders for Vision and Language Tasks

Gregor Geigle, Chen Cecilia Liu, Jonas Pfeiffer, Iryna Gurevych


MUX-PLMs: Pre-training Language Models with Data Multiplexing

Vishvak Murahari, Ameet Deshpande, Carlos E Jimenez, Izhak Shafran, Mingqiu Wang, Yuan Cao, Karthik R Narasimhan


Adversarial Clean Label Backdoor Attacks and Defenses on Text Classification Systems

Ashim Gupta, Amrith Krishna


Retrieval-Augmented Domain Adaptation of Language Models

Benfeng Xu, Chunxu Zhao, Wenbin Jiang, PengFei Zhu, Songtai Dai, Chao Pang, Zhuo Sun, Shuohuan Wang, Yu Sun


Hierarchical Multi-Instance Multi-Label Learning for Detecting Propaganda Techniques

Anni Chen, Bhuwan Dhingra


Probing Negation in Language Models

Shashwat Singh, Shashwat Goel, Saujas Vaduguru, Ponnurangam Kumaraguru


Towards Flow Graph Prediction of Open-Domain Procedural Texts

Keisuke Shirai, Hirotaka Kameko, Shinsuke Mori


Extracting Multi-valued Relations from Language Models

Sneha Singhania, Simon Razniewski, Gerhard Weikum


Syntax-Aware Graph-to-Graph Transformer for Semantic Role Labelling

Alireza Mohammadshahi, James Henderson


SPC: Soft Prompt Construction for Cross Domain Generalization

Wenbo Zhao, Arpit Gupta, Tagyoung Chung, Jing Huang


LSTMs Acquire More Than One Aspect of Gender with Few-shot Learning

Priyanka Sukumaran, Conor Houghton, Nina Kazanina


Mixed Orthographic/Phonemic Language Modeling: Beyond Orthographically Restricted Transformers (BORT)

Robert C Gale, Alexandra C. Salem, Gerasimos Fergadiotis, Steven Bedrick


Enhancing text comprehension for Question Answering with Contrastive Learning

Seungyeon Lee, Minho Lee


A Multilingual Evaluation of NER Robustness to Adversarial Inputs

Akshay Srinivasan, Sowmya Vajjala


Do not Mask Randomly: Effective Domain-adaptive Pre-training by Masking In-domain Keywords

Shahriar Golchin, Mihai Surdeanu, Nazgol Tavabi, Ata Kiapour


Relational Sentence Embedding for Flexible Semantic Matching

Bin Wang, Haizhou Li


Tucker Decomposition with Frequency Attention for Temporal Knowledge Graph Completion

Likang Xiao, Richong Zhang, Zijie Chen, Junfan Chen


Visual Coherence Loss for Coherent and Visually Grounded Story Generation

Xudong Hong, Vera Demberg, Asad Sayeed, Qiankun Zheng, Bernt Schiele


CLIP-based image captioning via unsupervised cycle-consistency in the latent space

Romain Bielawski, Rufin VanRullen


Token-level Fitting Issues of Seq2seq Models

Guangsheng Bao, Zhiyang Teng, Yue Zhang


Revealing the Blind Spot of Sentence Encoder Evaluation by HEROS

Cheng-Han Chiang, Yung-Sung Chuang, James Glass, Hung-yi Lee


One-Shot Exemplification Modeling via Latent Sense Representations

John Harvill, Hee Suk Yoon, Eunseop Yoon, Mark Hasegawa-Johnson, Chang Yoo


Enhancing Sentiment Knowledge via Self-Supervised Meta-Learning

Hyunjong Kim, Sungzoon Cho


Sen2Pro: A Probabilistic Perspective to Sentence Embedding from Pre-trained Language Model

Lingfeng Shen, Lemao Liu, Haiyun Jiang, Shuming Shi


Large Language Models with Controllable Working Memory

Daliang Li, Ankit Singh Rawat, Manzil Zaheer, Xin Wang, Michal Lukasik, Andreas Veit, Felix Yu, Sanjiv Kumar


Optimizing Test-Time Query Representations for Dense Retrieval

Mujeen Sung, Jungsoo Park, Jaewoo Kang, Danqi Chen, Jinhyuk Lee


Ambiguity Meets Uncertainty: Investigating Uncertainty Estimation for Word Sense Disambiguation

Zhu Liu, Ying Liu


DivHSK: Diverse Headline Generation using Self-Attention based Keyword Selection

Venkatesh E, Kaushal Maurya, Deepak Kumar, Maunendra Sankar Desarkar


Unsupervised Semantic Variation Prediction using the Distribution of Sibling Embeddings

Taichi Aida, Danushka Bollegala


Why Does Zero-Shot Cross-Lingual Generation Fail? An Explanation and a Solution

Tianjian Li, Kenton Murray


Fighting Bias With Bias: Promoting Model Robustness by Amplifying Dataset Biases

Yuval Reif, Roy Schwartz


AxomiyaBERTa: A Phonologically-aware Transformer Model for Assamese

Abhijnan Nath, Sheikh Mannan, Nikhil Krishnaswamy


Towards Reference-free Text Simplification Evaluation with a BERT Siamese Network Architecture

Xinran Zhao, Esin Durmus, Dit-Yan Yeung


Are Layout-Infused Language Models Robust to Layout Distribution Shifts? A Case Study with Scientific Documents

Catherine Chen, Zejiang Shen, Dan Klein, Gabriel Stanovsky, Doug Downey, Kyle Lo


Constructing Character-Context-Coupled Space Aligned with Associative Knowledge Relations for Interpretable Language Modeling

Fanyu Wang, Zhenping Xie


Enhancing Out-of-Vocabulary Estimation with Subword Attention

Raj Patel, Carlotta Domeniconi


Parameter-Efficient Finetuning for Robust Continual Multilingual Learning

Kartikeya Badola, Shachi Dave, Partha Talukdar


Target-Oriented Relation Alignment for Cross-Lingual Stance Detection

Ruike Zhang, Nan Xu, Hanxuan Yang, Yuan Tian, Wenji Mao


Know Where You're Going: Meta-Learning for Parameter-Efficient Fine-Tuning

Mozhdeh Gheini, Xuezhe Ma, Jonathan May


Shielded Representations: Protecting Sensitive Attributes Through Iterative Gradient-Based Projection

Shadi Iskander, Kira Radinsky, Yonatan Belinkov


DEnsity: Open-domain Dialogue Evaluation Metric using Density Estimation

ChaeHun Park, Seungil Lee, Daniel Rim, Jaegul Choo


Impact of Adversarial Training on Robustness and Generalizability of Language Models

Enes Altinisik, Hassan Sajjad, Husrev Sencar, Safa Messaoud, Sanjay Chawla


Exploring the Relationship between Alignment and Cross-lingual Transfer in Multilingual Transformers

Felix Gaschi, Patricio Cerda, Parisa Rastin, Yannick Toussaint


Boosting Zero-shot Cross-lingual Retrieval by Training on Artificially Code-Switched Data

Robert Litschko, Ekaterina Artemova, Barbara Plank


Another Dead End for Morphological Tags? Perturbed Inputs and Parsing

Alberto Munoz-Ortiz, David Vilares


Tokenization Impacts Multilingual Language Modeling: Assessing Vocabulary Allocation and Overlap Across Languages

Tomasz Limisiewicz, Jiri Balhar, David Mareček


Modular and On-demand Bias Mitigation with Attribute-Removal Subnetworks

Lukas Hauzenberger, Shahed Masoudian, Deepak Kumar, Markus Schedl, Navid Rekabsaz


Nonparametric Decoding for Generative Retrieval

Hyunji Lee, JaeYoung Kim, Hoyeon Chang, hanseok Oh, Sohee Yang, Vladimir Karpukhin, Yi Lu, Minjoon Seo


On Dataset Transferability in Active Learning for Transformers

Fran Jelenic, Josip Jukic, Nina Drobac, Jan Snajder


Assessing Word Importance Using Models Trained for Semantic Tasks

David Javorsky, Ondrej Bojar, Francois Yvon


Yes, this Way! Learning to Ground Referring Expressions into Actions with Inter-episodic Feedback from Supportive Teachers

Philipp Sadler, Sherzod Hakimov, David Schlangen


Measuring Intersectional Biases in Historical Documents

Nadav Borenstein, Karolina Stanczak, Thea Rolskov, Natacha Klein Käfer, Natalia da Silva Perez, Isabelle Augenstein


Exploring anisotropy and outliers in multilingual language models for cross-lingual semantic sentence similarity

Katharina Haemmerl, Alina Fastowski, Jindrich Libovicky, Alexander Fraser


TADA -- Efficient Task-Agnostic Domain Adaptation for Transformers

Chia-Chien Hung, Lukas Lange, Jannik Strötgen


RQUGE: Reference-Free Metric for Evaluating Question Generation by Answering the Question

Alireza Mohammadshahi, Thomas Scialom, Majid Yazdani, Pouya Yanki, Angela Fan, James Henderson, Marzieh Saeidi


Layerwise universal adversarial attack on NLP models

Olga Tsymboi, Danil Malaev, Andrei Petrovskii, Ivan Oseledets


Exploiting Abstract Meaning Representation for Open-Domain Question Answering

Cunxiang Wang, Qipeng Guo, Zhikun Xu, Xiangkun Hu, Xuefeng Bai, Zheng Zhang, Yue Zhang


Masked Latent Semantic Modeling: an Efficient Pre-training Alternative to Masked Language Modeling

Gabor Berend


RFiD: Towards Rational Fusion-in-Decoder for Open-Domain Question Answering

Cunxiang Wang, Haofei Yu, Yue Zhang


Predicting Numerals in Text Using Nearest Neighbor Language Models

Taku Sakamoto, Akiko Aizawa


Easy to Decide, Hard to Agree: Reducing Disagreements Between Saliency Methods

Josip Jukic, Martin Tutek, Jan Snajder


History Repeats: Overcoming Catastrophic Forgetting For Event-Centric Temporal Knowledge Graph Completion

Mehrnoosh Mirtaheri, Mohammad Rostami, Aram Galstyan


Aligning Instruction Tasks Unlocks Large Language Models as Zero-Shot Relation Extractors

Zhang Kai, Bernal Jiménez Gutiérrez, Yu Su


Nonparametric Masked Language Modeling

Sewon Min, Weijia Shi, Mike Lewis, Xilun Chen, Wen-tau Yih, Hannaneh Hajishirzi, Luke Zettlemoyer


Reliable Gradient-free and Likelihood-free Prompt Tuning

Maohao Shen, Soumya Ghosh, Prasanna Sattigeri, Subhro Das, Yuheng Bu, Gregory Wornell


Detecting Edit Failures In Large Language Models: An Improved Specificity Benchmark

Jason Hoelscher-Obermaier, Julia H. Persson, Esben Kran, Ioannis Konstas, Fazl Barez


The Larger They Are, the Harder They Fail: Language Models do not Recognize Identifier Swaps in Python

Antonio Barone, Fazl Barez, Ioannis Konstas, Shay B. Cohen


Taxonomy of Problems in Lexical Semantics

Bradley Hauer, Grzegorz Kondrak


From chocolate bunny to chocolate crocodile: Do Language Models Understand Noun Compounds?

Albert Coil, Vered Shwartz


Sequential Path Signature Networks for Personalised Longitudinal Language Modeling

Talia Tseriotou, Adam Tsakalidis, Peter Foster, Terence Lyons, Maria Liakata


SamToNe: Improving Contrastive Loss for Dual Encoder Retrieval Models with Same Tower Negatives

Fedor Moiseev, Gustavo Hernandez Abrego, Peter Dornbach, Imed Zitouni, Enrique Alfonseca, Zhe Dong


Data-Efficient Finetuning Using Cross-Task Nearest Neighbors

Hamish Ivison, Noah A. Smith, Hannaneh Hajishirzi, Pradeep Dasigi


Honey, I Shrunk the Language: Language Model Behavior at Reduced Scale.

Vijeta Deshpande, Dan Pechi, Shree Thatte, Vladislav Lialin, Anna Rumshisky


Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and Evaluation

Marius Mosbach, Tiago Pimentel, Shauli Ravfogel, Dietrich Klakow, Yanai Elazar


SConE: Simplified Cone Embeddings with Symbolic Operators for Complex Logical Queries

Chau Nguyen, Tim French, Wei Liu, Michael Stewart


Recurrent Attention Networks for Long-text Modeling

Xianming Li, Zongxi Li, Xiaotian Luo, Haoran Xie, Xing Lee, Yingbin Zhao, Fu Lee Wang, Qing Li


RHO ($\rho$): Reducing Hallucination in Open-domain Dialogues with Knowledge Grounding

Ziwei Ji, Zihan Liu, Nayeon Lee, Tiezheng Yu, Bryan Wilie, Min Zeng, Pascale Fung


Causal interventions expose implicit situation models for commonsense language understanding

Takateru Yamakoshi, James McClelland, Adele Goldberg, Robert Hawkins


Multimedia Generative Script Learning for Task Planning

Qingyun Wang, Manling Li, Hou Pong Chan, Lifu Huang, Julia Hockenmaier, Girish Chowdhary, Heng Ji


Mini-Model Adaptation: Efficiently Extending Pretrained Models to New Languages via Aligned Shallow Training

Kelly Marchisio, Patrick Lewis, Yihong Chen, Mikel Artetxe


ECOLA: Enhancing Temporal Knowledge Embeddings with Contextualized Language Representations

Zhen Han, Ruotong Liao, Jindong Gu, Yao Zhang, Zifeng Ding, Yujia Gu, Heinz Koeppl, Hinrich Schütze, Volker Tresp


Enhancing Hierarchical Text Classification through Knowledge Graph Integration

Ye Liu, Zhenya Huang, Kai Zhang, Kehang Wang, Yanghai Zhang, Qi Liu, Enhong Chen


On the Expressivity Role of LayerNorm in Transformers' Attention

Shaked Brody, Uri Alon, Eran Yahav


On Isotropy, Contextualization and Learning Dynamics of Contrastive-based Sentence Representation Learning

Chenghao Xiao, Yang Long, Noura Al Moubayed


Context-Aware Document Simplification

Liam Cripwell, Joel Legrand, Claire Gardent


EmbedTextNet: Dimension Reduction with Weighted Reconstruction and Correlation Losses for Efficient Text Embedding

Dae Yon Hwang, Bilal Taha, Yaroslav Nechaev


Revisiting the Architectures like Pointer Networks to Efficiently Improve the Next Word Distribution, Summarization Factuality, and Beyond

Haw-Shiuan Chang, Zonghai Yao, Alolika Gon, hong yu, Andrew McCallum


Byte-Pair Encoding is Approximately Optimal

Vilem Zouhar, Tim Vieira, Clara Meister, Juan Gastaldi, Mrinmaya Sachan, Ryan Cotterell


Nano: Nested Human-in-the-Loop Reward Learning for Few-shot Language Model Control

Xiang Fan, Yiwei Lyu, Paul Pu Liang, Ruslan Salakhutdinov, Louis-Philippe Morency