Accepted Papers
Contrastive Loss is All You Need to Recover Analogies as Parallel Lines
Narutatsu Ri, Fei-Tzin Lee, Nakul Verma
Fine-grained Text Style Transfer with Diffusion-Based Language Models
Yiwei Lyu, Tiange Luo, Jiacheng Shi, Todd C Hollon, Honglak Lee
Friendly Neighbors: Contextualized Sequence-to-Sequence Link Prediction
Adrian Kochsiek, Apoorv Umang Saxena, Inderjeet Jayakumar Nair, Rainer Gemulla
Grammatical information in BERT sentence embeddings as two-dimensional arrays
Vivi Nastase, Paola Merlo
Effectiveness of Data Augmentation for Parameter Efficient Tuning with Limited Data
Stephen Obadinma, Hongyu Guo, Xiaodan Zhu
Improving Zero-shot Relation Classification via Automatically-acquired Entailment Templates
Mahdi Rahimi, Mihai Surdeanu
One does not fit all! On the Complementarity of Vision Encoders for Vision and Language Tasks
Gregor Geigle, Chen Cecilia Liu, Jonas Pfeiffer, Iryna Gurevych
MUX-PLMs: Pre-training Language Models with Data Multiplexing
Vishvak Murahari, Ameet Deshpande, Carlos E Jimenez, Izhak Shafran, Mingqiu Wang, Yuan Cao, Karthik R Narasimhan
Adversarial Clean Label Backdoor Attacks and Defenses on Text Classification Systems
Ashim Gupta, Amrith Krishna
Retrieval-Augmented Domain Adaptation of Language Models
Benfeng Xu, Chunxu Zhao, Wenbin Jiang, PengFei Zhu, Songtai Dai, Chao Pang, Zhuo Sun, Shuohuan Wang, Yu Sun
Hierarchical Multi-Instance Multi-Label Learning for Detecting Propaganda Techniques
Anni Chen, Bhuwan Dhingra
Probing Negation in Language Models
Shashwat Singh, Shashwat Goel, Saujas Vaduguru, Ponnurangam Kumaraguru
Towards Flow Graph Prediction of Open-Domain Procedural Texts
Keisuke Shirai, Hirotaka Kameko, Shinsuke Mori
Extracting Multi-valued Relations from Language Models
Sneha Singhania, Simon Razniewski, Gerhard Weikum
Syntax-Aware Graph-to-Graph Transformer for Semantic Role Labelling
Alireza Mohammadshahi, James Henderson
SPC: Soft Prompt Construction for Cross Domain Generalization
Wenbo Zhao, Arpit Gupta, Tagyoung Chung, Jing Huang
LSTMs Acquire More Than One Aspect of Gender with Few-shot Learning
Priyanka Sukumaran, Conor Houghton, Nina Kazanina
Mixed Orthographic/Phonemic Language Modeling: Beyond Orthographically Restricted Transformers (BORT)
Robert C Gale, Alexandra C. Salem, Gerasimos Fergadiotis, Steven Bedrick
Enhancing text comprehension for Question Answering with Contrastive Learning
Seungyeon Lee, Minho Lee
A Multilingual Evaluation of NER Robustness to Adversarial Inputs
Akshay Srinivasan, Sowmya Vajjala
Do not Mask Randomly: Effective Domain-adaptive Pre-training by Masking In-domain Keywords
Shahriar Golchin, Mihai Surdeanu, Nazgol Tavabi, Ata Kiapour
Relational Sentence Embedding for Flexible Semantic Matching
Bin Wang, Haizhou Li
Tucker Decomposition with Frequency Attention for Temporal Knowledge Graph Completion
Likang Xiao, Richong Zhang, Zijie Chen, Junfan Chen
Visual Coherence Loss for Coherent and Visually Grounded Story Generation
Xudong Hong, Vera Demberg, Asad Sayeed, Qiankun Zheng, Bernt Schiele
CLIP-based image captioning via unsupervised cycle-consistency in the latent space
Romain Bielawski, Rufin VanRullen
Token-level Fitting Issues of Seq2seq Models
Guangsheng Bao, Zhiyang Teng, Yue Zhang
Revealing the Blind Spot of Sentence Encoder Evaluation by HEROS
Cheng-Han Chiang, Yung-Sung Chuang, James Glass, Hung-yi Lee
One-Shot Exemplification Modeling via Latent Sense Representations
John Harvill, Hee Suk Yoon, Eunseop Yoon, Mark Hasegawa-Johnson, Chang Yoo
Enhancing Sentiment Knowledge via Self-Supervised Meta-Learning
Hyunjong Kim, Sungzoon Cho
Sen2Pro: A Probabilistic Perspective to Sentence Embedding from Pre-trained Language Model
Lingfeng Shen, Lemao Liu, Haiyun Jiang, Shuming Shi
Large Language Models with Controllable Working Memory
Daliang Li, Ankit Singh Rawat, Manzil Zaheer, Xin Wang, Michal Lukasik, Andreas Veit, Felix Yu, Sanjiv Kumar
Optimizing Test-Time Query Representations for Dense Retrieval
Mujeen Sung, Jungsoo Park, Jaewoo Kang, Danqi Chen, Jinhyuk Lee
Ambiguity Meets Uncertainty: Investigating Uncertainty Estimation for Word Sense Disambiguation
Zhu Liu, Ying Liu
DivHSK: Diverse Headline Generation using Self-Attention based Keyword Selection
Venkatesh E, Kaushal Maurya, Deepak Kumar, Maunendra Sankar Desarkar
Unsupervised Semantic Variation Prediction using the Distribution of Sibling Embeddings
Taichi Aida, Danushka Bollegala
Why Does Zero-Shot Cross-Lingual Generation Fail? An Explanation and a Solution
Tianjian Li, Kenton Murray
Fighting Bias With Bias: Promoting Model Robustness by Amplifying Dataset Biases
Yuval Reif, Roy Schwartz
AxomiyaBERTa: A Phonologically-aware Transformer Model for Assamese
Abhijnan Nath, Sheikh Mannan, Nikhil Krishnaswamy
Towards Reference-free Text Simplification Evaluation with a BERT Siamese Network Architecture
Xinran Zhao, Esin Durmus, Dit-Yan Yeung
Are Layout-Infused Language Models Robust to Layout Distribution Shifts? A Case Study with Scientific Documents
Catherine Chen, Zejiang Shen, Dan Klein, Gabriel Stanovsky, Doug Downey, Kyle Lo
Constructing Character-Context-Coupled Space Aligned with Associative Knowledge Relations for Interpretable Language Modeling
Fanyu Wang, Zhenping Xie
Enhancing Out-of-Vocabulary Estimation with Subword Attention
Raj Patel, Carlotta Domeniconi
Parameter-Efficient Finetuning for Robust Continual Multilingual Learning
Kartikeya Badola, Shachi Dave, Partha Talukdar
Target-Oriented Relation Alignment for Cross-Lingual Stance Detection
Ruike Zhang, Nan Xu, Hanxuan Yang, Yuan Tian, Wenji Mao
Know Where You're Going: Meta-Learning for Parameter-Efficient Fine-Tuning
Mozhdeh Gheini, Xuezhe Ma, Jonathan May
Shielded Representations: Protecting Sensitive Attributes Through Iterative Gradient-Based Projection
Shadi Iskander, Kira Radinsky, Yonatan Belinkov
DEnsity: Open-domain Dialogue Evaluation Metric using Density Estimation
ChaeHun Park, Seungil Lee, Daniel Rim, Jaegul Choo
Impact of Adversarial Training on Robustness and Generalizability of Language Models
Enes Altinisik, Hassan Sajjad, Husrev Sencar, Safa Messaoud, Sanjay Chawla
Exploring the Relationship between Alignment and Cross-lingual Transfer in Multilingual Transformers
Felix Gaschi, Patricio Cerda, Parisa Rastin, Yannick Toussaint
Boosting Zero-shot Cross-lingual Retrieval by Training on Artificially Code-Switched Data
Robert Litschko, Ekaterina Artemova, Barbara Plank
Another Dead End for Morphological Tags? Perturbed Inputs and Parsing
Alberto Munoz-Ortiz, David Vilares
Tokenization Impacts Multilingual Language Modeling: Assessing Vocabulary Allocation and Overlap Across Languages
Tomasz Limisiewicz, Jiri Balhar, David Mareček
Modular and On-demand Bias Mitigation with Attribute-Removal Subnetworks
Lukas Hauzenberger, Shahed Masoudian, Deepak Kumar, Markus Schedl, Navid Rekabsaz
Nonparametric Decoding for Generative Retrieval
Hyunji Lee, JaeYoung Kim, Hoyeon Chang, hanseok Oh, Sohee Yang, Vladimir Karpukhin, Yi Lu, Minjoon Seo
On Dataset Transferability in Active Learning for Transformers
Fran Jelenic, Josip Jukic, Nina Drobac, Jan Snajder
Assessing Word Importance Using Models Trained for Semantic Tasks
David Javorsky, Ondrej Bojar, Francois Yvon
Yes, this Way! Learning to Ground Referring Expressions into Actions with Inter-episodic Feedback from Supportive Teachers
Philipp Sadler, Sherzod Hakimov, David Schlangen
Measuring Intersectional Biases in Historical Documents
Nadav Borenstein, Karolina Stanczak, Thea Rolskov, Natacha Klein Käfer, Natalia da Silva Perez, Isabelle Augenstein
Exploring anisotropy and outliers in multilingual language models for cross-lingual semantic sentence similarity
Katharina Haemmerl, Alina Fastowski, Jindrich Libovicky, Alexander Fraser
TADA -- Efficient Task-Agnostic Domain Adaptation for Transformers
Chia-Chien Hung, Lukas Lange, Jannik Strötgen
RQUGE: Reference-Free Metric for Evaluating Question Generation by Answering the Question
Alireza Mohammadshahi, Thomas Scialom, Majid Yazdani, Pouya Yanki, Angela Fan, James Henderson, Marzieh Saeidi
Layerwise universal adversarial attack on NLP models
Olga Tsymboi, Danil Malaev, Andrei Petrovskii, Ivan Oseledets
Exploiting Abstract Meaning Representation for Open-Domain Question Answering
Cunxiang Wang, Qipeng Guo, Zhikun Xu, Xiangkun Hu, Xuefeng Bai, Zheng Zhang, Yue Zhang
Masked Latent Semantic Modeling: an Efficient Pre-training Alternative to Masked Language Modeling
Gabor Berend
RFiD: Towards Rational Fusion-in-Decoder for Open-Domain Question Answering
Cunxiang Wang, Haofei Yu, Yue Zhang
Predicting Numerals in Text Using Nearest Neighbor Language Models
Taku Sakamoto, Akiko Aizawa
Easy to Decide, Hard to Agree: Reducing Disagreements Between Saliency Methods
Josip Jukic, Martin Tutek, Jan Snajder
History Repeats: Overcoming Catastrophic Forgetting For Event-Centric Temporal Knowledge Graph Completion
Mehrnoosh Mirtaheri, Mohammad Rostami, Aram Galstyan
Aligning Instruction Tasks Unlocks Large Language Models as Zero-Shot Relation Extractors
Zhang Kai, Bernal Jiménez Gutiérrez, Yu Su
Nonparametric Masked Language Modeling
Sewon Min, Weijia Shi, Mike Lewis, Xilun Chen, Wen-tau Yih, Hannaneh Hajishirzi, Luke Zettlemoyer
Reliable Gradient-free and Likelihood-free Prompt Tuning
Maohao Shen, Soumya Ghosh, Prasanna Sattigeri, Subhro Das, Yuheng Bu, Gregory Wornell
Detecting Edit Failures In Large Language Models: An Improved Specificity Benchmark
Jason Hoelscher-Obermaier, Julia H. Persson, Esben Kran, Ioannis Konstas, Fazl Barez
The Larger They Are, the Harder They Fail: Language Models do not Recognize Identifier Swaps in Python
Antonio Barone, Fazl Barez, Ioannis Konstas, Shay B. Cohen
Taxonomy of Problems in Lexical Semantics
Bradley Hauer, Grzegorz Kondrak
From chocolate bunny to chocolate crocodile: Do Language Models Understand Noun Compounds?
Albert Coil, Vered Shwartz
Sequential Path Signature Networks for Personalised Longitudinal Language Modeling
Talia Tseriotou, Adam Tsakalidis, Peter Foster, Terence Lyons, Maria Liakata
SamToNe: Improving Contrastive Loss for Dual Encoder Retrieval Models with Same Tower Negatives
Fedor Moiseev, Gustavo Hernandez Abrego, Peter Dornbach, Imed Zitouni, Enrique Alfonseca, Zhe Dong
Data-Efficient Finetuning Using Cross-Task Nearest Neighbors
Hamish Ivison, Noah A. Smith, Hannaneh Hajishirzi, Pradeep Dasigi
Honey, I Shrunk the Language: Language Model Behavior at Reduced Scale.
Vijeta Deshpande, Dan Pechi, Shree Thatte, Vladislav Lialin, Anna Rumshisky
Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and Evaluation
Marius Mosbach, Tiago Pimentel, Shauli Ravfogel, Dietrich Klakow, Yanai Elazar
SConE: Simplified Cone Embeddings with Symbolic Operators for Complex Logical Queries
Chau Nguyen, Tim French, Wei Liu, Michael Stewart
Recurrent Attention Networks for Long-text Modeling
Xianming Li, Zongxi Li, Xiaotian Luo, Haoran Xie, Xing Lee, Yingbin Zhao, Fu Lee Wang, Qing Li
RHO ($\rho$): Reducing Hallucination in Open-domain Dialogues with Knowledge Grounding
Ziwei Ji, Zihan Liu, Nayeon Lee, Tiezheng Yu, Bryan Wilie, Min Zeng, Pascale Fung
Causal interventions expose implicit situation models for commonsense language understanding
Takateru Yamakoshi, James McClelland, Adele Goldberg, Robert Hawkins
Multimedia Generative Script Learning for Task Planning
Qingyun Wang, Manling Li, Hou Pong Chan, Lifu Huang, Julia Hockenmaier, Girish Chowdhary, Heng Ji
Mini-Model Adaptation: Efficiently Extending Pretrained Models to New Languages via Aligned Shallow Training
Kelly Marchisio, Patrick Lewis, Yihong Chen, Mikel Artetxe
ECOLA: Enhancing Temporal Knowledge Embeddings with Contextualized Language Representations
Zhen Han, Ruotong Liao, Jindong Gu, Yao Zhang, Zifeng Ding, Yujia Gu, Heinz Koeppl, Hinrich Schütze, Volker Tresp
Enhancing Hierarchical Text Classification through Knowledge Graph Integration
Ye Liu, Zhenya Huang, Kai Zhang, Kehang Wang, Yanghai Zhang, Qi Liu, Enhong Chen
On the Expressivity Role of LayerNorm in Transformers' Attention
Shaked Brody, Uri Alon, Eran Yahav
On Isotropy, Contextualization and Learning Dynamics of Contrastive-based Sentence Representation Learning
Chenghao Xiao, Yang Long, Noura Al Moubayed
Context-Aware Document Simplification
Liam Cripwell, Joel Legrand, Claire Gardent
EmbedTextNet: Dimension Reduction with Weighted Reconstruction and Correlation Losses for Efficient Text Embedding
Dae Yon Hwang, Bilal Taha, Yaroslav Nechaev
Revisiting the Architectures like Pointer Networks to Efficiently Improve the Next Word Distribution, Summarization Factuality, and Beyond
Haw-Shiuan Chang, Zonghai Yao, Alolika Gon, hong yu, Andrew McCallum
Byte-Pair Encoding is Approximately Optimal
Vilem Zouhar, Tim Vieira, Clara Meister, Juan Gastaldi, Mrinmaya Sachan, Ryan Cotterell
Nano: Nested Human-in-the-Loop Reward Learning for Few-shot Language Model Control
Xiang Fan, Yiwei Lyu, Paul Pu Liang, Ruslan Salakhutdinov, Louis-Philippe Morency