Dr. Robert Loftin, Delft University of Technology (TU Delft)

Talk Date and Time: November 3, 2022 at 04:00 pm - 04:45 pm EST followed by 10 minutes of Q&A on Zoom and IRB-5105

Topic: On the Theoretical Limits of Ad-Hoc Cooperation in Multi-Agent Reinforcement Learning


Abstract:

Learning to cooperate with other agents (human or AI) will be essential for intelligent systems to succeed in many real-world applications. The challenge is that while we (or the agents we design) are trying to learn to cooperate with other agents, those agents may simultaneously be learning to cooperate with us. Practical and theoretical approaches to learning in cooperative settings therefore assume that other agents' behaviors are stationary, or else make very specific assumptions about other agents' learning processes. The goal of our work is to understand whether we can reliably learn to cooperate with others without such restrictive assumptions, which are unlikely to hold in real-world settings. In this talk, I will discuss a set of recent impossibility results which show that, in fact, no learning algorithm can be guaranteed to learn to cooperate with all possible adaptive partners. This is true even when every possible partner is guaranteed to eventually cooperate with some fixed strategy that our learner could have committed to. This talk will conclude with a discussion of the practical implications of these results for human-AI cooperation, and potential theoretical solutions to the issues they raise.

Bio:

Robert Loftin completed his PhD in computer science in 2019 at North Carolina State University under the supervision of David L. Roberts. He was previously a post-doc researcher at Microsoft Research Cambridge, and is currently doing a post-doc at TU Delft (Delft University of Technology) with Frans Oliehoek. His research interests include interactive machine learning, human-computer interaction, multi-agent reinforcement learning, and deep reinforcement learning.