The Algorithmic Learning in Games seminar is a joint effort between faculty at London School of Economics, the University of Waterloo, Imperial College London and, King's College London.
We meet once every fortnight to discuss problems and research in a range of topics related to:
Game Theory
Multi-Agent Reinforcement Learning
Learning in Games
Multi-Agent Systems
Recordings of our past sessions can be found here.
Next Talk
Our next talk will be a special PhD student session lead by Ed Plumb and Cole Wyeth on October 27th 2025 at 16:00 GMT. Please see further details below:
Ed Plumb - PhD Candidate, London School of Economics
Co-authors: Galit Ashkenazi-Golan, Domenico Mergoni Cecchelli
This paper examines the convergence behaviour of simultaneous best-response dynamics in random potential games. We provide a theoretical result showing that, for two-player games with sufficiently many actions, the dynamics converge quickly to a cycle of length two. This cycle lies within the intersection of the neighbourhoods of two distinct Nash equilibria. For three players or more, simulations show that the dynamics converge quickly to a Nash equilibrium with high probability. Furthermore, we show that all these results are robust, in the sense that they hold in non-potential games, provided the players' payoffs are sufficiently correlated. We also compare these dynamics to gradient-based learning methods in near-potential games with three players or more, and observe that simultaneous best-response dynamics converge to a Nash equilibrium of comparable payoff substantially faster.
Cole Wyeth - PhD Candidate, University of Waterloo
Co-authors: Marcus Hutter, Jan Leike, Jessica Taylor
A Bayesian player acting in an infinite multi-player game learns to predict the other players' strategies if his prior assigns positive probability to their play (or contains a grain of truth). Kalai and Lehrer's classic grain of truth problem is to find a reasonably large class of strategies that contains the Bayes-optimal policies with respect to this class, allowing mutually-consistent beliefs about strategy choice that obey the rules of Bayesian inference. Only small classes are known to have a grain of truth and the literature contains several related impossibility results. In this paper we present a formal and general solution to the full grain of truth problem: we construct a class of strategies wide enough to contain all computable strategies as well as Bayes-optimal strategies for every reasonable prior over the class. When the "environment" is a known repeated stage game, we show convergence in the sense of [KL93a] and [KL93b]. When the environment is unknown, agents using Thompson sampling converge to play -Nash equilibria in arbitrary unknown computable multi-agent environments. Finally, we include an application to self-predictive policies that avoid planning. While these results use computability theory only as a conceptual tool to solve a classic game theory problem, we show that our solution can naturally be computationally approximated arbitrarily closely.
Organisers
London School of Economics
University of Waterloo
Imperial College London
King's College London
Schedule
We intend to meet every other Monday from 16:00 - 17:00 GMT - please allow for slight changes to this schedule as needed.
Upcoming Speakers
Past Speakers
Recordings of past sessions can be found here or via the title of the respective talk in the spreadsheet below.
Sign-Up Form
We are open to any interested in joining the seminar, both as a member of the audience and as a speaker. To get in touch, please fill out the form below and we will get back to you as soon as possible.