Published articles:
D'Andrea, M. (2023) Playing against no-regret players. Operations Research Letters, Vol. 51, Issue 2, pages 142-145.
In increasingly different contexts, it happens that a human player has to interact with artificial players who make decisions following decision-making algorithms. How should the human player play against these algorithms to maximize his utility? Does anything change if he faces one or more artificial players? The main goal of the paper is to answer these two questions. Consider n-player games in normal form repeated over time, where we call the human player optimizer, and the (n - 1) artificial players, learners. We assume that learners play no-regret algorithms and we consider the concept of Stackelberg equilibrium. In a 2-player game, the optimizer can always guarantee an expected average utility of at least the Stackelberg value per round. However, we show, with counterexamples, that this result does not hold if the optimizer has to face more than one player. Therefore, we generalize the definition of Stackelberg equilibrium introducing the concept of correlated Stackelberg equilibrium. Finally, in the main result, we prove that the optimizer can guarantee at least the correlated Stackelberg value per round. Moreover, using a version of the strong law of large numbers, we show that our result is also true almost surely for the optimizer utility instead of the optimizer's expected utility.
D'Andrea, M., Gensbittel, F., Renault, J. (2024) Games played by Exponential Weights Algorithms. (Under review)
This paper studies the last-iterate convergence properties of the exponential weights algorithm with constant learning rates. We consider a repeated interaction in discrete time, where each player uses an exponential weights algorithm characterized by an initial mixed action and a fixed learning rate, so that the mixed action profile p^t played at stage t follows a homogeneous Markov chain. At first, we show that whenever a strict Nash equilibrium exists, the probability to play a strict Nash equilibrium at the next stage converges almost surely to 0 or 1. Secondly, we show that the limit of p^t, whenever it exists, belongs to the set of "Nash Equilibria with Equalizing Payoffs''. Thirdly, we show that in strong coordination games, where the payoff of a player is positive on the diagonal and 0 elsewhere, p^t converges almost surely to one of the strict Nash equilibria. We conclude with open questions.
Work in Progress:
D'Andrea, M., Scarsini, M. Strategic Queue with priorities.
Ph.D. manuscript: Learning Algorithms in Game-Theoretic Contexts.