EC 2024 Workshop on
Foundation Models and Game Theory


July 8th, 2024 at Yale University in New Haven 

Workshop Theme

Foundational models (FMs) such as Large Language Models (LLMs) are machine learning models that have been pre-trained with large datasets and can be efficiently adapted to various downstream tasks.  Given the tremendous impacts that FMs have been making, FMs could be a key step towards building AI agents that can autonomously make decisions to cooperate or compete with other AI agents and humans to achieve their ultimate goals in complex multi-agent environments.  In particular, FMs have been pushing the limits of the complexity of the environments in which such AI agents interact via natural language and multimodal data.  However, in many cases, the strategic reasoning capabilities of such agents has been severely lacking.  In parallel, game theory has been laying the strong foundation for the strategic reasoning capabilities of such AI agents.

It is therefore the goal of this workshop to bring together researchers working on foundation models, game theory, and their intersection to exchange ideas on how to push the frontiers of foundation models and game theory towards realizing better (e.g., more efficient, more sustainable, happier, safer) society with strategic AI agents.  The topics of the workshop  include but are not limited to:

This workshop builds on the DIMACS Workshop on Foundation Models, Large Language Models, and Game Theory, which was held on October 19-20, 2023.

Program

8:00 - 8:30 Breakfast

8:30 - 10:00 Session 1

Abstract: There is increasing interest in using LLMs as decision-making "agents." Doing so includes many degrees of freedom: which model should be used; how should it be prompted; should it be asked to introspect, conduct chain-of-thought reasoning, etc? Settling these questions -- and more broadly, determining whether an LLM agent is reliable enough to be trusted -- requires a methodology for assessing such an agent's economic rationality. This talk describes an effort to provide one. Our approach begins by surveying the economic literature on rational decision making, taxonomizing a large set of fine-grained "elements of rationality" that an agent should exhibit, along with dependencies between them. We then propose a benchmark distribution that quantitatively scores an LLM's performance on these elements and, combined with a user-provided rubric, produces a "STEER report card." Finally, we describe the results of a large-scale empirical experiment with 20 different LLMs, characterizing the both current state of the art and the impact of different model sizes on models' ability to exhibit rational behavior.

10:00 - 10:30 Coffee break

10:30 - 11:00 Session 2

11:00 - 12:00 Poster session

12:00 - 1:00 Lunch

1:00 - 3:00 Session 3

Abstract: We investigate auction mechanisms for AI-generated content, focusing on applications like ad creative generation. In our model, agents’ preferences over stochastically generated content are encoded as large language models (LLMs). We propose an auction format that operates on a token-by-token basis, and allows LLM agents to influence content creation through single dimensional bids. We formulate two desirable incentive properties and prove their equivalence to a monotonicity condition on output aggregation. This equivalence enables a second-price rule design, even absent explicit agent valuation functions. Our design is supported by demonstrations on a publicly available LLM.

Joint work with Vahab Mirrokni (Google Research), Renato Paes Leme (Google Research), Song Zuo (Google Research), and Haifeng Xu (U Chicago and Google Research)

3:00 - 3:30 Coffee break

3:30 - 5:30 Session 4

Keynote speakers

Kevin Leyton-Brown

Kevin Leyton-Brown is a professor of Computer Science and a Distinguished University Scholar at the University of British Columbia. He holds a Canada CIFAR AI Chair at the Alberta Machine Intelligence Institute and is an associate member of the Vancouver School of Economics. He is a Fellow of the Royal Society of Canada, the ACM, and AAAI. He studies artificial intelligence, mostly at the intersection of machine learning with either the design and operation of electronic markets or the design of heuristic algorithms. He is increasingly interested in large language models, particularly as components of agent architectures.

Paul Dütting

Paul Duetting is a Senior Research Scientist at Google Switzerland, specializing in the intersection of Algorithms, Contract Theory, and Mechanism Design. Previously, Paul was an Associate Professor of Mathematics at the London School of Economics, where he remains a visiting faculty member. Paul earned a PhD in Computer Science from EPFL Lausanne under the guidance of Monika Henzinger, and has held postdoctoral positions at Stanford University, Cornell University, and ETH Zurich. Paul has received numerous awards for his research, including Best Paper Awards at WWW 2024, EC 2019, and EC 2012, as well as the 2017/18 LSE Excellence in Education Award.

Accepted papers

Organizers

Kate Larson
University of Waterloo
Google DeepMind

Takayuki Osogami
IBM Research

David C. Parkes
Harvard University

David Pennock
Rutgers University

Segev Wasserkrug
IBM Research

Faculty of Data and Decision Sciences, Technion

Call for Papers

Each submission will be reviewed from various perspectives, with particular emphasis on the alignment with the workshop theme.  We will accept a mixture of poster papers and papers that will be presented orally.

Timetable for Authors

May 21May 22, 2024 (11:59pm AoE): Paper submission deadline.

May 31, 2024: Paper accept/reject notifications.

July 8, 2024: Workshop.

Submission Instructions

We encourage authors to format their submissions as instructed in the call for papers of EC'24, but papers in other formats will also be reviewed.  However, keep in mind that each paper will be reviewed at most a few hours.  In addition to standard technical papers (those typically submitted to the main conference of EC'24), we welcome short papers and extended abstracts (of arbitrary length) with preliminary results as well as position papers that advocate positions on topics relevant to the workshops.

Papers and abstracts should be submitted through OpenReview.  If you do not yet have your OpenReview profile, please make sure to create one with an institutional email (unless you have more than two weeks before submission).  Here is OpenReview's moderation policy: