Professor, Department of Brain and Cognitive Sciences, MIT
Title: What is Theory of Mind, and how would you know if a system had one?
Assistant Professor, Faculty of Science and Engineering, University of Groningen
Title: How is higher-order theory of mind reasoning beneficial in negotiations?
Abstract:
The ability to use (higher-order) theory of mind reasoning has been shown to have the potential for effecting better outcomes across a variety of situations. Despite these apparent benefits, however, humans represent the only species we know to be capable of higher-order theory of mind reasoning. In this talk, we investigate the way higher-order theory of mind contributes to achieving better outcomes, by comparing performance across variations of a particular negotiation setting known as Colored Trails. We also consider the ability of agents to signal their goal location to their trading partner. Simulation experiments show that agents may misrepresent their goal location when given the option to do so. However, such behaviour does not result in better outcomes for the agent.
Professor, Department of Computer Science, University of Toronto
Title: Purposeful Theory of Mind
Abstract: Theory of Mind is the ability to attribute mental states, such as beliefs and goals, as well as affective states, such as emotions, to oneself and to others. Research on humans has identified Theory of Mind as one of several social cognitive abilities that enable most humans to successfully understand and interact with others. In this talk we will explore ways in which AI agents can be endowed with (some) Theory of Mind capabilities, and how these capabilities can be used purposefully by these AI agents to enable or enhance their ability to perform a variety of tasks including diagnosis, explanation, assistance, and plan recognition.
Joint work with Maayan Shvo, Toryn Klassen, Ruthrash Hari, Christian Muise, Jorge Baier, Shrin Sohrabi, and others.
Professor, Department of Brain and Cognitive Sciences, MIT
Title: Engineering and reverse-engineering theories of mind for human and cooperative AI agents
Abstract: I will briefly present the computational and cognitive science foundations of engineering systems that attempt to simulate and scale human minds' intuitive theories of other minds, broadly known under the heading of "Bayesian Inverse Planning". Some of these ideas will have been introduced in Rebecca Saxe's morning keynote, in joint work we have done with the goals of scientific theory-building and theory-testing in cognitive science: Can we build a machine-implementable model of how human minds reason about others' mental states that is quantitatively predictive and explanatory, and that is robust and general enough to be the basis for scientifically grounded technologies and interventions that improve our lives? I will then discuss how, in an AI context, this work could provide foundations for building machines that think, learn and act productively and cooperatively with people, in ways that we humans have come to expect in our interactions with other humans -- both trusted friends and people we are just meeting for the first time. I will highlight recent progress based on advances in probabilistic programming languages, language models, and their integration, that together offer the potential to scale Bayesian approaches to inverse planning from the very simple settings where these ideas have been worked out in cognitive science to much more complex, real-world settings and deeper recursive forms of reasoning. I will show briefly how this work provides the basis for cognitive science accounts (and potentially scalable AI technologies) that capture many aspects of human social intelligence, such as communication, storytelling, norm formation and adaptation, social evaluation ,and moral judgment and decision-making.