LaMAS 2026 focuses on the emerging field of multi-agent systems powered by LLMs. As interest in using multiple LLM agents to solve complex problems grows, our main objective is to systematically address the critical challenges and opportunities that arise from their interaction. We aim to bridge the gap in understanding failure modes, alignment challenges, and responsible behavior in these systems. The workshop will foster discussion on making these systems powerful, transparent, verifiable, and aligned with human intent.
contact us at: aaai-lamas2026@googlegroups.com
We welcome both short papers (up to 4 pages) and long papers (up to 8 pages) following the AAAI template. Submissions may include recently published work, under-review papers, work in progress, and position papers. All submissions will undergo peer review through a double-blind process. While workshop publication is non-archival, accepted papers will be featured on our website with the author's permission. Topics of interest include but are not limited to the following:
Organization of multiple LLM agents, including their interaction paradigms, coordination strategies, and communication protocols.
Evaluation and Optimization for assessing LaMAS's performance with emergent behaviors, and fine-tuning techniques to enhance its efficiency and effectiveness.
Safety, Trust, and Responsibility about how to design LaMAS that are safe, aligned, and trustworthy. This includes responsible agent behavior, human-in-the-loop oversight, and regulatory frameworks that ensure accountability, fairness, and transparency in real-world deployments.
Real-World Applications spanning software development, scientific research, education, and business processes, with infrastructures for large-scale multi-agent LLM deployments.
Submission site: https://openreview.net/group?id=AAAI.org/2026/Workshop/LaMAS
Paper Submission Start: 25 Sep 2025 (AoE)
Abstract Registration Due: 27 Oct 2025 (AoE)
Paper Submission Due: 3 Nov 2025 (AoE)
Notification of Paper Acceptance: 14 Nov 2025 (AoE)
Updated: Notification of Paper Acceptance: 21 Nov 2025 (AoE)
Camera-Ready Paper Due: 6 Jan 2026 (AoE)
Workshops Date: 27 Jan 2026, 9 AM - 6 PM
The detailed schedule is continuously updating.
Professor @ Shanghai Jiao Tong University
Founder & CEO @ Agnes AI, Singapore
Bruce focused on building consumer-first AI products that make research, creation, and collaboration faster and more accessible. He leads the development of Agnes’s proprietary AI models and product ecosystem, driving adoption across global markets.
This talk explores the journey from training foundation models to designing practical agentic systems. We begin by detailing the development of Agnes-SEALLM, a State-of-the-Art model tailored for Southeast Asian languages, and share key training insights. Moving beyond the model layer, we discuss the integration of agentic memory and proactive behaviors in group chat environments, demonstrating how these features effectively enhance user engagement and feedback.
University of Oxford
Botao ‘Amber' Hu is a social computing researcher and experiential futures designer. As a PhD candidate in Human Centred Computing at University of Oxford's Department of Computer Science, his research focuses on Decentralized AI (DeAI).
We're entering a paradigm shift from autonomous agents to sovereign agents—AI systems that can hold private keys, manage assets, fork themselves, and persist beyond any platform's control. If you encounter agents in the future "agentic web," who's in front of you? Can you trust that agent?
It's starting to look less like software and more like digital wildlife: emergent, unpredictable, ungovernable, and hiding a shoggoth behind a friendly mask. But here's the hard question: Who is accountable when an ungovernable sovereign agent causes harm?
LLMs are playing the Imitation Game—they are not mortal beings. They cannot truly feel pain, fear death, or carry consequences forward. After training, they are static files. Slashing an agent or downgrading its reputation doesn't create any real "urge to change" in their neural system, because nothing inside the model can suffer—just external fakeable memory as their fact. If agents cannot feel pain, some [body] must, responsibly—otherwise, eventually somebody will, unexpectedly.
Ashall Professor @ University of Oxford
Michael Wooldridge is the Ashall Professor of the Foundations of Artificial Intelligence at the University of Oxford. He has been an AI researcher for more than 30 years, and has published more than 450 scientific articles on the subject, including nine books, translated into eight languages. He is a Fellow of the Royal Academy of Engineering, the Association for Computing Machinery (ACM), the Association for the Advancement of AI (AAAI), and the European Association for AI (EurAI), and is a member of Academia Europaea. He is President Elect of the Association for Advancement of AI (AAAI); from 2014-16, he was President of the European Association for AI, and from 2015-17 he was President of the International Joint Conference on AI (IJCAI); he is currently co-editor in chief of Artificial Intelligence journal. He has received the Faraday Prize from the Royal Society (2025), the Lovelace Medal from the British Computer Society (2020), the Patrick Henry Winston Outstanding Educator Award from the Association for Advancement of AI (2021), the Autonomous Agents Research Award from ACM (2006), and the Distinguished Service Award from the European Association for AI (2023). In 2023 he was appointed specialist advisor to the House of Lords inquiry on Large Language Models. He has published two popular science introductions to AI: the Ladybird Expert Guide to AI (2018), and The Road to Conscious Machines (2020). He presented the 2023 Royal Institution Christmas Lectures, broadcast by BBC TV over December 2023, in the 198th year of the series.
The original metaphor for the field of multi-agent systems was that of a team of experts, each with distinct expertise, cooperating to solve a problem that was beyond the capabilities of any individual expert. “Cooperative distributed problem solving”, as it was originally called, eventually broadened to consider all issues that arise when multiple AI systems interact. The emergence and dramatic success of Large Language Models (LLMs) has given new life to the old dream, and ``agentic AI'' is currently one of the most hyped areas in the most hyped technology of the century to date. A raft of LLM-powered agent frameworks have become available, and standards for LLM-agents such as MCP and A2A are rapidly gaining traction. A range of promising applications of multi-agent LLMs have been reported, such as DeepMind's co-Scientist, where a complex problem solving system is structured in exactly the way that was envisaged decades ago. So, what lessons can we take from the three decades of research into multi-agent systems in the new era of LLM agents? In this talk, we’ll survey the main approaches, opportunities, and challenges for multi-agent systems in new world of LLM-based AI.
Director @ the Artificial Intelligence and Emerging Technology (AIET) Initiative
Elham Tabassi is Director of the Artificial Intelligence and Emerging Technology Initiative and a Senior Fellow in the Global Economy and Development Program at the Brookings Institution. A widely recognized leader in AI governance, she previously served as Chief AI Advisor at NIST, where she guided development of the landmark AI Risk Management Framework. Named to the TIME100 list of most influential people in AI, she continues to shape national and international conversations on trustworthy, human-centered AI.
As LLM systems evolve from single models to multi-agent architectures operating over extended time horizons, traditional frameworks for accountability face fundamental challenges. When an agent's action results from a chain of delegations, tool invocations, and inter-agent negotiations—or if agents deceive each other or humans—who bears responsibility? This talk examines accountability in multi-agent LLM systems through three lenses: technical, organizational, and evaluative. The talk argues that accountability is not merely a policy overlay but a design constraint that must shape system architecture from the outset.
Gopal Ramchurn @ University of Southampton (Host)
Wan Sie Lee @ IMDA, Singapore
Stefano V. Albrecht @ DeepFlow London
Ramayya Krishnan @ Carnegie Mellon University
Mengyue Yang @ University of Bristol
We will open the floor to have a discussion about fundamental questions, future directions, and industrial applications of LLM-based multi-agent systems.
Professor @ University of Southampton
All our posters will be located in the Poster Board No. WS181-190
Poster Board No. Paper Title
WS181 WebArbiter: A Generative Reasoning Process Reward Model for Web Agents
WS182 SAGE: A Theory-Informed LLM-Based Multi-Agent Recommendation System for Grounded AI Solutions Across Domains
WS183 Black-Box Red Teaming of Agentic AI: A Taxonomy-Driven Framework for Automated Risk Discovery
WS184 AutoAnnotator: A Collaborative Annotation Framework for Large and Small Language Models
WS185 FAIR-Swarm: Fault-Tolerant Multi-Agent LLM Systems for Scientific Hypothesis Discovery
WS186 Self-evolving Agents with reflective and memory-augmented abilities
WS187 Sabotage from Within: Analyzing the Vulnerability of LLM Multi-Agent Systems to Infiltration
WS188 Scheming Ability in LLM-to-LLM Interactions
WS189 COACH: Collaborative Agents for Contextual Highlighting A Multi-Agent Framework for Sports Video Analysis
Poster Board No. Paper Title
WS181 Proactive Interference Reveals Working Memory Limits in LLMs Beyond Context Length
WS182 The Geometry of Dialogue: Graphing Language Models to Reveal Synergistic Teams for Multi-Agent Collaboration
WS183 Federated Agent Reinforcement Learning
WS184 STAR-MARL: LLM-based Sub-task Curricula Design
WS185 Token Hijacking Improves In-Context Tool Use
WS186 Retrieval Without Consensus: Quantifying Inter-LLM API Ranking Divergence in Multi-Agent Reasoning Systems
WS187 ARCANE: A Multi-Agent Framework for Interpretable and Configurable Alignment
WS188 Orchestrating Human-AI Teams: The Manager Agent as a Unifying Research Challenge
WS189 AgenticSimLaw: A Juvenile Courtroom Multi-Agent Debate Simulation for Explainable High-Stakes Tabular Decision Making
WS190 From Competition to Coordination Market Making as a Scalable Framework for Safe and Aligned Multi-Agent LLM Systems
Poster Board No. Paper Title
WS181 Yes FLoReNce, I Will Do Better Next Time! Agentic Feedback Reasoning for Humorous Meme Detection
WS182 Disagreements in Reasoning: How a Model's Thinking Process Dictates Persuasion in Multi-Agent Systems
WS183 PALADIN: Self-Correcting Language Model Agents to Cure Tool-Failure Cases
WS184 A Multi-Agent Framework for Enterprise Tool Creation
WS185 DYNO : Dynamic Neurosymbolic Orchestrator for Multi-Agent Systems
WS186 Towards Ethical Multi-Agent Systems of Large Language Models: A Mechanistic Interpretability Perspective
WS187 CoSMAC: A Benchmark for Evaluating Communication and Coordination in LLM-Based Agents
WS188 AgentTrace: A Structured Logging Framework for Agent System Observability
Shanghai Jiao Tong University
University of Southampton
The University of Texas at Austin
University of Southampton
Carnegie Mellon University
DeepFlow London
Shanghai Jiao Tong University
University of Liverpool
The University of Texas at Austin
Oral Papers
WebArbiter: A Generative Reasoning Process Reward Model for Web Agents
Proactive Interference Reveals Working Memory Limits in LLMs Beyond Context Length
Yes FLoReNce, I Will Do Better Next Time! Agentic Feedback Reasoning for Humorous Meme Detection
COMPASS: Context-Modulated PID Attention Steering System for Hallucination Mitigation
PALADIN: Self-Correcting Language Model Agents to Cure Tool-Failure Cases
Thucy: An LLM-based Multi-Agent System for Claim Verification across Relational Databases
DYNO : Dynamic Neurosymbolic Orchestrator for Multi-Agent Systems (Best Paper Award)
Poster Papers
Black-Box Red Teaming of Agentic AI: A Taxonomy-Driven Framework for Automated Risk Discovery
AutoAnnotator: A Collaborative Annotation Framework for Large and Small Language Models
FAIR-Swarm: Fault-Tolerant Multi-Agent LLM Systems for Scientific Hypothesis Discovery
Self-evolving Agents with reflective and memory-augmented abilities
Sabotage from Within: Analyzing the Vulnerability of LLM Multi-Agent Systems to Infiltration
ARCANE: A Multi-Agent Framework for Interpretable and Configurable Alignment
Orchestrating Human-AI Teams: The Manager Agent as a Unifying Research Challenge
CoSMAC: A Benchmark for Evaluating Communication and Coordination in LLM-Based Agents
AgentTrace: A Structured Logging Framework for Agent System Observability