Talk Date and Time: October 6, 2022 at 04:00 pm - 04:45 pm EST followed by 10 minutes of Q&A on Zoom and IRB-5105
Topic: Reinforcement Learning Beyond Agents With The Same Design
Abstract:
The standard approach to reinforcement learning is to fit policies and value functions specialized to a single agent design. This specialization precludes transferring existing models to agents with new designs that differ from the current one. The effect is that when parts on a robot break, or that robot is due for an upgrade, a new model must typically be trained from scratch, which is impractical for both the researcher and the end user. In this talk, I discuss the challenge of learning models that generalize beyond agents with the same design, and present our recent work on learning transferable policies by inferring agent morphology. Prior work has either assumed the physical connectivity of the agent (ie, morphology) is known in advance, or ignores this information. Instead, we propose to consider morphology a feature to be inferred during training, and show this relaxes assumptions in prior work and improves performance. The key idea behind our approach is to encode morphology as a sequence of discrete tokens with learned token representations, and process them with a Transformer-based policy. Our results on a standard benchmark suggest our Transformer generalizes better than current methods, and benefits from large datasets containing many unique agent designs.
Bio:
Brandon is a second-year PhD Student in the Machine Learning Department at the Carnegie Mellon University, Pittsburgh (CMU) advised by Prof. Ruslan (Russ) Salakhutdinov. Brandon's work focuses on using large multi-task models to accelerate Reinforcement Learning, focusing on applications in Embodied AI. His goal is to build general agents that help and work alongside us in our daily lives.