Generative Modeling meets HRI

GenAI-HRI


RSS 2024 Workshop


Jul 15, 2024 @ Delft, Netherlands

News: 

04/15: Call for papers is out! Submission deadline May 28th, AoE

In recent years, the rapid evolution of generative modeling, encompassing Variational Autoencoders (VAEs), Generative Adversarial Networks (GANs), and Diffusion Models, among others, has reshaped the landscape of image synthesis, enabling the generation of highly realistic and diverse visual content. Inspired by this progress, our workshop proposal seeks to explore the potential of applying generative modeling techniques to enhance human-robot interaction (HRI). We aim to gather the communities of robot learning and Human-Robot Interaction to discuss cutting-edge generative model techniques, modeling of human behaviors and robot motions, and opportunities to use them to achieve more intuitive human-robot interactions for robot teaching, data collection, language-base interactions and collaborative execution. 


Why are generative models important for research in HRI? HRI will benefit greatly from powerful large models that bring open-world knowledge and generalization to the classic HRI interaction workflows. Just as ChatGPT has become of popular use for non-technical users, it’s only a matter of time before these types of large models with vision and language capabilities will play a key role in generating and mediating interaction between humans and robots in daily life settings (home robots learning your home tasks from examples) and industrial deployments (co-bots in manufacturing). Generative models are also key for the creation of simulation environments (3D assets, scenes, tasks, language commands, and language-based task generation), and simulation environments are useful for data collection of human demonstrations, data generation, and policy training. It’s important for HRI researchers to foster collaborations that investigate how multi-agent interactions and human-like behaviors will play a role in these systems, whether in simulation or real settings. 


Why is HRI important to research in generative models? Conversely, HRI is pivotal for advancing research in generative models. Human interaction and feedback are essential for producing high-quality data for learning and value-aligned training. For example, reinforcement learning from human feedback (RLHF) has demonstrated significant advancements in model performance, enabling ChatGPT’s performance to surpass models learned from static language datasets. Generative models applied to robotics are fundamentally tied to human interaction. In data collection pipelines, we need to provide users with tools, methods, and interfaces to provide and curate high-quality data that can be used by learning algorithms. For model improvements, we need human feedback in the loop with policy learning iterations of fine-tuning during deployment. These are all core interaction problems that are studied in HRI and are now prime to be used in the loop with generative AI in both training and inference, bringing the knowledge from interactions and human-centered modeling into robot learning. 

 Topics

Motion and Behavior Modelling and Generation 


Human Interaction for Goal Specification


Large Language Models (LLMs) and Vision Language Models (VLMs) in HRI


AI-HRI Safety and Alignment 

Call For Papers

Authors are invited to submit short papers (3-4 pages excluding references) covering topics on generative modeling applied to human-robot interaction. We invite contributions describing on-going research, results that build on previously presented work, systems, datasets and benchmarks, and papers with demos (that could be displayed easily next to a poster).


Submission link https://openreview.net/group?id=roboticsfoundation.org/RSS/2024/Workshop/GenAI-HRI 


Submission

Submissions should use the official RSS LaTeX template.  Reviews will be single blind. Accepted papers will be presented as posters during the workshop and selected works will have an opportunity for a spotlight talk. Accepted papers will be available online on the workshop website (non-archival). A best paper award will be sponsored by NVIDIA. 


Important Dates 

Submission deadline: May 28th, 2024 (Anywhere on Earth [AoE])

Notification deadline: June 14th, 2024 (AoE)

Camera-ready deadline: July 1st, 2024 (AoE)

Workshop date: July 15th, 2024 (Full day)

Speakers and Panelists

Yilun Du

MIT

Andrea Bajcsy

CMU

Ben Burchfield

TRI

Siyuan Feng

TRI

Ted Xiao

Google DeepMind

Tesca Fitzgerald

Yale

Huazhe Xu

Tsinghua

Danica Kragic

KTH

Jens Kober

TU Delft

Maya Cakmak

UW

 Agenda


Full schedule to be posted soon 


Organizers

Felix Yanwei Wang

MIT

Sidd Karamcheti

Stanford

Vignesh Prasad

TU Darmstadt

Georgia Chalvatzaki

TU Darmstadt

Pete Florence

Google DeepMind

Nadia Figueroa

UPenn

Fabio Ramos

University of Sydney / NVIDIA

Julie Shah

MIT

Claudia D'Arpino

NVIDIA