Language Reward Modulation for Pretraining Reinforcement Learning
A. Adeniji, A. Xie, C. Sferrazza, Y. Seo, S. James, P. Abbeel
I. S. Sheidlower, R. M. Aronson, E. Short
Informing Reinforcement Learning Agents by Grounding Language to Markov Decision Processes
B. A. Spiegel, Z. Yang, W. Jurayj, B. Bachmann, S. Tellex, G. Konidaris
Concept-Based Interpretable Reinforcement Learning with Limited to No Human Labels
Z. Ye, S. Milani, F. Fang, G. Gordon
Transformers Learn Temporal Difference Methods for In-Context Reinforcement Learning
J. Wang, E. H. Blaser, H. Daneshmand, S. Zhang
SENSEI: Semantic Exploration Guided by Foundation Models to Learn Versatile World Models
C. Sancaktar, C. Gumbsch, A. Zadaianchuk, P. Kolev, G. Martius
G. H. Sarch, L. Jang, M. J. Tarr, W. W. Cohen, K. Marino, K. Fragkiadaki
Make the Pertinent Salient: Task-Relevant Reconstruction for Visual Control with Distraction
K. Kim, C. Fowlkes, R. Fox
[2-min video 🎥]
Pre-trained Language Models Improve the Few-shot Prompt Ability of Decision Transformer
Y. Yang, P. Xu
X. Zhang, P. Becker-Ehmck, P. van der Smagt, M. Karl
[2-min video 🎥]
Language Model-In-The-Loop: Data Optimal Approach to Recommend Actions in Text Games
A. V. Sudhakar, P. Parthasarathi, J. Rajendran, S. Chandar
ORSO: Accelerating Reward Design via Online Reward Selection and Policy Optimization
C. B. C. Zhang, Z.-W. Hong, A. Pacchiano, P. Agrawal
Skill-Based Reinforcement Learning with Intrinsic Reward Matching
A. Adeniji, A. Xie, P. Abbeel
KalMamba: Towards Efficient Probabilistic State Space Models for RL under Uncertainty
P. Becker, N. Freymuth, G. Neumann
[2-min video 🎥]
Collaborative Embodied Reasoning in Autonomous Driving
O. Chang, A. A. Kamat, W. Self
Dynamics Generalisation with Behaviour Foundation Models
S. Jeen, J. Cullen
Policy Learning with a Language Bottleneck
M. Srivastava, C. Colas, D. Sadigh, J. Andreas
LLM Policies for Text-based Reinforcement Learning: An Interactive Tutorial
M. Tec