Reinforcement Learning-Based Whole-Body Control for Dynamic Manipulation of Passively Connected Objects with a
Legged-Arm Robot
Period : 24.06 ~ 24.10
Reinforcement Learning-Based Whole-Body Control for Dynamic Manipulation of Passively Connected Objects with a
Legged-Arm Robot
Period : 24.06 ~ 24.10
Goal of this project
Dynamic manipulation of passively connected objects by legged-arm robots presents unique challenges due to nonlinearities, such as a floating base and the passive nature of the end effector. Although learning-based approaches offer potential for handling these complex dynamics, they often require extensive data, making them computationally demanding. On the other hand, optimization-based methods face limitations when dealing with unknown objects and adapting to variable environments. This paper proposes a hybrid control framework that selectively applies learning only to the end-effector trajectory while utilizing Whole-Body Impulse Control (WBIC) with Model Predictive Control (MPC) for overall stability and motion tracking. The proposed framework highlights the potential of RL-integrated control in reducing the learning burden and improving control efficiency and robustness in autonomous material-handling tasks in complex setting.
Publications
Reinforcement Learning-Based Whole-Body Control for Dynamic Manipulation of Passively Connected Objects with a Legged-Arm Robot
J. P. Jang, S. H. Hwang, Y.S. Chio, S. W. Hwang, #W. S. Kim
The Korea Robotics Society Annual Conference (KRoC), 2025
Videos
Demonstration : Achieving Stability with the Inverted Pendulum by tracking a end-effector trajectory created by RL
Tasks assigned to the controller
Contact constraint
Body orientation
CoM position
End-effector 6D position & velocity
Demonstration : Learning dynamics of the inverted pendulums through Soft Actor Critic(SAC)
Training environment
State : cart position,velocity, pole angle,angular velocity
Action : desired cart position and velocity
Reward: 1 + pole_angle_penalty + state limit penalties
Controller : F = K*(desired cart position - cart position) + D*(desired cart velocity - cart velocity)