Nonholonomic Yaw Control of an Underactuated Flying Robot with Model-based Reinforcement Learning

Nathan Lambert, Craig Schindler, Daniel Drew, and Kristofer Pister

Contact: nol@berkeley,edu

code: Github, paper: Arxiv, video: YouTube


Abstract:

Nonholonomic control is a candidate to control nonlinear systems with path-dependant states. We investigate an underactuated flying micro-aerial-vehicle, the ionocraft, that requires nonholonomic control in the yaw-direction for complete attitude control. Deploying an analytical control law involves substantial engineering design and is sensitive to inaccuracy in the system model. With specific assumptions on assembly and system dynamics, we derive a Lie bracket for yaw control of the ionocraft. As a comparison to the significant engineering effort required for an analytic control law, we implement a datadriven model-based reinforcement learning yaw controller in a simulated flight task. We demonstrate that a simple modelbased reinforcement learning framework can match the derived Lie bracket control – in yaw rate and chosen actions – in a few minutes of flight data, without a pre-defined dynamics function. This paper shows that learning-based approaches are useful as a tool for synthesis of nonlinear control laws previously only addressable through expert-based design.

Key points:

  1. We derived a novel control law with the Lie-bracket (type of nonlinear controller).

  2. We showed that model-based RL with random shooting can converge to a similar control law in very few samples.

  3. We compare in detail the policies received and the strengths and weakness of nonlinear control theory at the intersection of deep RL.

Video:

Future work - controlled flight in experiment: