Risk Averse Robust Adversarial Reinforcement Learning
Xinlei Pan, Daniel Seita, Yang Gao, John Canny
Xinlei Pan, Daniel Seita, Yang Gao, John Canny
Risk Averse Robust Adversarial Reinforcement Learning
Xinlei Pan, Daniel Seita, Yang Gao, John Canny
ICRA 2019
Updates:
Training Environment:
Training Results:
Testing all models without attacks or perturbations. The reward is divided into distance related reward (left subplot), progress related reward (middle subplot). We also present results for catastrophe reward per episode (right subplot). The blue vertical line indicates the beginning of adding perturbations during training.
Testing all models with random attacks. The three subplots follow the same convention as in the first figure.
Testing all models with adversarial attack. The three subplots follow the same convention as in the first figure.