Sample Efficient Reinforcement Learning 

Abstract: 

In many real-world reinforcement learning applications, training agent on the real environment is often prohibitively expensive if not impossible. Therefore, we have studied various ways of speeding up the convergence of learning with limited samples. In this talk, we will present recently completed works as well as an on-going effort to improve the efficiency of reinforcement learning. Specifically, we will share our methods (1) to utilize advises from multiple experts of varying quality, (2) to transfer previously gained knowledge to a novel task, and (3) to train on a virtual environment with limited samples from the physical one. Finally, we share our efforts on multi-agent reinforcement learning with a particular focus on mechanism design.


Bio: 

Chi-Guhn Lee is a Professor of Industrial Engineering and the Director of the Centre for Maintenance Optimization and Reliability Engineering (C-MORE) at the University of Toronto. His research interest includes reinforcement learning, Markov decision process, deep learning, supply chain optimization and physical asset management. Recent and on-going projects cover topics such as transfer learning, domain adaptation and Bayesian learning with applications from supply chain, equipment diagnosis. He has focused on both applications and theory, and published in machine learning conferences such as NeurIPS, ICLR, UAI as well journals such as Operations Research, IEEE Transactions on Industrial Electronics, Mechanical Systems and Signal Processing. He has also worked closely with private firms including Nestle, LG, IBM, General Motors, Magna International, Fujitsu, State Grid Corp of China to name a few.

E-mail cglee@mie.utoronto.ca

Summary