Reinforcement learning (RL) is used to directly design a control policy using data collected from the system. This paper considers the robustness of controllers trained via model-free RL. The discussion focuses on posing the (model-free) linear quadratic Gaussian (LQG) problem as a special instance of RL. A simple LQG example is used to demonstrate that RL with partial observations can lead to poor robustness margins. It is proposed to recover robustness by introducing random perturbations at the system input during the RL training. The perturbation magnitude can be used to trade off performance for increased robustness. Two simple examples are presented to demonstrate the proposed method for enhancing robustness during RL training.
|Original language||English (US)|
|Title of host publication||2019 American Control Conference, ACC 2019|
|Publisher||Institute of Electrical and Electronics Engineers Inc.|
|Number of pages||7|
|State||Published - Jul 2019|
|Event||2019 American Control Conference, ACC 2019 - Philadelphia, United States|
Duration: Jul 10 2019 → Jul 12 2019
|Name||Proceedings of the American Control Conference|
|Conference||2019 American Control Conference, ACC 2019|
|Period||7/10/19 → 7/12/19|
Bibliographical notePublisher Copyright:
© 2019 American Automatic Control Council.