Abstract
In modern machine learning, models can often fit training data in numerous ways, some of which perform well on unseen (test) data, while others do not. Remarkably, in such cases gradient descent frequently exhibits an implicit bias that leads to excellent performance on unseen data. This implicit bias was extensively studied in supervised learning, but is far less understood in optimal control (reinforcement learning). There, learning a controller applied to a system via gradient descent is known as policy gradient, and a question of prime importance is the extent to which a learned controller extrapolates to unseen initial states. This paper theoretically studies the implicit bias of policy gradient in terms of extrapolation to unseen initial states. Focusing on the fundamental Linear Quadratic Regulator (LQR) problem, we establish that the extent of extrapolation depends on the degree of exploration induced by the system when commencing from initial states included in training. Experiments corroborate our theory, and demonstrate its conclusions on problems beyond LQR, where systems are non-linear and controllers are neural networks. We hypothesize that real-world optimal control may be greatly improved by developing methods for informed selection of initial states to train on.
Original language | English |
---|---|
Pages (from-to) | 42275-42331 |
Number of pages | 57 |
Journal | Proceedings of Machine Learning Research |
Volume | 235 |
State | Published - 2024 |
Event | 41st International Conference on Machine Learning, ICML 2024 - Vienna, Austria Duration: 21 Jul 2024 → 27 Jul 2024 |
Funding
Funders | Funder number |
---|---|
Blavatnik Family Foundation | |
Yandex Initiative in Machine Learning | |
Adelis Research Fund for Artificial Intelligence | |
Tel Aviv University | |
European Research Council | |
Amnon and Anat Shashua | |
Google Research Gift | |
Israel Science Foundation | 1780/21 |
European Unions Horizon 2020 research and innovation pro-gramme | ERC HOLI 819080 |