Publication Date
4-1-2023
Document Type
Article
Publication Title
Physical Review Research
Volume
5
Issue
2
DOI
10.1103/PhysRevResearch.5.023085
Abstract
Reinforcement learning (RL) is an important field of research in machine learning that is increasingly being applied to complex optimization problems in physics. In parallel, concepts from physics have contributed to important advances in RL with developments such as entropy-regularized RL. While these developments have led to advances in both fields, obtaining analytical solutions for optimization in entropy-regularized RL is currently an open problem. In this paper, we establish a mapping between entropy-regularized RL and research in nonequilibrium statistical mechanics focusing on Markovian processes conditioned on rare events. In the long-time limit, we apply approaches from large deviation theory to derive exact analytical results for the optimal policy and optimal dynamics in Markov decision process (MDP) models of reinforcement learning. The results obtained lead to an analytical and computational framework for entropy-regularized RL which is validated by simulations. The mapping established in this work connects current research in reinforcement learning and nonequilibrium statistical mechanics, thereby opening avenues for the application of analytical and computational approaches from one field to cutting-edge problems in the other.
Funding Number
DMS-1854350
Funding Sponsor
Directorate for Mathematical and Physical Sciences
Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.
Department
Computer Engineering
Recommended Citation
Argenis Arriojas, Jacob Adamczyk, Stas Tiomkin, and Rahul V. Kulkarni. "Entropy regularized reinforcement learning using large deviation theory" Physical Review Research (2023). https://doi.org/10.1103/PhysRevResearch.5.023085