A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
Globally Optimal Hierarchical Reinforcement Learning for Linearly-Solvable Markov Decision Processes
[article]
2022
arXiv
pre-print
In this work we present a novel approach to hierarchical reinforcement learning for linearly-solvable Markov decision processes. Our approach assumes that the state space is partitioned, and the subtasks consist in moving between the partitions. We represent value functions on several levels of abstraction, and use the compositionality of subtasks to estimate the optimal values of the states in each partition. The policy is implicitly defined on these optimal value estimates, rather than being
arXiv:2106.15380v3
fatcat:z7rdlxnz5bdufecgxrsakghr2m