A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
On the Optimization Landscape of Dynamical Output Feedback Linear Quadratic Control
[article]
2022
The optimization landscape of optimal control problems plays an important role in the convergence of many policy gradient methods. Unlike state-feedback Linear Quadratic Regulator (LQR), static output-feedback policies are typically insufficient to achieve good closed-loop control performance. We investigate the optimization landscape of linear quadratic control using dynamical output feedback policies, denoted as dynamical LQR (dLQR) in this paper. We first show that the dLQR cost varies with
doi:10.48550/arxiv.2201.09598
fatcat:pvrax7e3fvdrnpscu7mnglg5dm