Continuous-Time Linear-Quadratic Regulator
Infinite horizon
For a linear time-invariant continuous-time system and a quadratic total cost where and , of its trajectory , the optimal controller can be derived based on the assumption that the value function takes the form When substituted into the Hamilton-Jacobi-Bellman Equation along with the system's dynamics we attain where and . To find the minimum, we may take the gradient of its argument (which is by design quadratic and convex) with respect to , set it to zero and find the solution (optimal control input)
The input can then be substituted back into (1). As the equation must hold for all , through basic manipulations we then attain the continuous-time algebraic Riccati equation (CARE)
Finite horizon
For a linear time-invariant continuous-time system and a quadratic total cost where , , and , of its trajectory , the optimal controller can be derived based on the assumption that the value function takes the form When substituted into the Hamilton-Jacobi-Bellman Equation along with the system's dynamics we attain where and . To find the minimum, we may take the gradient of its argument (which is by design quadratic and convex) with respect to , set it to zero and find the solution (optimal control input)
The input can then be substituted back into (1). As the equation must hold for all , through basic manipulations we then attain the continuous-time differential Riccati equation (CDRE) for a finite horizon . Supplemented with the boundary conditon , the CDRE forms a initial value problem (IVP) which can be solved using numerical integration. Numerical errors in the integration process may lead to the loss of positive-semi-definiteness. To overcome this, instead of integrating directly, we may use its factorized form a.k.a. the "square-root form" where As must be invertible must be (at least numerically) positive definite. Consequenlty we may use the cholesky factorization in order to form the boundary condition