Discrete-Time Linear-Quadratic Regulator
Infinite horizon
For a linear time-invariant discrete-time system and a quadratic total cost of its trajectory the optimal controller can be derived based on the assumption that the value function takes the form
When substituted into the Bellman Equation along with the system's dynamics we attain To find the minimum, we may take the gradient of its argument (which is by design quadratic and convex) with respect to , set it to zero and find the solution (optimal control input)
The input can then be substitued back into (1). As the equation must hold for all , through basic manipulations we then attain the discrete-time algebraic Riccati equation (DARE)
Finite horizon
For a linear time-invariant discrete-time system and a quadratic total cost of its trajectory the optimal controller can be derived based on the assumption that the value function takes the form When substituted into the Bellman Equation along with the system's dynamics we attain To find the minimum, we may take the gradient of its argument (which is by design quadratic and convex) with respect to , set it to zero and find the solution (optimal control input)
The input can then be substitued back into (1). As the equation must hold for all , through basic manipulations we then attain the discrete-time dynamic Riccati equation (DDRE) for a finite horizon .