Hamilton-Jacobi-Bellman Equation
Let us assume we are trying to minimize the total cost of a continuous-time system's trajectory , with dynamics in the form starting from the state .
The concept of the total cost can be generalized for any to a cost-to-go for which we may define a value function and optimal control policy the application of which results in the system following an optimal trajectory , .
For the previously defined value function the Hamilton-Jacobi-Bellman (HJB) equation can be derived1 as
1
An overview of the derivation is presented by Steven Brunton in one of his videos [Brunton2022-HJB] also available here. As a note, there is a small mistake, acknowledged by the presenter in the comments, at 9:11 of the video where "" should be replaced with "".