Keyboard shortcuts

Press or to navigate between chapters

Press S or / to search in the book

Press ? to show this help

Press Esc to hide this help

Hamilton-Jacobi-Bellman Equation

Let us assume we are trying to minimize the total cost of a continuous-time system’s trajectory , with dynamics in the form starting from the state .

The concept of the total cost can be generalized for any to a cost-to-go for which we may define a value function and optimal control policy the application of which results in the system following an optimal trajectory , .

For the previously defined value function the Hamilton-Jacobi-Bellman (HJB) equation can be derived1 as



  1. An overview of the derivation is presented by Steven Brunton in one of his videos Brunton2022-HJB also available here. As a note, there is a small mistake, acknowledged by the presenter in the comments, at 9:11 of the video where “” should be replaced with “”.