Nonlinear Control: Hamilton Jacobi Bellman (HJB) and Dynamic Programming

70,222
0
Publicado 2022-01-28
This video discusses optimal nonlinear control using the Hamilton Jacobi Bellman (HJB) equation, and how to solve this using dynamic programming.

Citable link for this video: doi.org/10.52843/cassyni.4t5069

This is a lecture in a series on reinforcement learning, following the new Chapter 11 from the 2nd edition of our book "Data-Driven Science and Engineering: Machine Learning, Dynamical Systems, and Control" by Brunton and Kutz

Book Website: databookuw.com/
Book PDF: databookuw.com/databook.pdf

Amazon: www.amazon.com/Data-Driven-Science-Engineering-Lea…

Brunton Website: eigensteve.com

This video was produced at the University of Washington

Todos los comentarios (21)
  • @charlescai6672
    Very good explanation to derivative of HJB equation. But there's a point I may have to add that I think there may be a typo in 'DERIVING HJB EQUATION': In dV/dt, minimizing the integral of L(x,u), the lower limits of integral should be t instead of 0. Only by the case, we can conclude in the second last equation that -L(x(t), u(t)) can be obtained from the time derivative of integral of function L(x,u)...
  • @higasa24351
    This is the first time I've ever seen the explanation of HJB-DP in a intuitive and fashionable way, not by following the text book lines one by one. Thank you so much for the great talk.
  • @ecologypig
    Excellent. Can see a lot of connections with Control and how the essence of Bellman equation are all over the place in different fields. Thanks Prof. Brunton!
  • @alanzhus2730
    Can't believe serious topic as this can have thousands of views hours after release. Youtube is really a magic place.
  • @ailsani8749
    I am a follower from his 'control bootcamp' series. Just trying to tell everyone new here that his video is life-saving.
  • @amaarquadri
    Wow it's so cool that these concepts from reinforcement learning apply so perfectly to nonlinear control.
  • @dmitry.bright
    thanks Steve for a great lecture; looking forward to more lectures on RL and non-linear control if possible with some simple examples. thank you very much!
  • @mingyucai6559
    Clear tutorial. Thanks Prof. Steve. Keep following your steps.
  • @hfkssadfrew
    Hey Steve, on 9:11 it should be integration from t to t_f, then that’s where the - comes from.
  • @geonheelee4717
    A Great Lecture. I hope the next lecture will open asap. In particular, I'm interest in detailed relationship between RL and optimal control.
  • @tuptge
    More on non linear control please! Im trying to make up my mind on topics for my postgrad thesis!
  • Steve I follow all of your lectures. Being a mechanical engineer I really got amazed by watching your turbulence lectures. I personally worked with CFD using scientific python and visualization and computation using python and published a couple of research articles. I'm very eager to work under your guidance in the field of CFD and Fluid dynamics using Machine learning specifically simulation and modelling of turbulence fluid flow field and explore the mysterious world of turbulence. How should I reach you for further communication?
  • @blitzkringe
    Please do more of this content. Thank you.
  • @prantel1
    At 11:47 the bounds of the integral should be from “t” to “tf”; not from 0 to tf. If you make that change then the derivative of the integral wrt to t will be -L(.,.)
  • @boldirio
    Great as always Steve! I was wondering if you have any experience in transfer learning, specifically domain adaptation? If so it would be a cool topic to go through! /J
  • Thanks dear steve for this wonderful tutorial I was wondering would it be ok if you solving an example for that?
  • @leventguvenc917
    Very nice video. In deriving the HJB equation, the lower limit of the integral should be t instead of 0.
  • This Hilbert space is include in f(x(k),u(k) * (x(0),y(k)-0) or outside the x(k) - (without double equation)?