Q-Learning: Model Free Reinforcement Learning and Temporal Difference Learning

96,293
0
Publicado 2022-01-14
Here we describe Q-learning, which is one of the most popular methods in reinforcement learning. Q-learning is a type of temporal difference learning. We discuss other TD algorithms, such as SARSA, and connections to biological learning through dopamine. Q-learning is also one of the most common frameworks for deep reinforcement learning.

Citable link for this video: doi.org/10.52843/cassyni.ss11hp

This is a lecture in a series on reinforcement learning, following the new Chapter 11 from the 2nd edition of our book "Data-Driven Science and Engineering: Machine Learning, Dynamical Systems, and Control" by Brunton and Kutz

Book Website: databookuw.com/
Book PDF: databookuw.com/databook.pdf

Amazon: www.amazon.com/Data-Driven-Science-Engineering-Lea…

Brunton Website: eigensteve.com

This video was produced at the University of Washington

Todos los comentarios (21)
  • I personally love the big picture perspective that Prof. Brunton always shows. Please, continue to make these high quality videos!
  • @usonian11
    Thank you for the outstanding production quality and content of these lectures! I especially enjoy the structure diagram organizing the different RL methods.
  • @OmerBoehm
    Thank you dear Prof Brunton for this outstanding lecture. The detailed explanations and focus on subtleties are so important , Looking forward to your next videos.
  • Professor I must sincerely thank you for the astonishing quality of this video. You were able to clearly explain an advanced concept without simplifying, going into the details and providing brilliant insights. Also I sincerely thank you for saving my GPA from my R.L. exam 😆
  • I enjoy your talks. They are very clear and well structured and have the right level of detail. Thank you,
  • @jashwantraj2987
    Prof. Burton, you are amazing. I never expected someone to take so much of time to explain a concept about TD. I'm one of the few people who hate reading text books to understand concepts. I rather see a video or learn about it from class. Thanks a lot
  • @areebayubi5469
    Thank you so much for using very relevant analogies and very clear explanations. I think I have a much better grasp of the concepts behind Temporal Difference learning now.
  • @kalimantros845
    I was hoping that your next video would have been about Q-learning, and here it comes!
  • @haotianhang3997
    Thank you! It's a great video. My understanding in TD learning was deepened a lot.
  • @TheSinashah
    CS PhD student here. This video provides such amazing content. Highly recommended.
  • @BoltzmannVoid
    this was the best explanation ever! thank you so much, professor!
  • @cruise0101
    Excellent class! Extremely easy to understand!
  • @marzs.szzzzz
    These are fantastic lectures, I use these as an alternative explaination to David Silvers DeepmindxUCL 2015 lectures on the same topic, the different perspective really suits how my brain understands RL. Thank you!!
  • Thank you for the clear picture. It was really well explained and others already mentioned, now I can say that I understand these techniques quite fairly well. 🙏
  • @FRANKONATOR123
    Hi Prof. Brunton. Great vídeo as always! Please keep producing quality ML content
  • @krullebolalex
    Thanks a bundle Steve, this was really well explained!
  • @complexobjects
    I do like the description of Q Learning. I had come up with another analogy for why it makes sense. If you took the action of going out to a party, and then happened to make some mistakes while there, we wouldn't want to say "you should never go out again." We'd want to reinforce the action of going out based on the best* possible outcome of that night, not the suboptimal action that was taken once there.
  • @TheFitsome
    This is the best RL tutorial on the internet.
  • @denchen1950
    The video quality is incredible lol and all the concept is discussed extremely clear OMG!! Brilliant masterpiece bro KEEP GOING !!