Lecture

Continuous Reinforcement Learning: Advanced Machine Learning

Description

This lecture covers the challenges of continuous-state reinforcement learning, such as the curse of dimensionality and the need for function approximation to estimate the value function. It explains how to learn the value function using Monte-Carlo and Temporal Difference methods, and how to update the value function through roll-outs. The lecture also delves into function approximation for the value function, providing examples of parametrizing the value function and choosing features. It discusses the transition from the value function to the policy, and introduces policy gradients as an alternative approach. Additionally, it explores Policy Gradients and Policy learning by Weighted Exploration with the Returns (POWER) for reinforcement learning, including human demonstration for imitation learning. The lecture concludes with examples of reinforcement learning policies after multiple trials.

This video is available exclusively on Mediaspace for a restricted audience. Please log in to MediaSpace to access it if you have the necessary permissions.

Watch on Mediaspace
About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.