Lecture

Linear Programming Techniques in Reinforcement Learning

Description

This lecture introduces the linear programming (LP) approach to reinforcement learning (RL), presenting it as an alternative convex viewpoint. It begins by revisiting the reinforcement learning setup, emphasizing the challenges faced in traditional methods, such as the need for approximate dynamic programming and the limitations of existing algorithms. The instructor discusses the Bellman optimality equation and its significance in defining optimal policies. The lecture then transitions to the primal and dual formulations of linear programming, detailing how these can be applied to solve Markov decision processes (MDPs). The occupancy measure is defined and visualized, illustrating its role in determining the value function. The lecture also covers the REPS algorithm, which applies proximal point methods to the dual LP, showcasing its effectiveness in practical applications like robotics. The session concludes with a summary of the LP approach's advantages and challenges, setting the stage for future discussions on policy gradient methods.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.