Lecture

Adaptive Gradient Methods

Description

This lecture covers adaptive gradient methods, including the Variable Metric Gradient Descent Algorithm and AdaGrad. It explains how these methods adapt locally by setting the Hessian matrix based on past gradient information. The lecture delves into the mathematical details of AdaGrad, highlighting its adaptive step-size and coordinate-wise extension. It also discusses the convergence rates for AdaGrad and introduces AcceleGrad, a combination of adaptive and accelerated algorithms. The lecture further explores UniXGrad, an accelerated extra-gradient algorithm for handling constraints, and ExtraNewton, an adaptive Newton's method with fast convergence rates.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.