Lecture

Gradient Descent for Linear MSE

Description

This lecture covers the concept of Gradient Descent for Linear Mean Squared Error (MSE) in machine learning. It explains the computation of the gradient, the complexity of computing the gradient, and the variant with an offset term. The lecture also delves into Stochastic Gradient Descent, the use of penalty functions for constrained optimization, and the implementation issues such as adaptive step-size selection and feature normalization. Additionally, it discusses non-convex optimization, stopping criteria, optimality conditions, and the transformation of constrained problems into unconstrained ones.

This video is available exclusively on Mediaspace for a restricted audience. Please log in to MediaSpace to access it if you have the necessary permissions.

Watch on Mediaspace
About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.