Optimization in Machine Learning: Gradient Descent
Graph Chatbot
Chat with Graph Search
Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.
DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.
Covers the concept of gradient descent in scalar cases, focusing on finding the minimum of a function by iteratively moving in the direction of the negative gradient.
Covers the history and inspiration behind artificial neural networks, the structure of neurons, learning through synaptic connections, and the mathematical description of artificial neurons.
Covers the BackProp algorithm, including initialization, signal propagation, error computation, weight updating, and complexity comparison with numerical differentiation.