Are you an EPFL student looking for a semester project?
Work with us on data science and visualisation projects, and deploy your project as an app on top of Graph Search.
This lecture introduces the concept of word embeddings, which aim to model the likelihood of a word and its context occurring together in a low-dimensional space. By mapping words and contexts into this space, the vector distance can be interpreted as a measure of their likelihood of co-occurrence. The instructor explains the process of learning the model from data, including formulating an optimization problem and defining a loss function to be minimized. The lecture covers topics such as obtaining negative samples, stochastic gradient descent, and computing derivatives. Additionally, alternative approaches like CBOW and GLOVE are discussed, along with the properties of word embeddings and their practical applications in document search, thesaurus construction, and document classification.