Lecture

Contrastive losses: Word2Vec and Skip-gram

Description

This lecture introduces contrastive losses for representation learning, focusing on Word2Vec and Skip-gram models. It covers the training objectives, context windows, negative sampling, and Noise Contrastive Estimation. The lecture also explores InfoNCE/CPC and its applications in image and graph data, as well as related topics like Deep Metric Learning and Energy-based models.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.