Are you an EPFL student looking for a semester project?
Work with us on data science and visualisation projects, and deploy your project as an app on top of Graph Search.
This lecture covers document classification methods, starting with the task of constructing a classifier to assign labels to unlabeled documents. It explores features like bag of words, phrases, and grammatical features, as well as challenges in dealing with high-dimensional feature spaces. Various classification algorithms such as k-Nearest-Neighbors and Naïve Bayes Classifier are discussed, along with their probabilistic estimates. The lecture also delves into transformer models, self-attention mechanisms, and the use of multi-head attention for learning different relationships. The importance of position embeddings, layer normalization, and multi-layer perceptrons in transformer networks is highlighted, along with the process of combining multi-head attention outputs. The lecture concludes with insights on finetuning transformer networks and the significance of transformer models in document classification.