This lecture covers generative models focusing on self-attention and transformers. Topics include autoencoders, Boltzmann machines, masked training, attention mechanisms, and maximum entropy principle. The slides discuss sampling methods, empirical means, and correlations in detail.