Covers the foundational concepts of deep learning and the Transformer architecture, focusing on neural networks, attention mechanisms, and their applications in sequence modeling tasks.
Explores the learning dynamics of deep neural networks using linear networks for analysis, covering two-layer and multi-layer networks, self-supervised learning, and benefits of decoupled initialization.
Covers the fundamentals of multilayer neural networks and deep learning, including back-propagation and network architectures like LeNet, AlexNet, and VGG-16.
Explores model-based deep reinforcement learning, focusing on Monte Carlo Tree Search and its applications in game strategies and decision-making processes.