Covers subquadratic attention mechanisms and state space models, focusing on their theoretical foundations and practical implementations in machine learning.
Covers transformer architecture and subquadratic attention mechanisms, focusing on efficient approximations and their applications in machine learning.
Introduces Dynamic Programming, focusing on saving computation by remembering previous calculations and applying it to solve optimization problems efficiently.