Lecture

GPUs: Multithreading and Architecture

Description

This lecture covers the architecture of GPUs, focusing on multithreading concepts. It explains the SIMT programming model, GPU microarchitecture, and the reasoning behind CUDA call ordering. The lecture delves into the execution models, memory systems, and the role of GPUs in machine learning. It discusses the limitations of GPUs for ML tasks and introduces tensor cores as a solution. The future of ML in GPUs is explored, highlighting the state-of-the-art training capabilities and the shift towards alternative platforms for inference.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.