Skip to main content
Graph
Search
fr
en
Login
Search
All
Categories
Concepts
Courses
Lectures
MOOCs
People
Practice
Publications
Startups
Units
Show all results for
Home
Lecture
Concurrency Basics: Threads and Synchronization
Graph Chatbot
Related lectures (32)
Previous
Page 3 of 4
Next
GPU Memory Hierarchy: Optimization
Explores GPU memory hierarchy, CUDA processing flow, optimizations, and parallelism efficiency on GPUs.
Lock-Free Synchronization: Performance and Scalability
Explores lock-free synchronization for performance and scalability in distributed systems, covering unique identifier generation, messaging queues, and atomic RDMA reads.
GPU Advanced: Memory Management and Optimization
Explores advanced GPU memory management, CUDA programming, and data-parallel computing for optimizing performance.
Locking Primitives: Avoiding Race Conditions in Threads
Covers locking primitives necessary to prevent race conditions in multithreaded programming, focusing on mutual exclusion and atomic operations.
Introduction to Concurrency
Covers the basics of concurrency, its importance, and how to write concurrent programs.
Principles of Parallel Computing: OpenMP
Explores the principles of parallel computing, focusing on OpenMP as a tool for creating concurrent code from serial code.
Readers-writer Lock
Introduces the readers-writer lock for handling concurrency in scenarios with frequent reads and occasional updates.
Parallel Computing: Principles and OpenMP
Covers the principles of parallel computing and introduces OpenMP for creating concurrent code from serial code.
Principles of Parallelism
Covers the basics of parallelism, including physical examples, historical context, multicore era, and parallel collections in Scala.
One place buffer: Producer-consumer queue
Introduces the concept of a one-place buffer in concurrent programming.