Skip to main content
Graph
Search
fr
en
Login
Search
All
Categories
Concepts
Courses
Lectures
MOOCs
People
Practice
Publications
Startups
Units
Show all results for
Home
Lecture
Information Theory: Source Coding
Graph Chatbot
Related lectures (29)
Previous
Page 2 of 3
Next
Information Theory: Entropy and Information Processing
Explores entropy in information theory and its role in data processing and probability distributions.
Conditional Entropy and Data Compression Techniques
Discusses conditional entropy and its role in data compression techniques.
Universal Source Coding
Covers the Lempel-Ziv universal coding algorithm and invertible finite state machines in information theory.
Entropy and Data Compression: Huffman Coding Techniques
Discusses entropy, data compression, and Huffman coding techniques, emphasizing their applications in optimizing codeword lengths and understanding conditional entropy.
Error Correction Codes: Theory and Applications
Covers error correction codes theory and applications, emphasizing the importance of minimizing distance for reliable communication.
Information Measures: Entropy and Information Theory
Explains how entropy measures uncertainty in a system based on possible outcomes.
Source Coding Theorem
Explores the Source Coding Theorem, entropy, Huffman coding, and conditioning's impact on entropy reduction.
Source Coding Theorems: Entropy and Source Models
Covers source coding theorems, entropy, and various source models in information theory.
Information Theory: Source Coding, Cryptography, Channel Coding
Covers source coding, cryptography, and channel coding in communication systems, exploring entropy, codes, error channels, and future related courses.
Information Theory: Entropy and Capacity
Covers concepts of entropy, Gaussian distributions, and channel capacity with constraints.