Model Compression: Techniques for Efficient NLP Models
Graph Chatbot
Chat with Graph Search
Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.
DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.
Covers the fundamentals of multilayer neural networks and deep learning, including back-propagation and network architectures like LeNet, AlexNet, and VGG-16.
Explores MP3 encoding, emphasizing reducing bits through lossy compression and utilizing psycho-acoustic models for efficient filtering and quantization.
Explores optimization strategies for deep learning accelerators, emphasizing data movement reduction through batching, dataflow optimizations, and compression.
Explores image compression through various approaches like pixel and block level compression, Discrete Cosine Transform, quantization, and entropy coding.
Explores decoding from neural models in modern NLP, covering encoder-decoder models, decoding algorithms, issues with argmax decoding, and the impact of beam size.