Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.
DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.
In this work package (WP), we investigate the possibility of discovering structure within dictionary learning. This could range from exploring groups of atoms that appear in clusters - a form of molecule learning - to learning graphical dependencies across ...
The aim of this work package (WP) is to explore approaches to learn structured sparse models, that is sparse models where the sparsity assumption seems not to be sufficient, or when there is hope to exploit some additional knowledge together with the spars ...
We develop an efficient learning framework to construct signal dictionaries for sparse representation by selecting the dictionary columns from multiple candidate bases. By sparse, we mean that only a few dictionary elements, compared to the ambient signal ...
We address the problem of microphone location calibration from a sparse coding perspective where the sensor positions are approximated over a discretized grid. We characterize the microphone signals as a sparse vector represented over a codebook of multi-c ...
Assume a multichannel data matrix, which due to the column-wise dependencies, has low-rank and joint-sparse representation. This matrix wont have many degrees of freedom. Enormous developments over the last decade in areas of compressed sensing and low-ran ...
In this work we present a new greedy algorithm for sparse approximation called LocOMP. LocOMP is meant to be run on local dictionaries made of atoms with much shorter supports than the signal length. This notably encompasses shift-invariant dictionaries an ...
Compressive sensing (CS) is a data acquisition and recovery technique for finding sparse solutions to linear inverse problems from sub-Nyquist measurements. CS features a wide range of computationally efficient and robust signal recovery methods, based on ...
We introduce the Multiplicative Update Selector and Estimator (MUSE) algorithm for sparse approximation in under-determined linear regression problems. Given ƒ = Φα* + μ, the MUSE provably and efficiently finds a k-sparse vector α̂ such that ∥Φα̂ − ƒ∥∞ ≤ ∥ ...
Sparse methods are widely used in image and audio processing for denoising and classification, but there have been few previous applications to neural signals for brain-computer interfaces (BCIs). We used the dictionary- learning algorithm K-SVD, coupled w ...
This paper presents a new method for learning overcomplete dictionaries adapted to efficient joint representation of stereo images. We first formulate a sparse stereo image model where the multi-view correlation is described by local geometric transforms o ...
Institute of Electrical and Electronics Engineers2011