Posez n’importe quelle question sur les cours, conférences, exercices, recherches, actualités, etc. de l’EPFL ou essayez les exemples de questions ci-dessous.
AVERTISSEMENT : Le chatbot Graph n'est pas programmé pour fournir des réponses explicites ou catégoriques à vos questions. Il transforme plutôt vos questions en demandes API qui sont distribuées aux différents services informatiques officiellement administrés par l'EPFL. Son but est uniquement de collecter et de recommander des références pertinentes à des contenus que vous pouvez explorer pour vous aider à répondre à vos questions.
In this work package (WP), we investigate the possibility of discovering structure within dictionary learning. This could range from exploring groups of atoms that appear in clusters - a form of molecule learning - to learning graphical dependencies across ...
Assume a multichannel data matrix, which due to the column-wise dependencies, has low-rank and joint-sparse representation. This matrix wont have many degrees of freedom. Enormous developments over the last decade in areas of compressed sensing and low-ran ...
This paper presents a new method for learning overcomplete dictionaries adapted to efficient joint representation of stereo images. We first formulate a sparse stereo image model where the multi-view correlation is described by local geometric transforms o ...
Institute of Electrical and Electronics Engineers2011
We introduce the Multiplicative Update Selector and Estimator (MUSE) algorithm for sparse approximation in under-determined linear regression problems. Given ƒ = Φα* + μ, the MUSE provably and efficiently finds a k-sparse vector α̂ such that ∥Φα̂ − ƒ∥∞ ≤ ∥ ...
The aim of this work package (WP) is to explore approaches to learn structured sparse models, that is sparse models where the sparsity assumption seems not to be sufficient, or when there is hope to exploit some additional knowledge together with the spars ...
We develop an efficient learning framework to construct signal dictionaries for sparse representation by selecting the dictionary columns from multiple candidate bases. By sparse, we mean that only a few dictionary elements, compared to the ambient signal ...
We address the problem of microphone location calibration from a sparse coding perspective where the sensor positions are approximated over a discretized grid. We characterize the microphone signals as a sparse vector represented over a codebook of multi-c ...
Sparse methods are widely used in image and audio processing for denoising and classification, but there have been few previous applications to neural signals for brain-computer interfaces (BCIs). We used the dictionary- learning algorithm K-SVD, coupled w ...
Compressive sensing (CS) is a data acquisition and recovery technique for finding sparse solutions to linear inverse problems from sub-Nyquist measurements. CS features a wide range of computationally efficient and robust signal recovery methods, based on ...
In this work we present a new greedy algorithm for sparse approximation called LocOMP. LocOMP is meant to be run on local dictionaries made of atoms with much shorter supports than the signal length. This notably encompasses shift-invariant dictionaries an ...