Laplacian Support Vector Analysis for Subspace Discriminative Learning
Graph Chatbot
Chat with Graph Search
Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.
DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.
Principal Component Analysis (PCA) has been widely used for manifold description and dimensionality reduction. Performance of PCA is however hampered when data exhibits nonlinear feature relations. In this work, we propose a new framework for manifold lear ...
In this paper we apply boosting to learn complex non-linear local visual feature representations, drawing inspiration from its successful application to visual object detection. The main goal of local feature descriptors is to distinctively repre- sent a s ...
Extracting low dimensional structure from high dimensional data arises in many applications such as machine learning, statistical pattern recognition, wireless sensor networks, and data compression. If the data is restricted to a lower dimensional subspace ...
Institute of Electrical and Electronics Engineers2012
Locality-Sensitive Hashing (LSH) approximates nearest neighbors in high dimensions by projecting original data into low-dimensional subspaces. The basic idea is to hash data samples to ensure that the probability of collision is much higher for samples tha ...
We consider the problem of classification of a pattern from multiple compressed observations that are collected in a sensor network. In particular, we exploit the properties of random projections in generic sensor devices and we take some first steps in in ...
Latent-variable calibrations using principal component regression and partial least-squares regression are often compromised by drift such as systematic disturbances and offsets. This paper presents a two-step framework that facilitates the evaluation and ...
This paper presents an application of the kernel principal component analysis aiming at aligning optical images before the application of change detection techniques. The approach relies on the extraction of nonlinear features from a selected subset of pix ...
Conventional linear subspace learning methods like principal component analysis (PCA), linear discriminant analysis (LDA) derive subspaces from the whole data set. These approaches have limitations in the sense that they are linear while the data distribut ...
Institute of Electrical and Electronics Engineers2011
In this paper we consider recovery of a high dimensional data matrix from a set of incomplete and noisy linear measurements. We introduce a new model which can efficiently restricts the degrees of freedom of data and, at the same time, is generic so that f ...
In this paper, we present a novel semi-supervised dimensionality reduction technique to address the problems of inefficient learning and costly computation in coping with high-dimensional data. Our method named the dual subspace projections (DSP) embeds hi ...