Publication

Nonparametric Construction of Multivariate Kernels

Victor Panaretos, Kjell Konis
2012
Journal paper
Abstract

We propose a nonparametric method for constructing multivariate kernels tuned to the configuration of the sample, for density estimation in R-d, d moderate. The motivation behind the approach is to break down the construction of the kernel into two parts: determining its overall shape and then its global concentration. We consider a framework that is essentially nonparametric, as opposed to the usual bandwidth matrix parameterization. The shape of the kernel to be employed is determined by applying the backprojection operator, the dual of the Radon transform, to a collection of one-dimensional kernels, each optimally tuned to the concentration of the corresponding one-dimensional projections of the data. Once an overall shape is determined, the global concentration is controlled by a simple sealing. It is seen that the kernel estimators thus developed are easy and extremely fast to compute, and perform at least as well in practice as parametric kernels with cross-validated or otherwise tuned covariance structure. Connections with integral geometry are discussed, and the approach is illustrated under a wide range of scenarios in two and three dimensions, via an R package developed for its implementation.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Related concepts (32)
Kernel density estimation
In statistics, kernel density estimation (KDE) is the application of kernel smoothing for probability density estimation, i.e., a non-parametric method to estimate the probability density function of a random variable based on kernels as weights. KDE answers a fundamental data smoothing problem where inferences about the population are made, based on a finite data sample. In some fields such as signal processing and econometrics it is also termed the Parzen–Rosenblatt window method, after Emanuel Parzen and Murray Rosenblatt, who are usually credited with independently creating it in its current form.
Multivariate kernel density estimation
Kernel density estimation is a nonparametric technique for density estimation i.e., estimation of probability density functions, which is one of the fundamental questions in statistics. It can be viewed as a generalisation of histogram density estimation with improved statistical properties. Apart from histograms, other types of density estimators include parametric, spline, wavelet and Fourier series. Kernel density estimators were first introduced in the scientific literature for univariate data in the 1950s and 1960s and subsequently have been widely adopted.
Nonparametric statistics
Nonparametric statistics is the type of statistics that is not restricted by assumptions concerning the nature of the population from which a sample is drawn. This is opposed to parametric statistics, for which a problem is restricted a priori by assumptions concerning the specific distribution of the population (such as the normal distribution) and parameters (such the mean or variance).
Show more
Related publications (34)

Density Estimation In Rkhs With Application To Korobov Spaces In High Dimensions

Fabio Nobile, Yoshihito Kazashi

A kernel method for estimating a probability density function from an independent and identically distributed sample drawn from such density is presented. Our estimator is a linear combination of kernel functions, the coefficients of which are determined b ...
SIAM PUBLICATIONS2023

Detecting whether a stochastic process is finitely expressed in a basis

Victor Panaretos, Neda Mohammadi Jouzdani

Is it possible to detect if the sample paths of a stochastic process almost surely admit a finite expansion with respect to some/any basis? The determination is to be made on the basis of a finite collection of discretely/noisily observed sample paths. We ...
ACADEMIC PRESS INC ELSEVIER SCIENCE2023

Robust discrete choice models with t-distributed kernel errors

Michel Bierlaire, Thomas Gasos, Prateek Bansal

Outliers in discrete choice response data may result from misclassification and misreporting of the response variable and from choice behaviour that is inconsistent with modelling assumptions (e.g. random utility maximisation). In the presence of outliers, ...
2023
Show more

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.