Learning gradients: Predictive models that infer geometry and statistical dependence
Repository Usage Stats
The problems of dimension reduction and inference of statistical dependence are addressed by the modeling framework of learning gradients. The models we propose hold for Euclidean spaces as well as the manifold setting. The central quantity in this approach is an estimate of the gradient of the regression or classification function. Two quadratic forms are constructed from gradient estimates: the gradient outer product and gradient based diffusion maps. The first quantity can be used for supervised dimension reduction on manifolds as well as inference of a graphical model encoding dependencies that are predictive of a response variable. The second quantity can be used for nonlinear projections that incorporate both the geometric structure of the manifold as well as variation of the response variable on the manifold. We relate the gradient outer product to standard statistical quantities such as covariances and provide a simple and precise comparison of a variety of supervised dimensionality reduction methods. We provide rates of convergence for both inference of informative directions as well as inference of a graphical model of variable dependencies. © 2010.
More InfoShow full item record
Research Professor of Mathematics
I am interested in novel constructions inspired by classical harmonic analysis that allow to analyse the geometry of manifolds and graphs and functions on such structures. These constructions are motivated by several important applications across many fields. In many situations we are confronted with large amounts of apparently unstructured high-dimensional data. I find fascinating to study the intrinsic geometry of such data, and exploiting in order to study, explore, visualize, characterize st
This author no longer has a Scholars@Duke profile, so the information shown here reflects their Duke status at the time this item was deposited.