jump to main area
:::
A- A A+

Seminars

Local Linear Regression on Manifolds and its Geometric Interpretation

  • 2014-01-06 (Mon.), 10:30 AM
  • Recreation Hall, 2F, Institute of Statistical Science
  • Prof. Ming-Yen Cheng
  • Department of Mathematics, National Taiwan University

Abstract

High-dimensional data analysis has been an active area, and the main focuses have been variable selection and dimension reduction. In practice, it occurs often that the variables are located on an unknown, lower-dimensional nonlinear manifold. Under this manifold assumption, one purpose of this article is regression and gradient estimation on the manifold, and another is developing a new tool for manifold learning.?? As regards the first aim, we suggest directly reducing the dimensionality to the intrinsic dimension of the manifold, and performing the popular local linear regression (LLR) on a tangent plane estimate. An immediate consequence is a dramatic reduction in the computational time when the ambient space dimension is much larger than the intrinsic dimension. We provide rigorous theoretical justification of the convergence of the proposed regression and gradient estimators by carefully analyzing the curvature, boundary, and non-uniform sampling effects.? A bandwidth selector that can handle heteroscedastic errors is proposed.? With reference to the second aim, we analyze carefully the asymptotic behaviour of our regression estimator both in the interior and near the boundary of the manifold, and make explicit its relationship with manifold learning, in particular estimating the Laplace-Beltrami operator of the manifold. In this context, we also make clear that it is important to? use a smaller bandwidth in the tangent plane estimation than in the LLR.? A simulation study and applications to the Isomap face data and a clinical computed tomography scan dataset are used to illustrate the computational speed and estimation accuracy of our methods.

Update:
scroll to top