Font Size: a A A

Learning From Data: Hermite Learning And Normal Estimation On Riemannian Manifolds

Posted on:2011-12-11Degree:DoctorType:Dissertation
Country:ChinaCandidate:L ShiFull Text:PDF
GTID:1100360305466582Subject:Computational Mathematics
Abstract/Summary:PDF Full Text Request
In this thesis, we investigate some algorithms in learning theory for purpose of regression, manifold learning and data analysis. Their design and asymptotic perfor-mance will be discussed in detail from the view point of approximation theory.In the first part, the problem of learning from data involving function values and gradients is studied in a framework of least-square regularized regression in reproduc-ing kernel Hilbert spaces. The algorithm is implemented by a linear system with the coefficient matrix involving both block matrices for generating Graph Laplacians and Hessians. The additional data for function gradients improve learning performance of the algorithm. Error analysis is done by means of sampling operators for the sample error and integral operators in Sobolev spaces for the approximation error.Normal estimation is an important topic for processing point cloud data and sur-face reconstruction in computer graphics. In the second part of the thesis, we consider the problem of estimating normals for a (unknown) submanifold of a Euclidean space of codimension 1 from random points on the manifold. We propose a kernel based learning algorithm in an unsupervised form of gradient learning. The algorithm can be implemented by solving a linear algebra problem. Error analysis is conducted under conditions on the true normals of the manifold and the sampling distribution.In the last part of this thesis, we consider the regression problem by learning with a regularization scheme in a data dependent hypothesis space. For a given set of sam-ples, functions in this hypothesis space are defined to be linear combinations of basis functions generated by a kernel function and sample data, thus are entirely determined by the combination coefficients. The data dependence nature of the kernel-based hy-pothesis space provides flexibility and adaptivity for the learning algorithms. The regu-larization scheme is essentially different from the standard one in a reproducing kernel Hilbert space:the kernel function is not necessarily symmetric or positive semi-definite and the regularizer, as a functional acting on the functions in such kinds of hypothesis spaces, is taken to be the p-th power of the (?)p-norm of the corresponding combination coefficients. The differences lead to additional difficulty in the error analysis.To be more specific, we mainly study two cases in this thesis:p = 1 and p= 2. When p = 1, the(?)1-regularizer often leads to sparsity of the solution vector which can greatly improve the efficiency of computations. When p = 2, the corresponding algorithm is linear and is easy to implement by solving a linear system. Both algorithms have been studied in the literature. In this thesis, we apply concentration techniques with (?)2-empirical covering numbers to get the best learning rate for the the learning algorithms. Since our aim is a capacity dependent analysis, we also show that the function spaces involved in the error analysis induced by the non-symmetric kernel function have nice behaviors in terms of the (?)2-empirical covering numbers of its unit ball.
Keywords/Search Tags:Hermite learning, normal estimation, manifold learning, sample dependent hypothesis space, coefficients regularization scheme
PDF Full Text Request
Related items