TY - JOUR

T1 - Approximation of functions over manifolds

T2 - A Moving Least-Squares approach

AU - Sober, Barak

AU - Aizenbud, Yariv

AU - Levin, David

N1 - Publisher Copyright:
© 2020 Elsevier B.V.

PY - 2021/2

Y1 - 2021/2

N2 - We present an algorithm for approximating a function defined over a d-dimensional manifold utilizing only noisy function values at locations sampled from the manifold with noise. To produce the approximation we do not require knowledge about the local geometry of the manifold or its local parameterizations. We do require, however, knowledge regarding the manifold's intrinsic dimension d. We use the Manifold Moving Least-Squares approach of Sober and Levin (2019) to reconstruct the atlas of charts and the approximation is built on top of those charts. The resulting approximant is shown to be a function defined over a neighborhood of a manifold, approximating the originally sampled manifold. In other words, given a new point, located near the manifold, the approximation can be evaluated directly on that point. We prove that our construction yields a smooth function, and in case of noiseless samples the approximation order is O(hm+1), where h is a local density of sample parameter (i.e., the fill distance) and m is the degree of a local polynomial approximation, used in our algorithm. In addition, the proposed algorithm has linear time complexity with respect to the ambient space's dimension. Thus, we are able to avoid the computational complexity, commonly encountered in high dimensional approximations, without having to perform non-linear dimension reduction, which inevitably introduces distortions to the geometry of the data. Additionally, we show numerically that our approach compares favorably to some well-known approaches for regression over manifolds.

AB - We present an algorithm for approximating a function defined over a d-dimensional manifold utilizing only noisy function values at locations sampled from the manifold with noise. To produce the approximation we do not require knowledge about the local geometry of the manifold or its local parameterizations. We do require, however, knowledge regarding the manifold's intrinsic dimension d. We use the Manifold Moving Least-Squares approach of Sober and Levin (2019) to reconstruct the atlas of charts and the approximation is built on top of those charts. The resulting approximant is shown to be a function defined over a neighborhood of a manifold, approximating the originally sampled manifold. In other words, given a new point, located near the manifold, the approximation can be evaluated directly on that point. We prove that our construction yields a smooth function, and in case of noiseless samples the approximation order is O(hm+1), where h is a local density of sample parameter (i.e., the fill distance) and m is the degree of a local polynomial approximation, used in our algorithm. In addition, the proposed algorithm has linear time complexity with respect to the ambient space's dimension. Thus, we are able to avoid the computational complexity, commonly encountered in high dimensional approximations, without having to perform non-linear dimension reduction, which inevitably introduces distortions to the geometry of the data. Additionally, we show numerically that our approach compares favorably to some well-known approaches for regression over manifolds.

KW - Dimension reduction

KW - High dimensional approximation

KW - Manifold learning

KW - Moving Least-Squares

KW - Out-of-sample extension

KW - Regression over manifolds

UR - http://www.scopus.com/inward/record.url?scp=85089812572&partnerID=8YFLogxK

U2 - 10.1016/j.cam.2020.113140

DO - 10.1016/j.cam.2020.113140

M3 - מאמר

AN - SCOPUS:85089812572

VL - 383

JO - Journal of Computational and Applied Mathematics

JF - Journal of Computational and Applied Mathematics

SN - 0377-0427

M1 - 113140

ER -