Abstract
We consider the problem of optimality in manifold reconstruction. A random sample \(\mathbb {X}_n = \{X_1,\ldots ,X_n\}\subset \mathbb {R}^D\) composed of points close to a d-dimensional submanifold M, with or without outliers drawn in the ambient space, is observed. Based on the tangential Delaunay complex (Discrete Comput Geom 51(1):221–267 2014), we construct an estimator \(\widehat{M}\) that is ambient isotopic and Hausdorff-close to M with high probability. The estimator \(\widehat{M}\) is built from existing algorithms. In a model with additive noise of small amplitude, we show that this estimator is asymptotically minimax optimal for the Hausdorff distance over a class of submanifolds satisfying a reach constraint. Therefore, even with no a priori information on the tangent spaces of M, our estimator based on Tangential Delaunay Complexes is optimal. This shows that the optimal rate of convergence can be achieved through existing algorithms. A similar result is also derived in a model with outliers. A geometric interpolation result is derived, showing that the Tangential Delaunay Complex is stable with respect to noise and perturbations of the tangent spaces. In the process, a decluttering procedure and a tangent space estimator both based on local principal component analysis (PCA) are studied.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Throughout many fields of applied science, data in \(\mathbb {R}^D\) can naturally be modeled as lying on a d-dimensional submanifold M. As M may carry a lot of information about the studied phenomenon, it is then natural to consider the problem of either approximating M geometrically, recovering it topologically, or both from a point sample \(\mathbb {X}_n = \{X_1,\ldots ,X_n\}\). It is of particular interest in high codimension (\(d \ll D\)) where it can be used as a preliminary processing of the data for reducing its dimension, and then avoiding the curse of dimensionality. This problem is usually referred to as manifold reconstruction in the computational geometry community, and rather called set/support estimation or manifold learning in the statistics literature.
The computational geometry community has now been active on manifold reconstruction for many years, mainly in deterministic frameworks. In dimension 3, [18] provides a survey of the state of the art. In higher dimension, the employed methods rely on variants of the ambient Delaunay triangulation [4, 13]. The geometric and topological guarantees are derived under the assumption that the point cloud—fixed and nonrandom—densely samples M at scale \(\varepsilon \), with \(\varepsilon \) small enough or going to 0.
In the statistics literature, most of the attention has been paid to approximation guarantees, rather than topological ones. The approximation bounds are given in terms of the sample size n, that is assumed to be large enough or going to infinity. To derive these bounds, a broad variety of assumptions on M have been considered. For instance, if M is a bounded convex set and \(\mathbb {X}_n\) does not contain outliers, a natural idea is to consider the convex hull \(\widehat{M} = \mathrm{Conv}(\mathbb {X}_n)\) to be the estimator. \(\mathrm{Conv}(\mathbb {X}_n)\) provides optimal rates of approximation for several loss functions [20, 29]. These rates depend crudely on the regularity of the boundary of the convex set M. In addition, \(\mathrm{Conv}(\mathbb {X}_n)\) clearly is ambient isotopic to M so that it has both good geometric and topological properties. Generalizations of the notion of convexity based on rolling ball-type assumptions such as r-convexity and reach bounds [15, 24] yield rich classes of sets with good geometric properties. In particular, the reach, as introduced by Federer [22], appears to be a key regularity and scale parameter [11, 24, 28].
This paper mainly follows up the two articles [4, 24], both dealing with the case of a d-dimensional submanifold \(M \subset \mathbb {R}^D\) with a reach regularity condition and where the dimension d is known.
On one hand, [4] focuses on a deterministic analysis and proposes a provably faithful reconstruction. The authors introduce a weighted Delaunay triangulation restricted to tangent spaces, the so-called Tangential Delaunay Complex. This paper gives a reconstruction up to ambient isotopy with approximation bounds for the Hausdorff distance along with computational complexity bounds. This work provides a simplicial complex based on the input point cloud and tangent spaces. However, it lacks stability up to now, in the sense that the assumptions used in the proofs of [4] do not resist ambient perturbations. Indeed, it heavily relies on the knowledge of the tangent spaces at each point and on the absence of noise.
On the other hand, [24] takes a statistical approach in a model possibly corrupted by additive noise, or containing outlier points. The authors derive an estimator that is proved to be minimax optimal for the Hausdorff distance \({d}_{H}\). Roughly speaking, minimax optimality of the proposed estimator means that it performs best in the worst possible case up to numerical constants, when the sample size n is large enough. Although theoretically optimal, the proposed estimator appears to be intractable in practice. At last, [28] proposes a manifold estimator based on local linear patches that is tractable but fails to achieve the optimal rates.
1.1 Contribution
Our main contributions (Theorems 2.7, 2.8 and 2.9) make a two-way link between the approaches of [4] and [24].
From a geometric perspective, Theorem 2.7 shows that the Tangential Delaunay Complex of [4] can be combined with local PCA to provide a manifold estimator that is optimal in the sense of [24]. This remains possible even if data is corrupted with additive noise of small amplitude. Also, Theorems 2.8 and 2.9 show that, if outlier points are present (clutter noise), the Tangential Delaunay Complex of [4] still yields the optimal rates of [24], at the price of an additional decluttering procedure.
From a statistical point of view, our results show that the optimal rates described in [24] can be achieved by a tractable estimator \(\widehat{M}\) that (1) is a simplicial complex of which vertices are the data points, and (2) such that \(\widehat{M}\) is ambient isotopic to M with high probability.
In the process, a stability result for the Tangential Delaunay Complex (Theorem 4.4) is proved. Let us point out that this stability is derived using an interpolation result (Theorem 4.1) which is interesting in its own right. Theorem 4.1 states that if a point cloud \(\mathscr {P}\) lies close to a submanifold M, and if estimated tangent spaces at each sample point are given, then there is a submanifold \(M'\) (ambient isotopic, and close to M for the Hausdorff distance) that interpolates \(\mathscr {P}\), with \(T_p M'\) agreeing with the estimated tangent spaces at each point \(p \in \mathscr {P}\). Moreover, the construction can be done so that the reach of \(M'\) is bounded in terms of the reach of M, provided that \(\mathscr {P}\) is sparse, points of \(\mathscr {P}\) lie close to M, and error on the estimated tangent spaces is small. Hence, Theorem 4.1 essentially allows to consider a noisy sample with estimated tangent spaces as an exact sample with exact tangent spaces on a proxy submanifold. This approach can provide stability for any algorithm that takes point cloud and tangent spaces as input, such as the so-called cocone complex [13].
1.2 Outline
This paper deals with the case where a sample \(\mathbb {X}_n = \{X_1,\ldots ,X_n\}\subset \mathbb {R}^D\) of size n is randomly drawn on/around M. First, the statistical framework is described (Sect. 2.1) together with minimax optimality (Sect. 2.2). Then, the main results are stated (Sect. 2.3).
Two models are studied, one where \(\mathbb {X}_n\) is corrupted with additive noise, and one where \(\mathbb {X}_n\) contains outliers. We build a simplicial complex \(\widehat{M}_{\mathtt {TDC}}(\mathbb {X}_n)\) ambient isotopic to M and we derive the rate of approximation for the Hausdorff distance \({d}_{H}(M,\widehat{M}_{\mathtt {TDC}})\), with bounds holding uniformly over a class of submanifolds satisfying a reach regularity condition. The derived rate of convergence is minimax optimal if the amplitude \(\sigma \) of the additive noise is small. With outliers, similar estimators \(\widehat{M}_{\mathtt {TDC \delta }}\) and \(\widehat{M}_{\mathtt {TDC +}}\) are built. \(\widehat{M}_{\mathtt {TDC}}\), \(\widehat{M}_{\mathtt {TDC \delta }}\) and \(\widehat{M}_{\mathtt {TDC +}}\) are based on the Tangential Delaunay Complex (Sect. 3), that is first proved to be stable (Sect. 4) via an interpolation result. A method to estimate tangent spaces and to remove outliers based on local Principal Component Analysis (PCA) is proposed (Sect. 5). We conclude with general remarks and possible extensions (Sect. 6). For ease of exposition, all the proofs are placed in the Appendix.
1.3 Notation
In what follows, we consider a compact d-dimensional submanifold without boundary \(M \subset \mathbb {R}^D\) to be reconstructed. For all \(p \in M\), \(T_p M\) designates the tangent space of M at p. Tangent spaces will either be considered vectorial or affine depending on the context. The standard inner product in \(\mathbb {R}^D\) is denoted by \(\langle \cdot ,\cdot \rangle \) and the Euclidean distance \(\left\| \cdot \right\| \). We let \(\mathscr {B}(p,r)\) denote the closed Euclidean ball of radius \(r>0\) centered at p. We let \(\wedge \) and \(\vee \) denote respectively the minimum and the maximum of real numbers. As introduced in [22], the reach of M, denoted by \(\mathrm {reach}(M)\) is the maximal offset radius for which the projection \(\pi _M\) onto M is well defined. Denoting by \({d}(\cdot ,M)\) the distance to M, the medial axis of M \(\mathrm {med}(M) = \{x \in \mathbb {R}^D \,|\, \exists a \ne b \in M, \,\left\| x-a\right\| = \left\| x-b\right\| = {d}(x,M) \}\) is the set of points which have at least two nearest neighbors on M. Then, \(\mathrm {reach}(M) = \inf _{p \in M} {d}(p,\mathrm {med}(M))\). We simply write \(\pi \) for \(\pi _M\) when there is no possibility of confusion. For any smooth function \(\Phi :\mathbb {R}^D \rightarrow \mathbb {R}^D\), we let \({d}_a \Phi \) and \({d}_a^2 \Phi \) denote the first and second order differentials of \(\Phi \) at \(a\in \mathbb {R}^D\). For a linear map A, \(A^t\) designates its transpose. Let \(\left\| A\right\| _\mathrm {op}= {\sup _x} \frac{\left\| A x\right\| }{\left\| x\right\| }\) and \(\left\| A\right\| _\mathscr {F} = \sqrt{\mathrm {trace}(A^t A)}\) denote respectively the operator norm induced by the Euclidean norm and the Frobenius norm. The distance between two linear subspaces \(U,V \subset \mathbb {R}^D\) of the same dimension is measured by the sine \(\angle (U,V) = {\max }_{u\in U} ~ {\max }_{v' \in V^{\perp }} \frac{\langle u,v' \rangle }{\left\| u\right\| \left\| v'\right\| } = \left\| \pi _U - \pi _V\right\| _\mathrm {op}\) of their largest principal angle. The Hausdorff distance between two compact subsets \(K,K'\) of \(\mathbb {R}^D\) is denoted by \({d}_{H}(K,K') = \sup _{x \in \mathbb {R}^D} |d(x,K) - d(x,K') |\). Finally, we let \(\cong \) denote the ambient isotopy relation in \(\mathbb {R}^D\).
Throughout this paper, \(C_\alpha \) will denote a generic constant depending on the parameter \(\alpha \). For clarity’s sake, \(c_\alpha \) and \(K_\alpha \) may also be used when several constants are involved.
2 Minimax Risk and Main Results
2.1 Statistical Model
Let us describe the general statistical setting we will use to define optimality for manifold reconstruction. A statistical model \(\mathscr {D}\) is a set of probability distributions on \(\mathbb {R}^D\). In any statistical experiment, \(\mathscr {D}\) is fixed and known. We observe an independent and identically distributed sample of size n (or i.i.d. n-sample) \(\mathbb {X}_n = \{X_1,\ldots ,X_n\}\) drawn according to some unknown distribution \(P \in \mathscr {D}\). If no noise is allowed, the problem is to recover the support of P, that is, the smallest closed set \(C\subset \mathbb {R}^D\) such that \(P(C)=1\). Let us give two examples of such models \(\mathscr {D}\) by describing those of interest in this paper.
Let \(\mathscr {M}_{D,d,\rho }\) be the set of all compact d-dimensional connected submanifolds \(M \subset \mathbb {R}^D\) without boundary satisfying \(\mathrm {reach}(M)\ge \rho \). The reach assumption is crucial to avoid arbitrarily curved and pinched shapes [15]. From a reconstruction point of view, \(\rho \) gives a minimal feature size on M, and then a minimal scale for geometric information. Every \(M\in \mathscr {M}_{D,d,\rho }\) inherits a measure induced by the d-dimensional Hausdorff measure on \(\mathbb {R}^D\supset M\). We denote this induced measure by \(v_M\). Beyond the geometric restrictions induced by the lower bound \(\rho \) on the reach, it also requires the natural measure \(v_M\) to behave like a d-dimensional measure, up to uniform constants. Denote by \(\mathscr {U}_M(f_\mathrm{min},f_\mathrm{max})\) the set of probability distributions Q having a density f with respect to \(v_M\) such that \(0< f_\mathrm{min} \le f(x) \le f_\mathrm{max}< \infty \) for all \(x\in M\). In particular, notice that such distributions \(Q \in \mathscr {U}_M(f_\mathrm{min},f_\mathrm{max})\) all have support M. Roughly speaking, when \(Q \in \mathscr {U}_M(f_\mathrm{min},f_\mathrm{max})\), points are drawn almost uniformly on M. This is to ensure that the sample visits all the areas of M with high probability. The noise-free model \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\) consists of the set of all these almost uniform measures on submanifolds of dimension d having reach greater than a fixed value \(\rho >0\).
Definition 2.1
(Noise-free model) \( \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho } = \bigcup _{M \in \mathscr {M}_{D,d,\rho }} \mathscr {U}_M(f_\mathrm{min},f_\mathrm{max}) \).
Notice that we do not explicitly impose a bound on the diameter of M. Actually, a bound is implicitly present in the model, as stated in the next lemma, the proof of which follows from a volume argument.
Lemma 2.2
There is \(C_d>0\) such that for all \(Q \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\) with associated M,
Observed random variables with distribution belonging to the noise-free model \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\) lie exactly on the submanifold of interest M. A more realistic model should allow some measurement error, as illustrated by Fig. 1 (a). We formalize this idea with the following additive noise model.
Definition 2.3
(Additive noise model) For \(\sigma < \rho \), we let \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\) denote the set of distributions of random variables \(X = Y + Z\), where Y has distribution \(Q \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\), and \(\left\| Z\right\| \le \sigma \) almost surely.
Let us emphasize that we do not require Y and Z to be independent, nor Z to be orthogonal to \(T_Y M\), as done for the “perpendicular” noise model of [24, 30]. This model is also slightly more general than the one considered in [28]. Notice that the noise-free model can be thought of as a particular instance of the additive noise model, since \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho } = \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma =0}\).
Eventually, we may include distributions contaminated with outliers uniformly drawn in a ball \(\mathscr {B}_0\) containing M, as illustrated in Fig. 1 (b). Up to translation, we can always assume that \(M \ni 0\). To avoid boundary effects, \(\mathscr {B}_0\) will be taken to contain M amply, so that the outlier distribution surrounds M everywhere. Since M has at most diameter \(K_{d,f_\mathrm{min},\rho }\) from Lemma 2.2 we arbitrarily fix \(\mathscr {B}_0 = \mathscr {B}(0,K_0)\), where \(K_0 = K_{d,f_\mathrm{min},\rho } + \rho \). Notice that the larger the radius of \(\mathscr {B}_0\), the easier to label the outlier points since they should be very far away from each other.
Definition 2.4
(Model with outliers/Clutter noise model) For \(0<f_\mathrm{min} \le f_\mathrm{max} < \infty \), \(0<\beta \le 1\), and \(\rho >0\), we define \(\mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta }\) to be the set of mixture distributions
where \(Q \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\) has support M such that \(0 \in M\), and \(U_{\mathscr {B}_0}\) is the uniform distribution on \(\mathscr {B}_0 = \mathscr {B}(0,K_0)\).
Alternatively, a random variable X with distribution \(P \in \mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta }\) can be represented as \(X = V X' + (1-V)X''\) , where \(V \in \{0,1\}\) is a Bernoulli random variable with parameter \(\beta \), \(X'\) has distribution in \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\) and \(X''\) has a uniform distribution over \(\mathscr {B}_0\), and such that \(V,X',X''\) are independent. In particular for \(\beta = 1\), \( \mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta =1} = \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\).
2.2 Minimax Risk
For a probability measure \(P \in \mathscr {D}\), denote by \(\mathbb {E}_P\)—or simply \(\mathbb {E}\)—the expectation with respect to the product measure \(P^{(n)}\). The quantity we will be interested in is the minimax risk associated to the model \(\mathscr {D}\). For \(n\ge 0\),
where the infimum is taken over all the estimators \(\widehat{M} = \widehat{M}(X_1,\ldots ,X_n)\) computed over an n-sample. \(R_n(\mathscr {D})\) is the best risk that an estimator based on an n-sample can achieve uniformly over the class \(\mathscr {D}\). It is clear from the definition that if \(\mathscr {D}' \subset \mathscr {D}\) then \(R_n(\mathscr {D'}) \le R_n(\mathscr {D})\). It follows the intuition that the broader the class of considered manifolds, the more difficult it is to estimate them uniformly well. Studying \(R_n(\mathscr {D})\) for a fixed n is a difficult task that can rarely be carried out. We will focus on the semi-asymptotic behavior of this risk. As \(R_n(\mathscr {D})\) cannot be surpassed, its rate of convergence to 0 as \(n\rightarrow \infty \) may be seen as the best rate of approximation that an estimator can achieve. We will say that two sequences \((a_n)_n\) and \((b_n)_n\) are asymptotically comparable, denoted by \(a_n \asymp b_n\), if there exist \(c,C>0\) such that for n large enough, \(cb_n \le a_n \le Cb_n\).
Definition 2.5
An estimator \(\widehat{M}\) is said to be (asymptotically) minimax optimal over \(\mathscr {D}\) if
In other words, \(\widehat{M}\) is (asymptotically) minimax optimal if it achieves, up to constants, the best possible rate of convergence in the worst case.
Studying a minimax rate of convergence is twofold. On one hand, deriving an upper bound on \(R_n\) boils down to provide an estimator and to study its quality uniformly on \(\mathscr {D}\). On the other hand, bounding \(R_n\) from below amounts to study the worst possible case in \(\mathscr {D}\). This part is usually achieved with standard Bayesian techniques [27]. For the models considered in the present paper, the rates were given in [24, 26].
Theorem 2.6
([26, Thm. 3]) We have
and for \(0 < \beta \le 1\) fixed,
Since the additive noise model \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\) has not yet been considered in the literature, the behavior of the associated minimax risk is not known. Beyond this theoretical result, an interesting question is to know whether these minimax rates can be achieved by a tractable algorithm. Indeed, that proposed in [24] especially rely on a minimization problem over the class of submanifolds \(\mathscr {M}_{D,d,\rho }\), which is computationally costly. In addition, the proposed estimators are themselves submanifolds, which raises storage problems. Moreover, no guarantee is given on the topology of the estimators. Throughout the present paper, we will build estimators that address these issues.
2.3 Main Results
Let us start with the additive noise model \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\), that includes in particular the noise-free case \(\sigma = 0\). The estimator \(\widehat{M}_{\mathtt {TDC}}\) is based on the Tangential Delaunay Complex (Sect. 3), with a tangent space estimation using a local PCA (Sect. 5).
Theorem 2.7
\(\widehat{M}_{\mathtt {TDC}} = \widehat{M}_{\mathtt {TDC}}(\mathbb {X}_n)\) is a simplicial complex with vertices included in \(\mathbb {X}_n\) such that the following holds. There exists \(\lambda _{d,f_\mathrm{min},f_\mathrm{max}}> 0\) such that if \(\sigma \le \lambda ({\log n}/{n} )^{1/d}\) with \(\lambda \le \lambda _{d,f_\mathrm{min},f_\mathrm{max}}\), then
Moreover, for n large enough,
It is interesting to note that the constants appearing in Theorem 2.7 do not depend on the ambient dimension D. Since \(R_n\bigl (\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\bigr ) \ge R_n\bigl (\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\bigr )\), we obtain immediately from Theorem 2.7 that \(\widehat{M}_{\mathtt {TDC}}\) achieves the minimax optimal rate \(({\log n}/{n} )^{2/d}\) over \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\) when \(\sigma \le c_{d,f_\mathrm{min},f_\mathrm{max}} ({\log n}/{n} )^{2/d}\). Note that the estimator of [28] achieves the rate \(({\log n}/{n} )^{2/(d+2)}\) when \(\sigma \le c_{d,f_\mathrm{min},f_\mathrm{max}} ({\log n}/{n} )^{2/(d+2)}\), so does the estimator of [25] for \(\sigma < \rho \) if the noise is centered and perpendicular to the submanifold. As a consequence, \(\widehat{M}_{\mathtt {TDC}}\) outperforms these two existing procedures whenever \(\sigma \ll ( \log n / n )^{2/(d+2)}\), with the additional feature of exact topology recovery. Still, for \(\sigma \gg ( \log n / n )^{1/d}\), \(\widehat{M}_{\mathtt {TDC}}\) may perform poorly compared to [25]. This might be due to the fact that the vertices of \(\widehat{M}_{\mathtt {TDC}}\) are sample points themselves, while for higher noise levels, a pre-process of the data based on local averaging could be more relevant.
In the model with outliers \(\mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta }\), with the same procedure used to derive Theorem 2.7 and an additional iterative pre-processing of the data based on local PCA to remove outliers (Sect. 5), we design an estimator of M that achieves a rate as close as wanted to the noise-free rate. Namely, for any positive \(\delta < 1/(d(d+1))\), we build \(\widehat{M}_{\mathtt {TDC \delta }}\) that satisfies the following similar statement.
Theorem 2.8
\(\widehat{M}_{\mathtt {TDC}\delta } = \widehat{M}_{\mathtt {TDC}\delta }(\mathbb {X}_n)\) is a simplicial complex with vertices included in \(\mathbb {X}_n\) such that
Moreover, for n large enough,
\(\widehat{M}_{\mathtt {TDC \delta }}\) converges at the rate at least \(( {\log n}/{n} )^{2/d - 2\delta }\), which is not the minimax optimal rate according to Theorem 2.6, but that can be set as close as desired to it. To our knowledge, \(\widehat{M}_{\mathtt {TDC\delta }}\) is the first explicit estimator to provably achieve such a rate in the presence of outliers. Again, it is worth noting that the constants involved in Theorem 2.8 do not depend on the ambient dimension D. The construction and computation of \(\widehat{M}_{\mathtt {TDC \delta }}\) is the same as \(\widehat{M}_{\mathtt {TDC}}\), with an extra pre-processing of the point cloud allowing to remove outliers. This decluttering procedure leads to compute, at each sample point, at most \(\log (1/\delta )\) local PCA’s, instead of a single one for \(\widehat{M}_{\mathtt {TDC}}\).
From a theoretical point of view, there exists a (random) number of iterations of this decluttering process, from which an estimator \(\widehat{M}_{\mathtt {TDC +}}\) can be built to satisfy the following.
Theorem 2.9
\(\widehat{M}_{\mathtt {TDC +}} = \widehat{M}_{\mathtt {TDC +}}(\mathbb {X}_n)\) is a simplicial complex of vertices contained in \(\mathbb {X}_n\) such that
Moreover, for n large enough,
\(\widehat{M}_{\mathtt {TDC +}}\) may be thought of as a limit of \(\widehat{M}_{\mathtt {TDC \delta }}\) when \(\delta \) goes to 0. As it will be proven in Sect. 5, this limit will be reached for \(\delta \) close enough to 0. Unfortunately this convergence threshold is also random, hence unknown.
The statistical analysis of the reconstruction problem is postponed to Sect. 5. Beforehand, let us describe the Tangential Delaunay Complex in a deterministic and idealized framework where the tangent spaces are known and no outliers are present.
3 Tangential Delaunay Complex
Let \(\mathscr {P}\) be a finite subset of \(\mathbb {R}^D\). In this section, we denote the point cloud \(\mathscr {P}\) to emphasize the fact that it is considered nonrandom. For \(\varepsilon ,\delta >0\), \(\mathscr {P}\) is said to be \(\varepsilon \)-dense in M if \(\sup _{x \in M}d(x,\mathscr {P}) \le \varepsilon \), and \(\delta \)-sparse if \({d}(p,\mathscr {P} {\setminus }\{p\}) \ge \delta \) for all \(p \in \mathscr {P}\). A \((\delta ,\varepsilon )\)-net (of M) is a \(\delta \)-sparse and \(\varepsilon \)-dense point cloud.
3.1 Restricted Weighted Delaunay Triangulations
We now assume that \(\mathscr {P} \subset M\). A weight assignment to \(\mathscr {P}\) is a function \(\omega :\mathscr {P} \rightarrow [0,\infty )\). The weighted Voronoi diagram is defined to be the Voronoi diagram associated to the weighted distance \({d}(x,p^\omega )^2 = \left\| x-p\right\| ^2 - \omega (p)^2\). Every \(p \in \mathscr {P}\) is associated to its weighted Voronoi cell \( \mathrm {Vor}^\omega (p)\). For \(\tau \subset \mathscr {P}\), let
be the common face of the weighted Voronoi cells of the points of \(\tau \). The weighted Delaunay triangulation \(\mathrm {Del}^\omega (\mathscr {P})\) is the dual triangulation to the decomposition given by the weighted Voronoi diagram. In other words, for \(\tau \subset \mathscr {P}\), the simplex with vertices \(\tau \), also denoted by \(\tau \), satisfies
Note that for a constant weight assignment \(\omega (p) \equiv \omega _0\), \(\mathrm {Del}^\omega (\mathscr {P})\) is the usual Delaunay triangulation of \(\mathscr {P}\). Under genericity assumptions on \(\mathscr {P}\) and bounds on \(\omega \), \(\mathrm {Del}^\omega (\mathscr {P})\) is an embedded triangulation with vertex set \(\mathscr {P}\) [4]. The reconstruction method proposed in this paper is based on \(\mathrm {Del}^\omega (\mathscr {P})\) for some weights \(\omega \) to be chosen later. As it is a triangulation of the whole convex hull of \(\mathscr {P}\) and fails to recover the geometric structure of M, we take restrictions of it in the following manner.
Given a family \(R = \{R_p\}_{p\in \mathscr {P}}\) of subsets \(R_p \subset \mathbb {R}^D\) indexed by \(\mathscr {P}\), the weighted Delaunay complex restricted to R is the sub-complex of \(\mathrm {Del}^\omega (\mathscr {P})\) defined by
In particular, we define the Tangential Delaunay Complex \(\mathrm {Del}^\omega (\mathscr {P},T)\) by taking \(R = T = \{T_p M \}_{p\in \mathscr {P}}\), the family of tangent spaces taken at the points of \(\mathscr {P} \subset M\) [4]. \(\mathrm {Del}^\omega (\mathscr {P},T)\) is a pruned version of \(\mathrm {Del}^\omega (\mathscr {P})\) where only the simplices with directions close to the tangent spaces are kept. Indeed, \(T_p M\) being the best linear approximation of M at p, it is very unlikely for a reconstruction of M to have components in directions normal to \(T_p M\) (see Fig. 2). As pointed out in [4], computing \(\mathrm {Del}^\omega (\mathscr {P},T)\) only requires to compute Delaunay triangulations in the tangent spaces that have dimension d. This reduces the computational complexity dependency on the ambient dimension \(D > d\). The weight assignment \(\omega \) gives degrees of freedom for the reconstruction. The extra degree of freedom \(\omega \) permits to stabilize the triangulation and to remove the so-called inconsistencies, the points remaining fixed. For further details, see [4, 5].
3.2 Guarantees
The following result sums up the reconstruction properties of the Tangential Delaunay Complex that we will use. For more details about it, the reader is referred to [4].
Theorem 3.1
([4, Thm. 5.3]) There exists \(\varepsilon _d>0\) such that for all \(\varepsilon \le \varepsilon _d \rho \) and all \(M \in \mathscr {M}_{D,d,\rho }\), if \(\mathscr {P} \subset M\) is an \(( \varepsilon ,2\varepsilon )\)-net, there exists a weight assignment \(\omega _*= \omega _{*\mathscr {P},T}\) depending on \(\mathscr {P}\) and \(T = \{T_p M\}_{p\in \mathscr {P}}\) such that
-
\({d}_{H} ( M, \mathrm {Del}^{\omega _*}(\mathscr {P},T) ) \le {C_{d}} {\varepsilon ^2}/{\rho }\),
-
M and \(\mathrm {Del}^{\omega _*}(\mathscr {P},T)\) are ambient isotopic.
Computing \(\mathrm {Del}^{\omega _*}(\mathscr {P},T)\) requires to determine the weight function \(\omega _*= \omega _{*\mathscr {P},T}\). In [4], a greedy algorithm is designed for this purpose and has a time complexity \({O}( Dn^2 + D 2^{O(d^2)}n )\).
Given an \((\varepsilon ,2\varepsilon )\)-net \(\mathscr {P}\) for \(\varepsilon \) small enough, \(\mathrm {Del}^{\omega _*}(\mathscr {P},T)\) recovers M up to ambient isotopy and approximates it at the scale \(\varepsilon ^2\). The order of magnitude \(\varepsilon ^2\) with an input \(\mathscr {P}\) of scale \(\varepsilon \) is remarkable. Another instance of this phenomenon is present in [14] in codimension 1. We will show that this \(\varepsilon ^2\) provides the minimax rate of approximation when dealing with random samples. Therefore, it can be thought of as optimal.
Theorem 3.1 suffers two major imperfections. First, it requires the knowledge of the tangent spaces at each sample point — since \({\omega _*} = {\omega _{*\mathscr {P}, T}}\)—and it is no longer usable if tangent spaces are only known up to some error. Second, the points are assumed to lie exactly on the submanifold M, and no noise is allowed. The analysis of \(\mathrm {Del}^{\omega _*}(\mathscr {P},T)\) is sophisticated [4]. Rather than redo the whole study with milder assumptions, we tackle this question with an approximation theory approach (Theorem 4.1). Instead of studying if \(\mathrm {Del}^{\omega _*}(\mathscr {P}',T')\) is stable when \(\mathscr {P}'\) lies close to M and \(T'\) close to T, we examine what \(\mathrm {Del}^{\omega _*}(\mathscr {P}',T')\) actually reconstructs, as detailed in Sect. 4.
3.3 On the Sparsity Assumption
In Theorem 3.1, \(\mathscr {P}\) is assumed to be dense enough so that it covers all the areas of M. It is also supposed to be sparse at the same scale as the density parameter \(\varepsilon \). Indeed, arbitrarily accumulated points would generate non-uniformity and instability for \(\mathrm {Del}^{\omega _*}(\mathscr {P},T)\) [4, 5]. At this stage, we emphasize that the construction of an \((\varepsilon ,2\varepsilon )\)-net can be carried out given an \(\varepsilon \)-dense sample. Given an \(\varepsilon \)-dense sample \(\mathscr {P}\), the farthest point sampling algorithm prunes \(\mathscr {P}\) and outputs an \((\varepsilon ,2 \varepsilon )\)-net \(\mathscr {Q} \subset \mathscr {P}\) of M as follows. Initialize at \(\mathscr {Q} = \{p_1\} \subset \mathscr {P}\), and while \({\max \nolimits _{p \in \mathscr {P}}} ~ {d}(p,\mathscr {Q}) > \varepsilon \), add to \(\mathscr {Q}\) the farthest point to \(\mathscr {Q}\) in \(\mathscr {P}\), that is, \(\mathscr {Q} \leftarrow \mathscr {Q} \cup \{ {\mathop {\hbox {argmax}}\limits \nolimits _{p \in \mathscr {P}}} ~ {d}(p,\mathscr {Q}) \}\). The output \(\mathscr {Q}\) is \(\varepsilon \)-sparse and satisfies \({d}_{H}(\mathscr {P},\mathscr {Q})\le \varepsilon \), so it is an \((\varepsilon ,2\varepsilon )\)-net of M. Therefore, up to the multiplicative constant 2, sparsifying \(\mathscr {P}\) at scale \(\varepsilon \) will not deteriorate its density property. Then, we can run the farthest point sampling algorithm to pre-process the data, so that the obtained point cloud is a net.
4 Stability Result
4.1 Interpolation Theorem
As mentioned above, if the data do not lie exactly on M and if we do not have the exact knowledge of the tangent spaces, Theorem 3.1 does not apply. To bypass this issue, we interpolate the data with another submanifold \(M'\) satisfying good properties, as stated in the following result.
Theorem 4.1
(Interpolation) Let \(M \in \mathscr {M}_{D,d,\rho }\). Let \(\mathscr {P} = \{p_1,\ldots ,p_q\} \subset \mathbb {R}^D\) be a finite point cloud and \(\widetilde{T} = \{ \widetilde{T}_1,\ldots ,\widetilde{T}_q \}\) be a family of d-dimensional linear subspaces of \(\mathbb {R}^D\). For \(\theta \le \pi /64\) and \(18 \eta <\delta \le \rho \), assume that
-
\(\mathscr {P}\) is \(\delta \)-sparse: \({\min \nolimits _{i\ne j}} \Vert {p_j - p_i}\Vert \ge \delta \),
-
the \(p_j\)’s are \(\eta \)-close to M: \({\max \nolimits _{1 \le j \le q}} {d}(p_j,M) \le \eta \),
-
\({\max \nolimits _{1 \le j \le q}} \, \angle (T_{\pi _M(p_j)} M, \widetilde{T}_j) \le \sin \theta \).
Then, there exist a universal constant \(c_0 \le 285\) and a compact d-dimensional connected submanifold \(M' \subset \mathbb {R}^D\) without boundary such that
-
1.
\(\mathscr {P} \subset M'\),
-
2.
\(\mathrm {reach}\bigl (M'\bigr ) \ge \bigl ( 1- c_0 \bigl (\frac{\eta }{\delta } +{\theta }\bigr )\, \frac{\rho }{\delta } \bigr ) \rho \),
-
3.
\( T_{p_j} M' = \widetilde{T}_j\) for all \(1 \le j \le q\),
-
4.
\({d}_{H}(M,M') \le \delta \theta + \eta \),
-
5.
M and \(M'\) are ambient isotopic.
Theorem 4.1 fits a submanifold \(M'\) to noisy points and perturbed tangent spaces with no change of topology and a controlled reach loss. We will use \(M'\) as a proxy for M. Indeed, if \(\widetilde{T}_1,\ldots ,\widetilde{T}_q\) are estimated tangent spaces at the noisy base points \(p_1,\ldots ,p_q\), \(M'\) has the virtue of being reconstructed by \(\mathrm {Del}^{\omega _*}(\mathscr {P},\widetilde{T})\) from Theorem 3.1. Since \(M'\) is topologically and geometrically close to M, we conclude that M is reconstructed as well by transitivity. In other words, Theorem 4.1 allows to consider a noisy sample with estimated tangent spaces as an exact sample with exact tangent spaces. \(M'\) is built pushing and rotating M towards the \(p_j\)’s locally along the vector \((p_j - \pi (p_j))\), as illustrated in Fig. 3. Since the construction is quite general and may be applied in various settings, let us provide an outline of the construction.
Let \(\phi (x) = \exp \bigl (\frac{ \left\| x\right\| ^2}{\left\| x\right\| ^2-1} \bigr ) \mathbb {1}_{\left\| x\right\| ^2 < 1}\). \(\phi \) is smooth and satisfies \(\phi (0) = 1, \left\| \phi \right\| _\infty \le 1\) and \({d}_0 \phi = 0\). For \(j = 1, \ldots , q\), it follows easily from the definition of \(\angle (T_{\pi (p_j)} M, \widetilde{T}_j)\)—e.g. by induction on the dimension—that there exists a rotation \(R_j\) of \(\mathbb {R}^D\) mapping \(T_{\pi (p_j)} M\) onto \(\widetilde{T}_j\) that satisfies \(\Vert {R_j-I_D}\Vert _\mathrm{op} \le 2 \sin ( \theta / 2 ) \le \theta \). For \(\ell > 0\) to be chosen later, and all \(a \in \mathbb {R}^D\), let us define \( \Phi :\mathbb {R}^D \rightarrow \mathbb {R}^D\) by
\(\Phi \) is designed to map \(\pi (p_j)\) onto \(p_j\) with \({d}_{\pi (p_j)} \Phi = R_j\). Roughly speaking, in balls of radii \(\ell \) around each \(\pi (p_j)\), \( \Phi \) shifts the points in the direction \(p_j - \pi (p_j)\) and rotates it around \(\pi (p_j)\). Off these balls, \( \Phi \) is the identity map. To guarantee smoothness, the shifting and the rotation are modulated by the kernel \(\phi \), as \(\Vert {a - \pi (p_j)}\Vert \) increases. Notice that \({d}_a \psi _j = (R_j - I_D)\) and \(\Vert {\psi _j(a)}\Vert \le \ell \theta + \eta \) whenever \(\phi \bigl (\frac{a-\pi (p_j)}{\ell } \bigr ) \ne 0\). Defining \(M' = \Phi (M)\), the facts that \(M'\) fits to \(\mathscr {P}\) and \(\widetilde{T}\) and is Hausdorff-close to M follow by construction. Moreover, Theorem 4.19 of [22] (reproduced as Lemma 7.1 in this paper) states that the reach is stable with respect to \(\mathscr {C}^2\)-diffeomorphisms of the ambient space. The estimate on \(\mathrm {reach}(M')\) relies on the following lemma stating differentials estimates on \(\Phi \).
Lemma 4.2
There exist universal constants \(C_1\le 7/2\) and \(C_2\le 28\) such that if \( 6 \eta < \ell \le \delta /3\) and \(\theta \le \pi /64\), \( \Phi :\mathbb {R}^D \rightarrow \mathbb {R}^D\) is a global \(\mathscr {C}^\infty \)-diffeomorphism. In addition, for all a in \(\mathbb {R}^D\),
The ambient isotopy follows easily by considering the weighted version \( \Phi _{(t)}(a) = a + t(\Phi (a)-a)\) for \(0\le t\le 1\) and the same differential estimates. We then take the maximum possible value \(\ell = \delta /3\) and \(M' = \Phi (M)\).
Remark 4.3
Changing slightly the construction of \(M'\), one can also build it such that the curvature tensor at each \(p_j\) corresponds to that of M at \(\pi (p_j)\). For this purpose it suffices to take a localizing function \(\phi \) identically equal to 1 in a neighborhood of 0. This additional condition would impact the universal constant \(c_0\) appearing in Theorem 4.1.
4.2 Stability of the Tangential Delaunay Complex
Theorem 4.1 shows that even in the presence of noisy sample points at distance \(\eta \) from M, and with the knowledge of the tangent spaces up to some angle \(\theta \), it is still possible to apply Theorem 3.1 to some virtual submanifold \({M}'\). Denoting \(\widetilde{M} = \mathrm {Del}^{\omega _*}(\mathscr {P},\widetilde{T})\), since \({d}_{H}(M,\widetilde{M}) \le {d}_{H}(M,M') + {d}_{H}(M',\widetilde{M})\) and since the ambient isotopy relation is transitive, \(M \cong M' \cong \widetilde{M}\). We get the following result as a straightforward combination of Theorems 3.1 and 4.1.
Theorem 4.4
(Stability of the Tangential Delaunay Complex) There exists \(\varepsilon _{d}>0\) such that for all \(\varepsilon \le \varepsilon _{d}\rho \) and all \(M \in \mathscr {M}_{D,d,\rho }\), the following holds. Let \(\mathscr {P} \subset \mathbb {R}^D\) be a finite point cloud and \(\widetilde{T} = \{\widetilde{T}_p\}_{p\in \mathscr {P}}\) be a family of d-dimensional linear subspaces of \(\mathbb {R}^D\) such that
-
\({\max \nolimits _{p \in \mathscr {P}}} \, {d}(p,M) \le \eta \),
-
\({\max \nolimits _{p \in \mathscr {P}}} \, \angle (T_{\pi _M(p)} {M},\widetilde{T}_p ) \le \sin \theta \),
-
\(\mathscr {P}\) is \(\varepsilon \)-sparse,
-
\({\max \nolimits _{x\in M}} \, {d}(x,\mathscr {P}) \le 2\varepsilon \).
If \(\theta \le \varepsilon /(1140 \rho )\) and \(\eta \le \varepsilon ^2/(1140 \rho )\), then
-
\({d}_{H} ( M, \mathrm {Del}^{\omega _*}(\mathscr {P},\widetilde{T}) ) \le C_{d}\varepsilon ^2/\rho \),
-
M and \(\mathrm {Del}^{\omega _*}(\mathscr {P},\widetilde{T})\) are ambient isotopic.
Indeed, applying the reconstruction algorithm of Theorem 3.1 even in the presence of noise and uncertainty on the tangent spaces actually recovers the submanifold \(M'\) built in Theorem 4.1. \(M'\) is isotopic to M and the quality of the approximation of M is at most impacted by the term \({d}_{H}(M,M') \le \varepsilon \theta + \eta \). The lower bound on \(\mathrm {reach}(M')\) is crucial, as constants appearing in Theorem 3.1 are not bounded for arbitrarily small reach.
It is worth noting that no extra analysis of the Tangential Delaunay Complex was needed to derive its stability. The argument is global, constructive, and may be applied to other reconstruction methods taking tangent spaces as input. For instance, a stability result similar to Theorem 4.4 could be derived readily for the so-called cocone complex [13] using the interpolating submanifold of Theorem 4.1.
5 Tangent Space Estimation and Decluttering Procedure
5.1 Additive Noise Case
We now focus on the estimation of tangent spaces in the model with additive noise \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\). The proposed method is similar to that of [2, 28]. A point \(p \in M\) being fixed, \(T_p M\) is the best local d-dimensional linear approximation of M at p. Performing a Local Principal Component Analysis (PCA) in a neighborhood of p is likely to recover the main directions spanned by M at p, and therefore to yield a good approximation of \(T_p M\). For \(j=1,\ldots ,n\) and \(h>0\) to be chosen later, define the local covariance matrix at \(X_j\) by
where \(\overline{X}_j= \frac{1}{N_j} \sum _{i \ne j} {X_i \mathbb {1}_{\mathscr {B}(X_j,h)}(X_i)}\) is the barycenter of sample points contained in the ball \(\mathscr {B}(X_j,h)\), and \(N_j = |\mathscr {B}(X_j,h)\cap \mathbb {X}_n|\). Let us emphasize the fact that the normalization \(1/(n-1)\) in the definition of \(\widehat{\Sigma }_j\) stands for technical convenience. In fact, any other normalization would yield the same guarantees on tangent spaces since only the principal directions of \(\widehat{\Sigma }_j\) play a role. Set \(\widehat{T}_j(h)\) to be the linear space spanned by the d eigenvectors associated with the d largest eigenvalues of \(\widehat{\Sigma }_j(h)\). Computing a basis of \(\widehat{T}_j(h)\) can be performed naively using a singular value decomposition of the full matrix \(\widehat{\Sigma }_j(h)\), although fast PCA algorithms [31] may lessen the computational dependence on the ambient dimension. We also denote by \(\mathtt {TSE}(.,h)\) the function that maps any vector of points to the vector of their estimated tangent spaces, with
Proposition 5.1
Set \(h = \bigl (c_{d,f_\mathrm{min},f_\mathrm{max}} \frac{\log n}{n-1} \bigr )^{1/d}\) for \(c_{d,f_\mathrm{min},f_\mathrm{max}}\) large enough. Assume that \(\sigma /h \le 1/4\). Then for n large enough, for all \(Q \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\),
with probability larger than \(1-4 ( {1}/{n} )^{{2}/{d}}\).
An important feature given by Proposition 5.1 is that the statistical error of our tangent space estimation procedure does not depend on the ambient dimension D. The intuition behind Proposition 5.1 is the following: if we assume that the true tangent space \(T_{X_j}M\) is spanned by the first d vectors of the canonical basis, we can decompose \(\widehat{\Sigma }_j\) as
where \(\widehat{R}\) comes from the curvature of the submanifold along with the additive noise, and is of order \(N_j(h)(h^{3}/(\rho (n-1))+h \sigma ) \lesssim h^{d+2}(h/\rho + \sigma /h)\), provided that h is roughly smaller than \((\log (n)/(n-1))^{1/d}\). On the other hand, for a bandwidth h of order \((\log (n)/(n-1))^{1/d}\), \(\widehat{A}_j(h)\) can be proved (Lemma 9.5) to be close to its deterministic counterpart
where \(\pi _{T_{X_j} M}\) denotes orthogonal projection onto \(T_{X_j}M\) and expectation is taken conditionally on \(X_j\). The bandwidth \((\log (n)/(n-1))^{1/d}\) may be thought of as the smallest radius that allows enough sample points in balls to provide an accurate estimation of the covariance matrices. Then, since \(f_\mathrm{min} >0\), Lemma 9.4 shows that the minimum eigenvalue of A(h) is of order \(h^{d+2}\). At last, an eigenvalue perturbation result (Proposition 10.1) shows that \(\widehat{T}_j(h)\) must be close to \(T_{X_j}M\) up to \((h^{d+3}/\rho + h^{d+1} \sigma )/(h^{d+2}) \approx h/\rho + \sigma /h\). The complete derivation is provided in Sect. 1.
Then, it is shown in Lemma 9.1, based on the results of [12], that letting \(\varepsilon = c_{d,f_\mathrm{min},f_\mathrm{max}} (h \vee \rho \sigma /h)\) for \(c_{d,f_\mathrm{min},f_\mathrm{max}}\) large enough, entails \(\mathbb {X}_n\) is \(\varepsilon \)-dense in M with probability larger than \(1 - ({1}/{n})^{2/d}\). Since \(\mathbb {X}_n\) may not be sparse at the scale \(\varepsilon \), and for the stability reasons described in Sect. 3, we sparsify it with the farthest point sampling algorithm (Sect. 3.3) with scale parameter \(\varepsilon \). Let \(\mathbb {Y}_n\) denote the output of the algorithm. If \(\sigma \le h/4\), and \(c_{d,f_\mathrm{min},f_\mathrm{max}}\) is large enough, we have the following.
Corollary 5.2
With the above notation, for n large enough, with probability at least \(1-5( {1}/{n} )^{2/d}\),
-
\({\max \nolimits _{X_j \in \mathbb {Y}_n}} \, {d}(X_j,M) \le \frac{\varepsilon ^2}{1140 \rho }\),
-
\({\max \nolimits _{X_j \in \mathbb {Y}_n}} \angle ( T_{\pi _M(X_j)} {M},\widehat{T}_j(h) ) \le \frac{\varepsilon }{2280\rho }\),
-
\(\mathbb {Y}_n\) is \(\varepsilon \)-sparse,
-
\({\max \nolimits _{x\in M}} \, {d}(x,\mathbb {Y}_n) \le 2 \varepsilon \).
In other words, the previous result shows that \(\mathbb {Y}_n\) satisfies the assumptions of Theorem 4.4 with high probability. We may then define \(\widehat{M}_{\mathtt {TDC}}\) to be the Tangential Delaunay Complex computed on \(\mathbb {Y}_n\) and the collection of estimated tangent spaces \(\mathtt {TSE}(\mathbb {X}_n,h)_{\mathbb {Y}_n}\), that is elements of \(\mathtt {TSE}(\mathbb {X}_n,h)\) corresponding to elements of \(\mathbb {Y}_n\), where h is the bandwidth defined in Proposition 5.1.
Definition 5.3
With the above notation, define \(\widehat{M}_{\mathtt {TDC}} = \mathrm {Del}^{\omega _*}(\mathbb {Y}_n,\mathtt {TSE}(\mathbb {X}_n,h)_{\mathbb {Y}_n} )\).
Combining Theorem 4.4 and Corollary 5.2, it is clear that \(\widehat{M}_{\mathtt {TDC}}\) satisfies Theorem 2.7.
5.2 Clutter Noise Case
Let us now focus on the model with outliers \(\mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta }\). We address the problem of decluttering the sample \(\mathbb {X}_n\), that is, to remove outliers (see Fig. 4). We follow ideas from [24]. To distinguish whether \(X_j\) is an outlier or belongs to M, we notice again that points drawn from M approximately lie on a low dimensional structure. On the other hand, the neighborhood points of an outlier drawn far away from M should typically be distributed in an isotropic way. Let \(k_1,k_2,h>0\), \(x\in \mathbb {R}^D\) and \(T\subset \mathbb {R}^D\) a d-dimensional linear subspace. The slab at x in the direction T is the set \(S(x,T,h) = \{x\} \oplus \mathscr {B}_{T}(0,k_1h) \oplus \mathscr {B}_{T^\perp }\!(0,k_2h^2) \subset \mathbb {R}^D\), where \(\oplus \) denotes the Minkowski sum, and \(\mathscr {B}_T,\mathscr {B}_{T^\perp }\) are the Euclidean balls in T and \(T^\perp \) respectively (Fig. 5).
Following notation of Sect. 2.1, for \(P \in \mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta }\), let us write \(P = \beta Q + (1-\beta )U_{\mathscr {B}_0}\). For h small enough, by definition of the slabs, \(U_{\mathscr {B}_0}( S(x,T_{\pi (x)} M,h) ) \asymp (k_1h)^d (k_2 h^2)^{D-d} \asymp h^{2D-d}\). Furthermore, Fig. 5 indicates that for \(k_1\) and \(k_2\) small enough, \(Q ( S(x,T_{\pi (x)} M,h) ) \asymp \mathrm{Vol}( S(x,T_{\pi (x)} M,h) \cap M ) \asymp h^d\) if \({d}(x,M) \le h^2\), and \(Q ( S(x,T_{\pi (x)} M,h) ) = 0\) if \({d}(x,M)> h^2\). Coming back to \(P = \beta Q + (1-\beta )U_{\mathscr {B}_0}\), we roughly get
as h goes to 0, for \(k_1\) and \(k_2\) small enough. Since \(h^{2D-d} \ll h^{d}\), the measure P(S(x, T, h)) of the slabs clearly is discriminatory for decluttering, provided that tangent spaces are known.
Based on this intuition, we define the elementary step of our decluttering procedure as the map \(\mathtt {SD}_t(.,.,h)\), that sends a vector \(\mathscr {P} = (p_1, \ldots , p_r ) \subset \mathbb {R}^D\) and a corresponding vector of (estimated) tangent spaces \(T_\mathscr {P} =( T_{1}, \ldots , T_r )\) onto a subvector of \(\mathscr {P}\) according to the rule
where t is a threshold to be fixed. This procedure relies on counting how many sample points lie in the slabs in direction of the estimated tangent spaces (see Fig. 5).
Since tangent spaces are unknown, the following result gives some insight on the relation between the accuracy of the tangent space estimation and the decluttering performance that can be reached.
Lemma 5.4
Let \(K>0\) be fixed. There exist constants \(k_1(K)\) and \(k_2(\rho ,K)\) such that for every \(h \le 1\) and x in \(\mathbb {R}^D\), \(S(x,T,h) \subset \mathscr {B}(x,h/2)\). Moreover, for every \(h \le h_+\wedge 1\) we have
where \(S'(x,T_{\pi _M(x)}M,h)\) is a larger slab with parameters \(k'_1(\rho ,K)\) and \(k'_2(\rho ,K)\), which are such that \(S'(x,T_{\pi _M(x)}M,h) \cap M = \emptyset \). In addition, there exists \(k_3(\rho ,K)\) such that for all x and y in M,
Possible values for \(k_1\) and \(k_2\) are, respectively, \(\frac{1}{ 16(K \vee 1)}\) and \(\frac{1}{16(\rho \vee K \vee 1)}\), and \(k_3\) can be taken as \( k_1 \wedge \frac{\rho k_2}{1 + 2K}\).
The proof of Lemma 5.4, mentioned in [24], follows from elementary geometry, combined with the definition of the reach and Proposition 8.1.
Roughly, Lemma 5.4 states that the decluttering performance is of order the square of the tangent space precision, hence will be closely related to the performance of the tangent space estimation procedure \(\mathtt {TSE}\). Unfortunately, a direct application of \(\mathtt {TSE}\) to the corrupted sample \(\mathbb {X}_n\) leads to slightly worse precision bounds, in terms of angle deviation. Typically, the angle deviation would be of order \(n^{-1/(d+1)}\). However, this precision is enough to remove outliers points which are at distance at least \(n^{-2/(d+1)}\) from M. Then running our \(\mathtt {TSE}\) on this refined sample \(\mathtt {SD}_t(\mathbb {X}_n,\mathtt {TSE}(\mathbb {X}_n),n^{-1/(d+1)})\) leads to better angle deviation rates, hence better decluttering performance, and so on.
Let us introduce an iterative decluttering procedure in a more formal way. We choose the initial bandwidth \(h_0 = \bigl ( c_{d,f_\mathrm{min},f_\mathrm{max},\rho } \frac{\log n}{\beta (n-1)} \bigr )^{\gamma _0}\), with \(\gamma _0 = 1/(d+1)\), and define the first set \(\mathbb {X}^{(-1)} = \mathbb {X}_n\) as the whole sample. We then proceed recursively, setting \(h_{k+1} = \bigl ( c_{d,f_\mathrm{min},f_\mathrm{max},\rho } \frac{\log n}{\beta (n-1)} \bigr )^{\gamma _{k+1}}\), with \(\gamma _{k+1}\) satisfying \(\gamma _{k+1} = (2 \gamma _k +1)/(d+2)\). This recursion formula is driven by the optimization of a trade-off between imprecision terms in tangent space estimation, as may be seen from (5). An elementary calculation shows that
With this updated bandwidth we define
In other words, at step \(k +1\) we use a smaller bandwidth \(h_{k+1}\) in the tangent space estimation procedure \(\mathtt {TSE}\). Then we use this better estimation of tangent spaces to run the elementary decluttering step \(\mathtt {SD}\). The performance of this procedure is guaranteed by the following proposition. With a slight abuse of notation, if \(X_j\) is in \(\mathbb {X}^{(k)}\), \(\mathtt {TSE}(\mathbb {X}^{(k)},h)_j\) will denote the corresponding tangent space of \(\mathtt {TSE}(\mathbb {X}^{(k)},h)\).
Proposition 5.5
In the clutter noise model, for t, \(c_{d,f_\mathrm{min},f_\mathrm{max},\rho }\) and n large enough, \(k_1\) and \(k_2\) small enough, the following properties hold with probability larger than \(1-7( {1}/{n} )^{2/d}\) for all \(k\ge 0\).
Initialization
-
For all \(X_j\) \(\in \) \(\mathbb {X}^{(-1)}\) such that \({d}(X_j,M) \le h_0/\sqrt{2}\), \(\angle \bigl ( \mathtt {TSE}(\mathbb {X}^{(-1)},h_0)_j, T_{\pi (X_j)} {M} \bigr ) \le C_{d,f_\mathrm{min},f_\mathrm{max}} h_0/\rho \).
-
For every \(X_j\) \(\in \) \(M \cap \mathbb {X}^{(-1)}\), \(X_j\) \(\in \) \(\mathbb {X}^{(0)}\).
-
For every \(X_j\in \) \(\mathbb {X}^{(-1)}\), if \({d}(X_j,M) > h_0^2/\rho \), then \(X_j\) \(\notin \) \(\mathbb {X}^{(0)}\).
Iterations
-
For all \(X_j\) \(\in \) \(\mathbb {X}^{(k)}\) such that \({d}(X_j,M) \le h_{k+1}/\sqrt{2}\), \(\angle \bigl ( \mathtt {TSE}(\mathbb {X}^{(k)},h_{k+1})_j, T_{\pi (X_j)} {M} \bigr ) \le C_{d,f_\mathrm{min},f_\mathrm{max}} h_{k+1}/\rho \).
-
For every \(X_j\) \(\in \) \(M \cap \mathbb {X}^{(k)}\), \(X_j\) \(\in \) \(\mathbb {X}^{(k+1)}\).
-
For every \(X_j\) \(\in \) \(\mathbb {X}^{(k)}\), if \({d}(X_j,M) > h_{k+1}^2/\rho \), then \(X_j\) \(\notin \) \(\mathbb {X}^{(k+1)}\).
This result is threefold. Not only can we distinguish data and outliers within a decreasing sequence of offsets of radii \(h_k^2/\rho \) around M, but we can also ensure that no point of M is removed during the process with high probability. Moreover, it also provides a convergence rate for the estimated tangent spaces \(\mathtt {TSE}(\mathbb {X}_k,h_{k+1})\). Now fix a precision level \(\delta \). If k is larger than \((\log (1/\delta ) - \log (d(d+1)) /(\log (d+2) - \log (2))\), then \(1/d > \gamma _k \ge 1/d - \delta \). Let us define \(k_\delta \) as the smallest integer satisfying \(\gamma _{k} \ge 1/d -\delta \), and denote by \(\mathbb {Y}_n^\delta \) the output of the farthest point sampling algorithm applied to \(\mathbb {X}^{(k_\delta )}\) with parameter \(\varepsilon = c_{d,f_\mathrm{min}f_\mathrm{max}} h_{k_\delta }\), for \(c_{d,f_\mathrm{min}f_\mathrm{max}}\) large enough. Define also \(\widehat{T}^\delta \) as the restriction of \(\mathtt {TSE}(\mathbb {X}^{(k_\delta )},h_{k_\delta })\) to the elements of \(\mathbb {Y}_n^{\delta }\).
According to Proposition 5.5, the decluttering procedure removes no data point on M with high probability. In other words, \(\mathbb {X}^{(k_\delta )} \cap M = \mathbb {X}_n \cap M\), and as a consequence, \({\max }_{x\in M} \, {d}(x,\mathbb {X}^{(k_\delta )}) \le c_{d,f_\mathrm{min}} ( {\log n}/{(\beta n)})^{1/d} \ll h_{k_\delta }\) with high probability (see Lemma 9.1). As a consequence, we obtain the following.
Corollary 5.6
With the above notation, for n large enough, with probability larger than \(1- 8 ( {1} /n )^{2/d}\),
-
\({\max \nolimits _{X_j \in \mathbb {Y}_n^\delta }} {d}(X_j,M) \le \frac{\varepsilon ^2}{1140 \rho }\),
-
\({\max \nolimits _{X_j \in \mathbb {Y}_n^\delta }} \angle ( T_{\pi _M(X_j)} {M} , \widehat{T}^\delta _j ) \le \frac{\varepsilon }{2280\rho }\),
-
\(\mathbb {Y}_n^\delta \) is \(\varepsilon \)-sparse,
-
\({\max \nolimits _{x\in M}} \, {d}(x,\mathbb {Y}_n^\delta ) \le 2 \varepsilon \).
We are now able to define the estimator \(\widehat{M}_{\mathtt {TDC \delta }}\).
Definition 5.7
With the above notation, define \( \widehat{M}_{\mathtt {TDC \delta }} = \mathrm {Del}^{\omega _*}(\mathbb {Y}^\delta _n,\widehat{T}^\delta ). \)
Combining Theorem 4.4 and Corollary 5.6, it is clear that \(\widehat{M}_{\mathtt {TDC \delta }}\) satisfies Theorem 2.8.
Finally, we turn to the asymptotic estimator \(\widehat{M}_{\mathtt {TDC +}}\). Set \(h_{\infty } = \bigl (c_{d,f_\mathrm{min},f_\mathrm{max},\rho } \frac{\log n}{\beta (n-1)} \bigr )^{1/d}\), and let \(\widehat{k}\) denote the smallest integer such that \(\min \{ {d}(X_j,M)\, |\, {d}(X_j,M)> h_{\infty }^2/\rho \} > h_{\widehat{k}}^2/\rho \). Since \(\mathbb {X}_n\) is a (random) finite set, we can always find such a random integer \(\widehat{k}\) that provides a sufficient number of iterations to obtain the asymptotic decluttering rate. For this random iteration \(\widehat{k}\), we can state the following result.
Proposition 5.8
Under the assumptions of Corollary 5.6, for every \(X_j\) \(\in \) \(X^{(\widehat{k} +1)}\), we have
As before, taking \(\mathbb {Y}_n^{+}\) as the result of the farthest point sampling algorithm based on \(\mathbb {X}^{(\widehat{k}+1)}\), and \(T^+\) the vector of tangent spaces \(\mathtt {TSE}(\mathbb {X}^{(\widehat{k}+1)},h_{\infty })_j\) such that \(\mathbb {X}^{(\widehat{k}+1)}_j \) \(\in \) \(\mathbb {Y}_n^+\), we can construct our last estimator.
Definition 5.9
With the above notation, define \( \widehat{M}_{\mathtt {TDC +}} = \mathrm {Del}^{\omega _*}(\mathbb {Y}^+_n,T^{+} ). \)
In turn, Proposition 5.8 implies that \(\widehat{M}_{\mathtt {TDC +}}\) satisfies Theorem 2.9.
6 Conclusion
In this work, we gave results on explicit manifold reconstruction with simplicial complexes. We built estimators \(\widehat{M}_{\mathtt {TDC}}\), \(\widehat{M}_{\mathtt {TDC \delta }}\) and \(\widehat{M}_{\mathtt {TDC+}}\) in two statistical models. We proved minimax rates of convergence for the Hausdorff distance and consistency results for ambient isotopic reconstruction. Since \(\widehat{M}_{\mathtt {TDC}}\) is minimax optimal in the additive noise model for \(\sigma \) small, and uses the Tangential Delaunay Complex of [4], the latter is proven to be optimal. Moreover, rates of [24] are proven to be achievable with simplicial complexes that are computable using existing algorithms. To prove the stability of the Tangential Delaunay Complex, a generic interpolation result was derived. In the process, a tangent space estimation procedure and a decluttering method both based on local PCA were studied.
In the model with outliers, the proposed reconstruction method achieves a rate of convergence that can be as close as desired to the minimax rate of convergence, depending on the number of iterations of the decluttering procedure. Though this procedure seems to be well adapted to our reconstuction scheme—which is based on tangent spaces estimation—we believe that it could be of interest in the context of other applications. Also, further investigation may be carried out to compare this decluttering procedure to existing ones [9, 19].
As briefly mentioned below Theorem 2.7, our approach is likely to be suboptimal in cases where noise level \(\sigma \) is large. In such cases, with additional structure on the noise such as centered and independent from the source, other statistical procedures such as deconvolution [24] could be adapted to provide vertices to the Tangential Delaunay Complex. Tangential properties of deconvolution are still to be studied.
The effective construction of \(\widehat{M}_{{\mathtt {TDC}}\delta }\) can be performed using existing algorithms. Namely, Tangential Delaunay Complex, farthest point sampling, local PCA and point-to-linear subspace distance computation for slab counting. A crude upper bound on the time complexity of a naive step-by-step implementation is
since the precision \(\delta \) requires no more than \(\log {(1/\delta )}\) iterations of the decluttering procedure. It is likely that better complexity bounds may be obtained using more refined algorithms, such as fast PCA [31], that lessens the dependence on the ambient dimension D. An interesting development would be to investigate a possible precision/complexity trade-off, as done in [3] for community detection in graphs for instance.
Even though Theorem 4.1 is applied to submanifold estimation, we believe it may be applied in various settings. Beyond its statement, the way that it is used is quite general. When intermediate objects (here, tangent spaces) are used in a procedure, this kind of proxy method can provide extensions of existing results to the case where these objects are only approximated.
As local PCA is performed throughout the paper, the knowledge of the bandwidth h is needed for actual implementation. In practice its choice is a difficult question and adaptive selection of h remains to be considered.
In the process, we derived rates of convergence for tangent space estimation. The optimality of the method will be the object of a future paper.
References
Alexander, S.B., Bishop, R.L.: Gauss equation and injectivity radii for subspaces in spaces of curvature bounded above. Geom. Dedicata 117, 65–84 (2006)
Arias-Castro, E., Lerman, G., Zhang, T.: Spectral clustering based on local PCA. J. Mach. Learn. Res. 18(9), 1–57 (2017)
Arias-Castro, E., Verzelen, N.: Community detection in dense random networks. Ann. Stat. 42(3), 940–969 (2014)
Boissonnat, J.-D., Ghosh, A.: Manifold reconstruction using tangential Delaunay complexes. Discrete Comput. Geom. 51(1), 221–267 (2014)
Boissonnat, J.-D., Guibas, L.J., Oudot, S.Y.: Manifold reconstruction in arbitrary dimensions using witness complexes. Discrete Comput. Geom. 42(1), 37–70 (2009)
Boucheron, S., Bousquet, O., Lugosi, G.: Theory of classification: a survey of some recent advances. ESAIM Probab. Stat. 9, 323–375 (2005)
Boucheron, S., Lugosi, G., Massart, P.: Concentration Inequalities: A Nonasymptotic Theory of Independence. Oxford University Press, Oxford (2013)
Bousquet, O.: A Bennett concentration inequality and its application to suprema of empirical processes. C. R. Math. Acad. Sci. Paris 334(6), 495–500 (2002)
Buchet, M., Dey, T.K., Wang, J., Wang, Y.: Declutter and resample: towards parameter free denoising. In: Aronov, B., Katz, M.J. (eds.) The 33rd International Symposium on Computational Geometry (SoCG’17). LIPIcs. Leibniz Int. Proc. Inform., vol. 77, Art. No. 23. Schloss Dagstuhl Leibniz-Zentrum für Informatik, Wadern (2017)
do Carmo, M.P.: Riemannian Geometry. Mathematics: Theory & Applications. Birkhäuser, Boston, MA (1992)
Chazal, F., Cohen-Steiner, D., Lieutier, A.: A sampling theory for compact sets in Euclidean space. In: Proceedings of the 22nd Annual Symposium on Computational Geometry (SCG’06), pp. 319–326. ACM, New York (2006)
Chazal, F., Glisse, M., Labruère, C., Michel, B.: Convergence rates for persistence diagram estimation in topological data analysis. J. Mach. Learn. Res. 16, 3603–3635 (2015)
Cheng, S.-W., Dey, T.K., Ramos, E.A.: Manifold reconstruction from point samples. In: Proceedings of the 16th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA’05), pp. 1018–1027. ACM, New York (2005)
Clarkson, K.L.: Building triangulations using \(\varepsilon \)-nets. In: Proceedings of the 38th Annual ACM Symposium on Theory of Computing (STOC’06), pp. 326–335. ACM, New York (2006)
Cuevas, A., Rodríguez-Casal, A.: On boundary estimation. Adv. Appl. Probab. 36(2), 340–354 (2004)
Davis, C., Kahan, W.M.: The rotation of eigenvectors by a perturbation. III. SIAM J. Numer. Anal. 7, 1–46 (1970)
De Marco, G., Gorni, G., Zampieri, G.: Global inversion of functions: an introduction. NoDEA Nonlinear Differ. Equ. Appl. 1(3), 229–248 (1994)
Dey, T.L.: Curve and surface reconstruction: algorithms with mathematical analysis. Cambridge Monographs on Applied and Computational Mathematics, vol. 23. Cambridge University Press, Cambridge (2007)
Donoho, D.L.: De-noising by soft-thresholding. IEEE Trans. Inform. Theory 41(3), 613–627 (1995)
Dümbgen, L., Walther, G.: Rates of convergence for random approximations of convex sets. Adv. Appl. Probab. 28(2), 384–393 (1996)
Dyer, R., Vegter, G., Wintraecken, M.: Riemannian simplices and triangulations. In: Arge, L., Pach, J. (eds.) The 31st International Symposium on Computational Geometry (SoCG’15). LIPIcs. Leibniz Int. Proc. Inform. (LIPIcs), vol. 34, pp. 255–269. Schloss Dagstuhl Leibniz-Zentrum für Informatik, Wadern (2015)
Federer, H.: Curvature measures. Trans. Am. Math. Soc. 93(3), 418–491 (1959)
Federer, H.: Geometric Measure Theory. Die Grundlehren der Mathematischen Wissenschaften, vol. 153. Springer, New York (1969)
Genovese, C.R., Perone-Pacifico, M., Verdinelli, I., Wasserman, L.: Manifold estimation and singular deconvolution under Hausdorff loss. Ann. Stat. 40(2), 941–963 (2012)
Genovese, C.R., Perone-Pacifico, M., Verdinelli, I., Wasserman, L.: Minimax manifold estimation. J. Mach. Learn. Res. 13, 1263–1291 (2012)
Kim, A.K.H., Zhou, H.H.: Tight minimax rates for manifold estimation under Hausdorff loss. Electron. J. Stat. 9(1), 1562–1582 (2015)
LeCam, L.: Convergence of estimates under dimensionality restrictions. Ann. Stat. 1(1), 38–53 (1973)
Maggioni, M., Minsker, S., Strawn, N.: Multiscale dictionary learning: non-asymptotic bounds and robustness. J. Mach. Learn. Res. 17, 1–51 (2016)
Mammen, E., Tsybakov, A.B.: Asymptotical minimax recovery of sets with smooth boundaries. Ann. Stat. 23(2), 502–524 (1995)
Niyogi, P., Smale, S., Weinberger, S.: Finding the homology of submanifolds with high confidence from random samples. Discrete Comput. Geom. 39(1–3), 419–441 (2008)
Sharma, A., Paliwal, K.K.: Fast principal component analysis using fixed-point algorithm. Pattern Recognit. Lett. 28(10), 1151–1155 (2007)
Acknowledgements
We would like to thank Jean-Daniel Boissonnat, Frédéric Chazal, Pascal Massart, and Steve Oudot for their insight and the interest they brought to this work. We are also grateful to the reviewers whose comments helped enhancing substantially this paper. This work was supported by ANR project TopData ANR-13-BS01-0008 and by the Advanced Grant of the European Research Council GUDHI (Geometric Understanding in Higher Dimensions). E. Aamari was supported by the Conseil régional d’Île-de-France under a doctoral allowance of its program Réseau de Recherche Doctoral en Mathématiques de l’Île-de-France (RDM-IdF).
Author information
Authors and Affiliations
Corresponding author
Additional information
Editor in Charge: Kenneth Clarkson
Appendices
Appendix
A Interpolation Theorem
This section is devoted to prove the interpolation results of Sect. 4.1. For the sake of completeness, let us state a stability result for the reach with respect to \(\mathscr {C}^2\)-diffeomorphisms.
Lemma 7.1
([22, Thm. 4.19]) Let \(A \subset \mathbb {R}^D\) with \(\mathrm {reach}(A) \ge \rho > 0\) and \( \Phi :\mathbb {R}^D \rightarrow \mathbb {R}^D\) be a \(\mathscr {C}^1\)-diffeomorphism such that \( \Phi \),\( \Phi ^{-1}\), and \({d}\Phi \) are Lipschitz with Lipschitz constants K,N and R respectively. Then
Writing \(\phi _\ell (\cdot ) = \phi (\cdot /\ell )\), we recall that \({\psi _j(a)} = (R_j-I_D)(a - \pi (p_j)) + (p_j - \pi (p_j)) \) and
Let us denote \(b_1 = \sup _x \left\| {d}_x \phi \right\| \), \(b_2 = \sup _x \Vert {{d}^2_x \phi }\Vert _\mathrm {op}\), and write \(C_1 = 1+b_1\), \(C_2 = b_2+2b_1\). Straightforward computation yields \(C_1 \le 7/2\) and \(C_2 \le 28\).
Proof of Lemma 4.2
First notice that the sum appearing in (1) consists of at most one term. Indeed, since \(\phi \equiv 0\) outside \(\mathscr {B}(0,1)\), if \(\phi _\ell (a-\pi (p_j) ) \ne 0\) for some \(j \in \{1,\ldots ,q\}\), then \(\Vert {a - \pi (p_j)}\Vert \le \ell \). Consequently, for all \(i \ne j\),
where we used that \( 6 \eta \le \ell \le \delta / 3\). Therefore, \(\phi _\ell (a-\pi (p_i) ) = 0\) for all \(i\ne j\). In other words, if a \(p_j\) actually appears in \(\Phi (a)\) then the others do not.
Global diffeomorphism: As the sum in (1) is at most composed of one term, chain rule yields
where the last line follows from \(b_1 \le 5/2\), \(6 \eta \le \ell \) and \(\theta \le \pi /64\). Therefore, \({d}_a \Phi \) is invertible for all \(a\in \mathbb {R}^D\), and \(({d}_a \Phi )^{-1} = \sum _{i=0}^\infty (I_D- {d}_a \Phi )^i\). \( \Phi \) is a local diffeomorphism according to the local inverse function theorem. Moreover, \(\left\| \Phi (a)\right\| \rightarrow \infty \) as \(\left\| a\right\| \rightarrow \infty \), so that \( \Phi \) is a global \(\mathscr {C^\infty }\)-diffeomorphism by the Hadamard–Cacciopoli theorem [17].
Differentials estimates: (i) First order: From the estimates above,
(ii) Inverse: Write for all \(a \in \mathbb {R}^D\):
where the first inequality holds since \(\left\| {d}_a \Phi - {{I_D}}\right\| _\mathrm {op} < 1\), and \(\left\| \cdot \right\| _\mathrm {op}\) is sub-multiplicative.
(iii) Second order: Again, since the sum (1) includes at most one term,
\(\square \)
Proof of Theorem 4.1
Set \(\ell = \delta /3\) and \(M' = \Phi (M)\).
-
Interpolation: For all j, \(p_j = \Phi (\pi (p_j)) \in M'\) by construction since \(\phi _\ell (0)=1\).
-
Tangent spaces: Since \({d}_x \phi _l|_{x=0} = 0\), for all \(j \in \{1,\ldots ,q \}\), \({d}_a \Phi |_{a=\pi (p_j)} = R_j\). Thus,
$$\begin{aligned} T_{p_j} {M}'&= T_{ \Phi (\pi (p_j))} \Phi (M) = {d}_a \Phi |_{a=\pi (p_j)} (T_{\pi (p_j)} M )= R_j (T_{\pi (p_j)} M ) = T_j, \end{aligned}$$by definition of \(R_j\).
-
Proximity to M: The bound on \(d_H(M,M') = d_H\bigl (M, \Phi (M) \bigr )\) follows from the correspondence
$$\begin{aligned} \left\| \Phi (a) - a\right\|&\le \sup _{a \in \mathbb {R}^D}\max _{1\le j\le q} \phi _\ell (a-\pi (p_j) ) \left\| \psi _j(a)\right\| \le \ell \theta + \eta \le \delta \theta + \eta . \end{aligned}$$ -
Isotopy: Consider the continuous family of maps
$$\begin{aligned} \Phi _{(t)}(a) = a + t \sum _{j = 1}^q \phi _\ell (a-\pi (p_j) ) {\psi _j(a)} , \end{aligned}$$for \(0 \le t \le 1\). Since \( \Phi _{(t)} - {{I_D}} = t ( \Phi - {{I_D}})\), the arguments above show that \( \Phi _{(t)}\) is a global diffeomorphism of \(\mathbb {R}^D\) for all \(t \in [0,1]\). Moreover, \( \Phi _{(0)} = {{I_D}}\) and \( \Phi _{(1)} = \Phi \). Thus, \(M = \Phi _{(0)}(M)\) and \(M' = \Phi _{(1)}(M)\) are ambient isotopic.
-
Reach lower bound: The differentials estimates of order 1 and 2 of \( \Phi \) translate into estimates on Lipschitz constants of \( \Phi , \Phi ^{-1}\) and \({d}\Phi \). Applying Lemma 7.1 leads to
$$\begin{aligned} \mathrm {reach}(M')&\ge \dfrac{( 1-C_1 ({\eta }/{\ell } + \theta ) )^2}{({1+ C_1 ({\eta }/{\ell } + \theta )})/{\rho }+ C_2 ({\eta }/{\ell ^2} +{\theta }/{\ell })}\\&= \rho \cdot \dfrac{( 1-C_1 ({\eta }/{\ell } + \theta ) )^2}{1+ C_1 ({\eta }/{\ell } + \theta )+ C_2 ({\eta }/{\ell ^2} +{\theta }/{\ell }) \rho }. \end{aligned}$$Now, replace \(\ell \) by its value \(\delta /3\), and write \(c_1 = 3 C_1 \le 21/2 \le 11\) and \(c_2 = 3^2 C_2 \le 252\). We derive
$$\begin{aligned} \mathrm {reach}(M')&\ge \biggl (1-2c_1 \biggl (\frac{\eta }{\delta } + \theta \biggr ) \biggr ) \biggl ( 1 - c_1 \biggl (\frac{\eta }{\delta } + \theta \biggr ) - c_2 \biggl (\frac{\eta }{\delta ^2} + \frac{\theta }{\delta } \biggr ) \rho \biggr ) \rho \\&\ge \biggl ( 1 - 3 c_1 \biggl (\frac{\eta }{\delta } + \theta \biggr ) - c_2 \biggl (\frac{\eta }{\delta ^2} + \frac{\theta }{\delta } \biggr ) \rho \biggr ) \rho \\&\ge \biggl ( 1 - (3 c_1 +c_2) \biggl (\frac{\eta }{\delta ^2} + \frac{\theta }{\delta } \biggr ) \rho \biggr ) \rho , \end{aligned}$$where for the last line we used that \(\delta / \rho \le 1\). The desired lower bound follows taking \(c_0 = 3c_1+ c_2 \le 285\).
\(\square \)
B Some Geometric Properties under Reach Regularity Condition
1.1 B.1 Reach and Projection on the Submanifold
In this section we state intermediate results that connect the reach condition to orthogonal projections onto the tangent spaces. They are based on the following fundamental result.
Proposition 8.1
([22, Thm. 4.18]) For all x and y in M,
where \((y-x)_{\perp }\) denotes the projection of \(y-x\) onto \(T_x M^{\perp }\).
From Proposition 8.1 we may deduce the following property about trace of Euclidean balls on M.
Proposition 8.2
Let \(x \in \mathbb {R}^D\) be such that \({d}(x,M) = \Delta \le h \le {\rho }/{8}\), and let y denote \(\pi (x)\). Then,
where \(r_h^2 + \Delta ^2 = h^2\), \((r_h^-)^2 = \bigl (1-\frac{ \Delta }{\rho }\bigr )r_h^2\), and \((r_h^+)^2 = \bigl (1+\frac{ 2 \Delta }{\rho }\bigr )r_h^2\).
Proof
Let z be in \(M \cap \mathscr {B}(x,h)\), and denote by \(\delta \) the quantity \(\Vert z-y\Vert \). We may write
hence \(\delta ^2 \le h^2 - \Delta ^2 - 2 \langle z-y,y-x \rangle \). Denote, for u in \(\mathbb {R}^D\), by \(u_\perp \) its projection onto \(T_y M^{\perp }\). Since \(\langle z-y,y-x \rangle = \langle (z-y)_{\perp },y-x \rangle \), Proposition 8.1 ensures that
Since \(\Delta \le h \le \rho /8\), it comes \(\delta ^2 \le (1+2 {\Delta }/{\rho })r_h^2\). On the other hand, (2) and Proposition 8.1 also yield
Hence, if \(\delta ^2 \le \big (1- \frac{\Delta }{\rho } \big )r_h^2\), we have
\(\square \)
Also, the following consequence of Proposition 8.1 will be of particular use in the decluttering procedure.
Proposition 8.3
Let h and \(h_k\) be bandwidths satisfying \(h_k^2/\rho \le h \le h_k\). Let x be such that \(d(x,M) \le h/\sqrt{2}\) and \(\pi _{M} (x) = 0\), and let z be such that \(\Vert z-x\Vert \le h\) and \(d(z,M) \le h_k^2/\rho \). Then
where \(z_{\perp }\) denotes the projection of z onto \(T_0 M^{\perp }\).
Proof
Let y denote \(\pi _M (z)\). The triangle inequality yields \(\Vert y\Vert \le \Vert y-z \Vert + \Vert z-x\Vert + \Vert x\Vert \le h_k^{2}/\rho + (1 + 1/\sqrt{2}) h \le 3 h_k\). Proposition 8.1 ensures that \(\Vert y_{\perp } \Vert \le \Vert y \Vert ^2/(2\rho ) \le (9 h_k^2)/(2\rho )\). Since \(\Vert z_{\perp } \Vert \le \Vert y_{\perp }\Vert + h_k^2/\rho \), we have \(\Vert z_{\perp } \Vert \le 6 h_k^2/\rho \) . \(\square \)
At last, let us prove Lemma 5.4, which gives properties of intersections of ambient slabs with M.
Proof of Lemma 5.4
Set \(k_1=\frac{1}{16(K \vee 1)}\), \(k_2 = \frac{1}{16(K \vee \rho \vee 1)}\), and \(k_3 = k_1 \wedge \frac{\rho k_2}{1+2K} \wedge 1\). For all \(h> 0\), and \(z \in S(x,T,h)\), the triangle inequality yields \(\Vert z-x\Vert \le \Vert \pi _T(z-x)\Vert + \Vert \pi _{T^{\perp }}(z-x)\Vert \le (k_1 + k_2h)h\). Since \(h \le 1\) and \(k_1 + k_2 \le 1/2\), we get \(z \in \mathscr {B}(x,h/2)\).
Now, suppose that \( h/\sqrt{2} \ge {d}(x,M) \ge h^2 / \rho \) and \(\angle (T_{\pi (x)} M, T ) \le K h / \rho \). For short we write \(T_0 = T_{\pi (x)} M\). Let \(z \in S(x,T,h)\), since \(h \le 1\), it comes
with \(k'_1 = k_1 + k_2\). On the other hand
with \(k'_2 = k_1 K/\rho + k_2\). Hence \(S(x,T,h) \subset S'(x,T_0,h)\), for the constants \(k'_1\) and \(k'_2\) defined above. It remains to prove that \(S'(x,T_0,h) \cap M = \emptyset \). To see this, let \(z \in S'(x,T_0,h)\), and \(y = \pi (x)\). Since \(k'_1 + k'_2 \le 1/4\), we have \(\Vert y-z \Vert \le \Vert y-x \Vert + \Vert x-z \Vert \le h(1/\sqrt{2} + 1/4)\). For the normal part, we may write
Since \(k'_2 \le 1/(8\rho )\), we have \(\Vert \pi _{T_0^{\perp }}(z-y) \Vert > \Vert y-z\Vert ^2 /(2\rho )\), hence Proposition 8.1 ensures that \(z \notin M\).
At last, suppose that \(x \in M\) and \(y \in \mathscr {B}(x,k_3h)\cap M\). Since \(k_3 \le k_1\), we have \(\Vert \pi _{T}(y-x)\Vert \le k_1h\). Next, we may write
Since \(y \in M\), Proposition 8.1 entails \(\Vert \pi _{T_0^{\perp }}(y-x) \Vert \le \Vert y-x\Vert ^2/(2\rho ) \le k_3^2 h^2/(2\rho )\). It comes
Hence \(y \in S(x,T,h)\). \(\square \)
1.2 B.2 Reach and Exponential Map
In this section we state results that connect Euclidean and geodesic quantities under reach regularity condition. We start with a result linking reach and principal curvatures.
Proposition 8.4
([30, Prop. 6.1]) For all \(x\in M\), writing \(\textit{II}_x\) for the second fundamental form of M at x, for all unitary \(w\in T_x {M}\), we have \(\left\| \textit{II}_x(w,w)\right\| \le 1/\rho \).
For all \(x \in {M}\) and \(v\in T_x {M}\), let us denote by \(\exp _x(v)\) the exponential map at x of direction v. According to the following proposition, this exponential map turns out to be a diffeomorphism on balls of radius at most \(\pi \rho \).
Proposition 8.5
([1, Corr. 1.4]) The injectivity radius of M is at least \(\pi \rho \).
Denoting by \({d}_M(\cdot ,\cdot )\) the geodesic distance on M, we are in position to connect geodesic and Euclidean distance. In what follows, we fix the constant \(\alpha = 1 + \frac{1}{4 \sqrt{2}}\).
Proposition 8.6
For all \(x,y \in {M}\) such that \(\left\| x-y\right\| \le \rho /4\),
Moreover, writing \(y = \exp _x(r v)\) for \(v\in T_x {M}\) with \(\left\| v\right\| =1\) and \(r \le \rho /4\),
with \(\left\| R(r,v)\right\| \le {r^2}/({2\rho })\).
Proof
The first statement is a direct consequence of Proposition 6.3 in [30]. Let us define \(u(t) =\exp _x(tv) - \exp _x(0) -tv\) and \(w(t) = \exp _x(tv)\) for all \(0\le t \le r\). It is clear that \(u(0) = 0\) and \(u'(0) = 0\). Moreover, \(\Vert {u''(t)}\Vert = \Vert {\textit{II}_{w(t)}(w'(t),w'(t))}\Vert \le 1/\rho \). Therefore, a Taylor expansion at order two gives \(\left\| R(r,v)\right\| = u(r) \le r^2/(2\rho )\). Applying the first statement of the proposition gives \(r\le \alpha \left\| x-y\right\| \). \(\square \)
The next proposition gives bounds on the volume form expressed in polar coordinates in a neighborhood of points of M.
Proposition 8.7
Let \(x \in {M}\) be fixed. Denote by J(r, v) the Jacobian of the volume form expressed in polar coordinates around x, for \(r \le {\rho }/{4}\) and v a unit vector in \(T_x {M}\). In other words, if \(y = \exp _x(rv)\), \(d_yV =J(r,v)\,\, dr dv.\) Then
where \(c_d = 2^{-d}\) and \(C_d=2^d\). As a consequence, if \(\mathscr {B}_M(x,r)\) denotes the geodesic ball of radius r centered at x, then, if \(r \le {\rho }/{4}\),
with \(c'_d = c_d V_d\) and \(C'_d = C_d V_d\), where \(V_d\) denotes the volume of the unit d-dimensional Euclidean ball.
Proof
Denoting \(A_{r,v} = {d}_{rv} \exp _x\), the Area Formula [23, Sect. 3.2.5] asserts that \(J(r,v) = r^{d-1} \sqrt{\det ( A_{r,v}^t A_{r,v} )}\). Note that from Proposition 6.1 in [30] together with the Gauss equation [10, p. 130], the sectional curvatures in M are bounded by \(|\kappa | \le 2/\rho ^2\). Therefore, the Rauch theorem [21, Lemma 5] states that
for all \(w \in T_x {M}\). As a consequence,
Since \(\mathrm{Vol}(\mathscr {B}_M(x,r))= \int _{s=0}^{r} \int _{v \in {\mathscr {S}}_{d-1}} J(s,v)\, ds dv\), where \({\mathscr {S}}_{d-1}\) denotes the unit \((d-1)\)-dimensional sphere, the bounds on the volume easily follow. \(\square \)
C Some Technical Properties of the Statistical Model
1.1 C.1 Covering and Mass
Lemma 9.1
Let \(Q_0 \in \mathscr {U}_M(f_\mathrm{min},f_\mathrm{max})\). Then for all \(p \in {M}\) and \( r \le \rho /4\),
where \(a_d > 0\). As a consequence, for n large enough and for all \(Q \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\), with probability larger than \(1-({1}/{n})^{2/d}\),
Similarly, for n large enough and for all \(P \in \mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta }\), with probability larger than \(1-({1}/{n})^{2/d}\),
Proof
The first statement is a direct corollary of Proposition 8.7, since for all \(r \le \rho /4\),
where \(a_d\) can be taken to be equal to \(c'_d\) of Proposition 8.7. Let us now prove the second statement. By definition, sample \(X_i \in \mathbb {X}_n\), that has distribution \(Q \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\), can be written as \(X_i = Y_i + Z_i\), with \(Y_i\) having distribution \(Q_0 \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\), and \(\left\| Z_i\right\| \le \sigma \). From the previous point, letting \(a = a_d f_\mathrm{min}\), \(Q_0\) fulfils the so-called (a, d)-standard assumption of [12] for \(r\le \rho /4\). Looking carefully at the proof of Lemma 10 in [12] shows that its conclusion still holds for measures satisfying the (a, d)-standard assumption for small radii only. Therefore, writing \(\mathbb {Y}_n = \{Y_1,\ldots ,Y_n\}\), for \(r \le \rho /8\) we obtain
The claim follows using that \(d_H(\mathbb {X}_n,\mathbb {Y}_n) \le \sigma \), and setting \(r = C_{d,f_\mathrm{min}} ( {\log n}/{n})^{1/d}\) with \(C_{d,f_\mathrm{min}}^d \frac{a}{2^{d+1}} \ge 1+ 2/d\).
To prove the last point, notice that for all \(k=0,\ldots ,n\), conditionally on the event \(\lbrace |\mathbb {X}_n \cap {M}|=k\rbrace \), \(\mathbb {X}_n \cap {M}\) has the distribution of a k-sample of \(Q_0\). Therefore,
Hence,
whenever \(r\le \rho /8\) and \(a r^d \le 2^d\). Taking \(r = C_{d,f_\mathrm{min}}' \bigl ( \frac{\log n}{\beta n}\bigr )^{1/d}\) with \(C_{d,f_\mathrm{min}}'^d \frac{\beta a}{2^{d+1}} \ge 1+ 2/d\) yields the result. \(\square \)
We now focus on proving Lemma 2.2. For its proof, we need the following piece of notation. For all bounded subsets \(K \subset \mathbb {R}^D\) and \(\varepsilon >0\), we let \(\mathrm {cv}_K(\varepsilon )\) denote the Euclidean covering number of K. That is, \(\mathrm {cv}_K(\varepsilon )\) is the minimal number k of Euclidean open balls of radii \(\varepsilon \) and centered at elements of K that are needed to cover K.
Lemma 9.2
Let \(K \subset \mathbb {R}^D\) be a bounded subset. If K is path connected, then for all \(\varepsilon >0\), \(\mathrm {diam}(K) \le 2\varepsilon \mathrm {cv}_K(\varepsilon )\).
Proof
Let \(p,q \in K\) and \(\gamma :[0,1] \rightarrow K\) be a continuous path joining \(\gamma (0) = p\) and \(\gamma (1) = q\). Writing \(N = \mathrm {cv}_K(\varepsilon )\), let \(x_1,\ldots ,x_N \in \mathbb {R}^D\) be the centers of a covering of K by open balls of radii \(\varepsilon \). We let \(U_i\) denote \(\{ t , \left\| \gamma (t) - x_{i}\right\| < \varepsilon \} \subset [0,1]\). By construction of the covering, there exists \(x_{(1)} \in \{x_1,\ldots ,x_N\}\) such that \(\Vert {p - x_{(1)}}\Vert < \varepsilon \). Then \(U_{(1)} \ni \gamma (0) = p\) is a non-empty open subset of [0, 1], so that \(t_{(1)} = \sup U_{(1)}\) is positive. If \(t_{(1)} =1\), then \(\Vert {q-x_{(1)}}\Vert \le \varepsilon \), and in particular \(\left\| q-p\right\| \le 2 \varepsilon \). If \(t_{(1)} < 1\), since \(U_{(1)}\) is an open subset of [0, 1], we see that \(\gamma (t_{(1)}) \notin U_{(1)}\). But \(\bigcup _{i = 1}^N U_i\) is an open cover of [0, 1], which yields the existence of \(U_{(2)}\) such that \(\gamma (t_{(1)}) \in U_{(2)}\), and for all \(t < t_{(1)}\), \(\gamma (t) \notin U_{(2)}\). Then consider \(t_{(2)} = \sup U_{(2)}\), and so on. Doing so, we build by induction a sequence of numbers \(0< t_{(1)}< \cdots < t_{(k)} \le 1\) and distinct centers \(x_{(1)},\ldots ,x_{(k)} \in \{x_1,\ldots ,x_N\}\) (\(k \le N\)) such that \(\Vert {p - x_{(1)}}\Vert < \varepsilon \), \(\Vert {q - x_{(k)}}\Vert \le \varepsilon \), with \(\Vert {\gamma (t_{(i)}) - x_{(i)}}\Vert \le \varepsilon \) for \(1 \le i \le k\) and \(\Vert {\gamma (t_{(i)}) - x_{(i+1)}}\Vert < \varepsilon \) for \(1 \le i \le k-1\). In particular, \(\Vert {x_{(i)} - x_{(i+1)}}\Vert \le 2\varepsilon \) for all \(1\le i \le k-1\). To conclude, write
Since this bound holds for all \(p,q \in K\), we get the announced bound on the diameter of K. \(\square \)
We are now in position to prove Lemma 2.2.
Proof of Lemma 2.2
Let \(\varepsilon \le {\rho }/{4}\), and \(x_1,\ldots ,x_{\mathrm {cv}_M(\varepsilon )}\) be a minimal covering of M. According to Lemma 9.1, for all k,
for some \(a_d > 0\). A straightforward packing argument [12, Sect. B.1] yields that the covering number of the support M of Q satisfies
for all \(\varepsilon \le {\rho }/{4}\), where \(c_d = 2^d/a_d\). Applying this bound with \(\varepsilon = \rho /4\), together with Lemma 9.2 yields
where \(C_d = 2^{3d-1}/a_d\). \(\square \)
Now we allow for some outliers. We consider a random variable X with distribution P, that can be written as \(X = V(Y+Z) + (1-V)X''\), with \(\Vert Z\Vert \le s h\), \(s \le 1/4\), such that \(\mathbb {P}(V=1) = \beta \) and V is independent from \((Y,Z,X'')\), Y has law Q in \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho }\), and \(X''\) has uniform distribution on \(\mathscr {B}(0,K_0)\) (recall that \(K_0\) is defined below Lemma 2.2). Note that \(s=0\) corresponds to the clutter noise case, whereas \(\beta =1\) corresponds to the additive noise case.
For a fixed point x, let p(x, h) denote \(P(\mathscr {B}(x,h))\). We have \(\mathbb {P}(VY \in \mathscr {B}(x,(1-s)h) ) \le \mathbb {P}(VX \in \mathscr {B}(x,h))\le \mathbb {P}(VY \in \mathscr {B}(x,2h))\). Hence we may write
where \(q(x,h) = Q(\mathscr {B}(x,h))\), and \(q'(x,h)\) = \((h/{K_0})^D\). Bounds on the quantities above are to be found in the following lemma.
Lemma 9.3
There exists \(h_+(\rho ,\beta ,f_\mathrm{min},f_\mathrm{max},d) \le \rho / \sqrt{12d}\) such that, if \(h \le h_+\), for every x such that \({d}(x,M) \le h\), we have
-
\(\mathscr {B}(x,2h) \cap M \subset \mathscr {B}(\pi _{{M}}(x),4h) \cap M,\)
-
\( q(x,2h) \le C_{d}{ f_\mathrm{max}} h^{d}\).
Moreover, if \({d}(x,M) \le h/\sqrt{2}\), we have
-
\(\mathscr {B}(\pi _{{M}}(x),h/8) \cap M \subset \mathscr {B}(x,3h/4)\),
-
\(c_{d}{ f_\mathrm{min}} h^d \le q(x,3h/4)\),
-
\( p(x,h) \le 2 \beta q(x,2h)\).
Proof
Set \(h_1(\rho ) = \rho /(16 \alpha )\), and let x be such that \({d}(x,{M}) \le h\), and \(h \le h_1\). According to Proposition 8.2, \(\mathscr {B}(x,2h) \cap {M} \subset \mathscr {B}(\pi _{{M}}(x), r_{2h}^+) \cap M\), with \(r_{2h}^+ = \sqrt{(1 + 2 \Delta / \rho )} r_{2h} \le 2 r_{2h} \le 4h\). According to Proposition 8.6, if y \(\in \) \(\mathscr {B}(\pi _M(x),4h) \cap M\), then \(d_M(\pi _M(x),y) \le 4\alpha h \le \rho /4\). Proposition 8.7 then yields \(q(x,2h) \le C_d f_\mathrm{max}h^d\).
Now if \({d}(x,M) \le h/\sqrt{2}\), \(\mathscr {B}(\pi _{{M}}(x),r_{3h/4}^-) \cap M \subset \mathscr {B}(x,3h/4) \cap M\) from Proposition 8.2, with \(r_{3h/4}^- = \sqrt{(1 - \Delta / \rho )}r_{3h/4} \ge r_{3h/4}/2 \ge h/8\). Since we have \(\mathscr {B}_M(\pi _M(x),h/8) \subset \mathscr {B}(\pi _M(x),h/8)\cap M\), a direct application of Proposition 8.7 entails \(c_{d}{ f_\mathrm{min}} h^d \le q(x,3h/4)\).
Applying Proposition 8.7 again, there exists \(h_2(f_\mathrm{min},d,D,\beta ,\rho )\) such that if \(h \le h_1 \wedge h_2\), then for any x such that \({d}(x,{M}) \le h/\sqrt{2}\) we have \((1-\beta ) q'(x,h) \le \beta c_{d,f_\mathrm{min}} h^{d} \), along with \(q(x,2h) \ge q(x,3h/4) \ge c_{d,f_\mathrm{min}} h^{d}\). We deduce that \( p(x,h) \le 2 \beta q(x,2h)\). Taking \(h_+ = h_1 \wedge h_2\wedge \rho / \sqrt{12d}\) leads to the result. \(\square \)
1.2 C.2 Local Covariance Matrices
In this section we describe the shape of the local covariance matrices involved in tangent space estimation. Without loss of generality, the analysis will be conducted for \(\widehat{\Sigma }_1\) (at sample point \(X_1\)), abbreviated as \(\widehat{\Sigma }\). We further assume that \(d(X_1,M) \le h/\sqrt{2}\), \(\pi _{{M}}(X_1) = 0\), and that \(T_{0}{M}\) is spanned by the d first vectors of the canonical basis of \(\mathbb {R}^D\).
The two models (additive noise and clutter noise) will be treated jointly, by considering a random variable X of the form
where \(\mathbb {P}(V=1) = \beta \) and V is independent from \((Y,Z,X'')\), Y has distribution in \(\mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\), \(\Vert Z\Vert \le \sigma \), and \(X''\) has uniform law on \(\mathscr {B}(0,K_0)\) (recall that \(K_0\) is defined above Definition 2.4). For short we denote by s the quantity \(\sigma /h\), and recall that we take \(s \le 1/4\), along with \(h \le h_+\) (defined in Lemma 9.3).
Let \(U(X_i,h)\), \(i=2, \ldots , n\), denote \( \mathbb {1}_{\mathscr {B} (X_1,h )}(X_i)\), let \(Y_i \in M\) and \(Z_i\) such that \(X_i = Y_i + Z_i\), with \(\Vert Z_i\Vert \le s h\), and let \(V_2, \ldots V_n\) denote random variables such that \(V_i= 1\) if \(X_i\) is drawn from the signal distribution (see Sect. 2.1). It is immediate that the \((U(X_i,h),V_i)\)’s are independent and identically distributed, with distribution (U(X, h), V).
With a slight abuse of notation, we will denote by \(\mathbb {P}\) and \(\mathbb {E}\) conditional probability and expectation with respect to \(X_1\). The following expectations will be of particular interest.
where for any x in \(\mathbb {R}^D\) \(x_{\top }\) and \(x_{\perp }\) denote respectively the projection of x onto \(T_0 M\) and \(T_0M^{\perp }\).
The following lemma gives useful results on both m(h) and \(\Sigma (h)\), provided that \(X_1\) is close enough to M.
Lemma 9.4
If \(d(X_1,M) \le h/\sqrt{2}\), for \(h \le h_+\), then
with
where \(\mu _{min}(A(h))\) denotes the smallest eigenvalue of A(h). Furthermore,
Proof
Let \(x=y+z\) be in \(B(X_1,h)\), with \(y \in M\) and \(\Vert z\Vert \le s h\). Since \(s \le 1/4\), \(\Vert y\Vert \le 2h\). According to Proposition 8.2 combined with Proposition 8.6, we may write, for \(h \le h_+\) and y in \(\mathscr {B}(X_1 ,2h) \cap {M}\),
in local polar coordinates. Moreover, if \(y \in \mathscr {B}(X_1,(1-s)h)\), then \(x \in \mathscr {B}(X_1,h)\). Then, according to Proposition 8.2, we have \(\mathscr {B}(\pi _{{M}}(X_1),r_{3h/4}^-) \cap M \subset \mathscr {B}(X_1,(1-s)h) \cap M\). Let u be a unit vector in \(T_0 {M}\). Then \(\langle u,x-m_{\top }(h) \rangle ^2\) \(=\) \(\langle u,rv + R(r,v)+z-m_{\top }(h) \rangle ^2\) \(\ge \) \(\langle u,rv -m_{\top }(h) \rangle ^2 /2 - 3 (R(r,v)+z)^2\) \(\ge \) \(\langle u,rv -m_{\top }(h) \rangle ^2 /2-6 r^4/(4\rho ^2) - 6 s^2 h^2\) according to Proposition 8.6. Hence we may write
according to Proposition 8.7 (bound on J(r, v)) and Proposition 8.2 (the geodesic ball \(\mathscr {B}_M(\pi _M(X_1),r_{3h/4}^-)\) is included in the Euclidean ball \(\mathscr {B}(\pi _M(X_1),r_{3h/4}^-)\) \(\subset \) \(\mathscr {B}(X_1,(1-s)h) \cap {M}\)). Then
where \(\sigma _{d-1}\) denotes the surface of the \((d-1)\)-dimensional unit sphere. On the other hand,
Since \(r_{3h/4}^- \le h \le h_+ \le \rho /\sqrt{12d}\), we conclude that
since, for \(d(X_1,M) \le h/\sqrt{2}\) and \(h \le h_+\), \(r_{3h/4}^- \ge r_{3h/4}/2 \ge h/8\), according to Proposition 8.2.
Now, since for any \(x=y+z \in \mathscr {B}(X_1,h)\), y \(\in \) \(M \cap \mathscr {B}(0,2h)\) and \(\Vert z\Vert \le s h\), we have \(\Vert y_{\perp }\Vert \le 2h^2/\rho \), according to Proposition 8.1. Jensen’s inequality yields that \(\Vert m(h)_{\perp }\Vert \le 2h^2/\rho + s h\) and \(\Vert m(h)_{\top }\Vert \le \Vert m(h)\Vert \le 2h\). \(\square \)
The following Lemma 9.5 is devoted to quantifying the deviations of empirical quantities such as local covariance matrices, means and number of points within balls from their deterministic counterparts. To this aim we define \(N_0(h)\) and \(N_1(h)\) as the number of points drawn from respectively noise and signal in \(\mathscr {B}(X_1,h) \cap M\), namely
Lemma 9.5
Recall that \(h_0 = \bigl ( \kappa \frac{\log n}{\beta (n-1)} \bigr )^{1/(d+1)}\) (as defined in Sect. 5.2), and \(h_{\infty } = h_0^{(d+1)/d}\), for \(\kappa \) to be fixed later.
If \(h_0 \le h_+\) and \(d(X_1, M) \le h_+ /\sqrt{2}\), then, with probability larger than \(1- 4 ({1}/{n})^{2/d +1}\), the following inequalities hold, for all \(h \le h_0\).
Moreover, for all \((h_{\infty } \vee \sqrt{2} {d}(X_1,M)) \le h \le h_0\), and n large enough,
Proof
The first two inequalities are straightforward applications of [6, Thm. 5.1]. The proofs of the two last results are detailed below. They are based on Talagrand–Bousquet’s inequality (see, e.g., [8, Thm. 2.3]) combined with the so-called peeling device.
Define \(h_-=(h_{\infty } \vee \sqrt{2} {d}(X_1,M))\), where we recall that in this analysis \(X_1\) is fixed, and let \(f_{T,h}\) denote the function
for \(h_- \le h \le h_0\), T a \(d \times d\) matrix such that \(\Vert T\Vert _{\mathscr {F}}=1\), x in \(\mathbb {R}^D\), v in \(\{0,1\}\), and \(\langle T, B \rangle = \mathrm {trace}(T^t A)\), for any square matrices T and A. Now we define the weighted empirical process
with \(r(h) = \beta q(2h) h^2\), along with the constrained empirical processes
for \(h_- \le u \le h_0\). Since \(\Vert f_{T,h} \Vert _{\infty } \le \sup _{x \in M} \Vert x-m(h)\Vert ^2 U(x,h) \le 4h^2\), and
for \(s \le 1/4\), a direct application of Theorem 2.3 in [8] yields, with probability larger than \(1-e^{-x}\),
To get a bound on \(\mathbb {E} Z(u)\), we introduce some independent Rademacher random variables \(\sigma _2, \ldots , \sigma _n\), i.e., \(\mathbb {P}(\sigma _j =1) = \mathbb {P}(\sigma _j=-1)=1/2\). With a slight abuse of notation, expectations with respect to the \((X_i,V_i)\)’s and \(\sigma _i\)’s, \(i=2, \ldots , n\), will be denoted by \(\mathbb {E}_{(X,V)}\) and \(\mathbb {E}_{\sigma }\) in what follows. According to the symmetrization principle (see, e.g., [7, Lem. 11.4]), we have
For a fixed sequence \((X_i,V_i)\), \(i=2, \ldots , n\), we may write
Jensen’s inequality ensures that
hence
For the term \(E_{11}\), note that, when \((X_i,V_i)_{i=2, \ldots ,n}\) is fixed, \(\sup _{h \le u} \bigl ( \bigl \Vert \sum _{i \ge 2} \sigma _i V_i U(X_i,h) X_i X_i^t \bigr \Vert _{\mathscr {F}} - \mathbb {E}_{\sigma } \bigl \Vert \sum _{i \ge 2} \sigma _i V_i U(X_i,h) X_i X_i^t \bigr \Vert _{\mathscr {F}} \bigr )\) is in fact a supremum of at most \(N_1(u)\) processes. According to the bounded difference inequality (see, e.g., [7, Thm. 6.2]), each of these processes is subGaussian with variance bounded by \(16h^4N_1(u)\) (see Theorem 2.1 of [7]). Hence a maximal inequality for sub-Gaussian random variables (see Sect. 2.5, p. 31, of [7]) ensures that
Hence \(\mathbb {E}_{(X,V)} E_{11} \le 4h^2\sqrt{2 \beta (n-1) q(2u) \log (n-1)}\). \(E_2\) may also be decomposed as
Jensen’s inequality yields that \(E_{22} \le 2u \sqrt{N_1(u)}\), and the same argument as for \(E_{11}\) (expectation of a supremum of \(n-1\) sub-Gaussian processes with variance bounded by \(4u^2N_1(u)\)) gives \(E_{22} \le 2u \sqrt{2 N_1(u) \log (n-1)}\). Hence
Similarly, we may write
At last, we may decompose \(E_4\) as
using the same argument. Combining all these terms leads to
hence we get
To derive a bound on the weighted process Z, we make use of the so-called peeling device (see, e.g., [7, Sect. 13.7, p. 387]). Set \(p = \lceil \log (h_0/h_\infty ) \rceil \le 1 + \log (h_0/h_\infty )\), so that \(e^{-{p}}h_0 \le h_-\). According to Lemma 9.3, if \(I_j\) denotes the slice \([e^{-j}h_0, e^{-(j-1)}h_0] \cap [h_-,h_0]\), then, for every h in \(I_j\), we have
where \(c_d\) depends only on the dimension, provided that \(h_0 \le h_+\). Now we may write
Since \(q(2h_{j-1}) \ge q(2h_-)\), we deduce that
Now, according to Lemma 9.3, \(\beta q(2h_-) \ge c_d \kappa \log n /(n-1)\). On the other hand, \(p \le 1 + \log (h_0/h_\infty ) \le \log (\beta (n-1)/\kappa )/d \le \log n /d\), for \(\kappa \ge 1\). For n large enough, taking \(x = ( 1 + 2/d ) \log n\) in the previous inequality, we get
The last concentration inequality of Lemma 9.5 may be derived the same way, considering the functions
where T is an element of \(\mathbb {R}^d\) satisfying \(\Vert T\Vert \le 1\). \(\square \)
1.3 C.3 Decluttering Rate
In this section we prove that, if the angle between tangent spaces is of order h, then we can distinguish between outliers and signal at order \(h^2\). We recall that the slab S(x, T, h) is the set of points y such that \( \Vert \pi _{T}(y-x) \Vert \le k_1 h\) and \(\Vert \pi _{T^{\perp }}(y-x)\Vert \le k_2h^2\), \(k_1\) and \(k_2\) defined in Lemma 5.4, and where \(\pi _{T}\) denotes the orthogonal projection onto T.
Lemma 9.6
Recall that \(h_0 = \bigl ( \kappa \frac{\log n}{\beta (n-1)} \bigr )^{1/(d+1)}\), and \(h_{\infty } = h_0^{(d+1)/d}\). Let K be fixed, and \(k_1\), \(k_2\) defined accordingly as in Lemma 5.4. If \(h_0 \le h_+\), for \(\kappa \) large enough (depending on d, \(\rho \) and \(f_\mathrm{min}\)) and n large enough, there exists a threshold t such that, for all \(h_{\infty } \le h \le h_0\), we have, with probability larger than \(1-3({1}/{n})^{2/d +1}\),
Proof
Suppose that \(d(X_1,M) \ge h/\sqrt{2} \). Then, according to Lemma 5.4, \(S(X_1,T,h) \subset \mathscr {B}(X_1,h/2)\), with \(\mathscr {B}(X_1, h/2) \cap M = \emptyset \), hence \(P_n(S(X_1,T,h)) \le P_n(\mathscr {B}(X_1,h/2))\). Theorem 5.1 in [6] yields that, for all \(h_{\infty } \le h \le h_0\), with probability larger than \(1-({1}/{n})^{2/d +1}\),
Since \(\log (n)/(n-1) \le \beta h^d/\kappa \), we may write
for n large enough so that \(h \le 1\).
If \(h/ \sqrt{2} \ge d(X_1,M) \ge h^2/\rho \) and \(\angle (T_{\pi (X_1)} M, T ) \le K h / \rho \), then Lemma 5.4 provides a big slab \(S'(x,T_{\pi (x)}M,h)\) so that \(S(x,T,h) \subset S'(x,T_{\pi (x)}M,h)\) and \(S'(x,T_{\pi (x)}M,h) \cap M = \emptyset \). Thus, \(P_n(S(x,T,h)) \le P_n(S'(x,T_{\pi (x)}M,h))\). Another application of Theorem 5.1 in [6] yields that, for all \(h_{\infty } \le h \le h_0\), with probability larger than \(1-({1}/{n})^{2/d +1}\),
hence, denoting by \(\omega _{r}\) the volume of the r-dimensional unit ball, we get
when n is large enough.
Now, if \(X_1 \in M\) and \(\angle (T_{\pi (X_1)} M, T ) \le K h / \rho \), Lemma 5.4 entails that \(\mathscr {B}(X_1,k_3h) \cap M \subset S(X_1,T,h)\), hence \(P_n(S(X_1,T,h)) \ge P_n(\mathscr {B}(X_1,k_3h) \cap M)\). The last application of Theorem 5.1 in [6] yields that, for all \(h_{\infty } \le h \le h_0\), with probability larger than \(1-({1}/{n})^{2/d +1}\),
Thus we deduce that
according to Lemma 9.3 (since \(k_3 \le 1\)). Choosing \(\kappa \) large enough (depending on d, \(\rho \) and \(f_\mathrm{min}\)) and then n large enough leads to the result. \(\square \)
D Matrix Decomposition and Principal Angles
In this section we expose a standard matrix perturbation result, adapted to our framework. For real symmetric matrices, we let \(\mu _i(\cdot )\) denote their i-th largest eigenvalue and \(\mu _\mathrm{min}(\cdot )\) the smallest one.
Theorem 10.1
(Sin \(\theta \) theorem [16], this version from Lemma 19 in [2]) Let \(O \in \mathbb {R}^{D\times D}\), \(B\in \mathbb {R}^{d\times d}\) be positive semi-definite symmetric matrices such that
Let \(T_0\) (resp. T) be the vector space spanned by the first d vectors of the canonical basis (resp. by the first d eigenvectors of O). Then
E Local PCA for Tangent Space Estimation and Decluttering
This section is dedicated to the proofs of Sect. 5. We begin with the case of additive noise (and no outliers), that is Proposition 5.1.
1.1 E.1 Proof of Proposition 5.1
Without loss of generality, the local PCA analysis will be conducted at base point \(X_1\), the results on the whole sample then follow from a standard union bound. For convenience, we assume that \(\pi _M(X_1) = 0\) and that \(T_{0}{M}\) is spanned by the d first vectors of the canonical basis of \(\mathbb {R}^D\). We recall that \(X_i = Y_i + Z_i\), with \(Y_i \in M\) and \(\Vert Z_i\Vert \le s h\), for \(s \le 1/4\). In particular, \(\Vert X_1\Vert \le \Vert Z_1\Vert \le sh \le h/4\).
We adopt the following notation for the local covariance matrix based on the whole sample \(\mathbb {X}_n\).
Note that the tangent space estimator \(\mathtt {TSE}(\mathbb {X}_n,h)_1\) is the space spanned by the first d eigenvectors of \(\widehat{\Sigma }(h)\). From now on we suppose that all the inequalities of Lemma 9.5 are satisfied, defining then a global event of probability larger than \(1-4 ( {1}/{n} )^{2/d + 1}\).
We consider \(h=h_0 \le h_+\), so that Lemmas 9.3 and 9.4 hold. We may then decompose the local covariance matrix as follows.
The first term may be written as
where
According to Lemma 9.4 (with \(\beta =1\)), \(\mu _\mathrm{min}(A(h)) \ge c_d f_\mathrm{min} h^{d+2}\). On the other hand, using Proposition 8.1 and Lemma 9.4 we may write
since \(h \le h_+\) and \(s \le 1/4\). In addition, we can write
with \(\Vert R_2\Vert _{\mathscr {F}} \le C_d \frac{f_\mathrm{max}}{f_\mathrm{min} \sqrt{\kappa }} q(2h) h^2\) according to Lemma 9.5 (with \(\beta =1\)). In turn, the term \(\widehat{\Sigma }_2\) may be decomposed as
with
according to Lemma 9.5. A similar bound on \(R_3\) may be derived,
according to Proposition 8.1 and Lemma 9.4. If we choose \(h = \bigl ( \kappa \frac{\log n}{n-1} \bigr )^{1/d}\), for \(\kappa \) large enough (depending on d, \(f_\mathrm{min}\) and \(f_\mathrm{max}\)), we have
Now, provided that \(\kappa \ge 1\), according to Lemma 9.5, we may write
which, for n large enough, leads to
according to Proposition 10.1.
1.2 E.2 Proof of Proposition 5.5
The proof of Proposition 5.5 follows the same path as the derivation of Proposition 5.1, with some technical difficulties due to the outliers (\(\beta < 1\)). We emphasize that in this framework, there is no additive noise (\(\sigma =0\)). As in the previous section, the analysis will be conducted for \(X_1\) \(\in \) \(\mathbb {X}^{(k)}\), for some fixed \(k \ge -1\), \(k=-1\) referring to the initialization step. Results on the whole sample then follow from a standard union bound. As before, we assume that \(\pi _{{M}}(X_1) = 0\) and that \(T_{0}{M}\) is spanned by the first d vectors of the canonical basis of \(\mathbb {R}^D\). In what follows, denote by \(\widehat{t}\) the map from \(\mathbb {R}^D\) to \(\{0,1\}\) such that \(\widehat{t}(X_i)=1\) if and only if \(X_i\) is in \(\mathbb {X}^{(k)}\).
We adopt the following notation for the local covariance matrix based on \(\mathbb {X}^{(k)}\) (after \(k+1\) iterations of the outlier filtering procedure).
Also recall that we define \(N_0(h)\) and \(N_1(h)\) as the number of points drawn from respectively clutter and signal in \(\mathscr {B}(X_1,h) \cap M\) (based on the whole sample \(\mathbb {X}_n\)). At last, we suppose that all the inequalities of Lemmas 9.5 and 9.6 are satisfied, defining then a global event of probability larger than \(1-7 ( {1}/{n} )^{2/d + 1}\).
We recall that we consider \(h_\infty \le h \le h_{k}\), \(k \ge -1\) (with \(h_{-1} = h_0\)), and \(X_1\) in \(\mathbb {X}^{(k)}\) such that \({d}(X_1,M) \le h/\sqrt{2}\). We may then decompose the local covariance matrix as
The proof of Proposition 5.5 will follow by induction.
Initialization step (\(k=-1\)): In this case \(\mathbb {X}^{(k)} = \mathbb {X}_n\), \(h=h_0\), \({d}(X_1,M) \le h_0/\sqrt{2}\), and \(\widehat{t}\) is always equal to 1. Then the first term \(\widehat{\Sigma }_1^{(k)}\) of (4) may be written as
where
According to Lemma 9.4, \(\mu _\mathrm{min}(A(h)) \ge c_d f_\mathrm{min} \beta h^{d+2}\), and \(\Vert R_1\Vert _{\mathscr {F}} \le 34 \frac{N_1(h) h^3 }{\rho (n-1)}\) according to Proposition 8.1. Moreover, we can write
with \(\Vert R_2\Vert _{\mathscr {F}} \le C_d \frac{f_\mathrm{max}}{f_\mathrm{min} \sqrt{\kappa }} \beta q(2h) h^2\) according to Lemma 9.5. Term \(\widehat{\Sigma }^{(k)}_2\) in inequality (4) may be bounded by
In turn, term \(\widehat{\Sigma }^{(k)}_3\) may be decomposed as
with
according to Lemma 9.5. We may also write
according to Proposition 8.1 and Lemma 9.4. As in the additive noise case (see proof of Proposition 5.1), provided that \(\kappa \) is large enough (depending on d, \(f_\mathrm{min}\), and \(f_\mathrm{max}\)), we have
Since \((n-1)h_0^{d} =\frac{\kappa \log n}{\beta h}\), if we ask \(\kappa \ge \rho \), then for n large enough we eventually get
according to Lemma 9.5. Then, Proposition 10.1 can be applied to obtain
According to Lemma 9.6, we may choose \(\kappa \) large enough (with respect to \(K=\sqrt{2} K^{(0)}\), d, \(f_\mathrm{min}\) and \(\rho \)) and then a threshold t so that, if \(X_1 \in M\), then \(X_1 \in \mathbb {X}^{(0)}\), and if \(d(X_1,M) \ge h_0^2/ \rho \), then \(X_1 \notin \mathbb {X}^{(0)}\).
Iteration step: Now we assume that \(k \ge 0\), and that \({d}(X_i,M) \ge h_k^2/\rho \) implies \(\widehat{t}(X_i)=0\), with \(h_k = \bigl ( \kappa \frac{\log n}{\beta (n-1)} \bigr )^{\gamma _k}\), \(\gamma _k\) being between \(1/(d+1)\) and 1 / d. Let \(h_{\infty } \le h \le h_k\), and suppose that \({d}(X_1,M) \le h_k/\sqrt{2}\). As in the initialization step, \(\widehat{\Sigma }_1^{(k)}\) may be written as
with \(\mu _\mathrm{min}(A(h)) \ge c_d f_\mathrm{min} \beta h^{d+2}\), \(\Vert R_1\Vert _{\mathscr {F}} \le 34 \frac{N_1(h) h^3 }{\rho (n-1)}\), and \(\Vert R_2\Vert _{\mathscr {F}} \le C_d \frac{f_\mathrm{max}}{f_\mathrm{min} \sqrt{\kappa }} \times \beta q(2h) h^2\).
We can decompose \(\widehat{\Sigma }_2^{(k)}\) as
with \(\Vert R_4\Vert _{\mathscr {F}} \le \frac{16N_0(h) h^2}{n-1}\) and \(\Vert R_3\Vert \le \frac{128 N_0(h) h h_k^2}{(n-1)\rho }\), according to Proposition 8.3, for n large enough so that \(h_0^2/\rho \le h_{\infty }\). Term \(\widehat{\Sigma }^{(k)}_3\) may also be written as
with
according to Lemma 9.5. We may also write
according to Propositions 8.1, 8.3 and Lemma 9.4. As done before, we may choose \(\kappa \) large enough (depending on d, \(f_\mathrm{min}\) and \(f_\mathrm{max}\), but not on k) such that
Now choose \(h = h_{k+1} = \bigl ( \kappa \frac{\log n}{\beta (n-1)} \bigr )^{(2\gamma _k +1)/(d+2)}\), with \(\kappa \ge 1\). This choice is made to optimize residual terms of the form \(h/\rho + h_k^2 N_0(h)/h\) coming from \({\Vert R_1 + R_3 + R_5 \Vert _{\mathscr {F}}}/{\mu _\mathrm{min}(A(h_{k+1}))}\). Then we get, according to Lemma 9.5,
where again, \(K_{d,f_\mathrm{min},f_\mathrm{max},\beta }\) does not depend on k. At last, we may apply Proposition 10.1 to get
Then, according to Lemma 9.6, we may choose \(\kappa \) large enough (not depending on k) and t (not depending on k either) so that if \(X_1 \in M\), then \(X_1 \in \mathbb {X}^{(k+1)}\), and if \(d(X_1,M) \ge h_{k}^2/\rho \), then \(X_1 \notin \mathbb {X}^{(k+1)}\). Proposition 5.5 then follows from a straightforward union bound on the sample \(\{X_1, \ldots , X_n\}\).
1.3 E.3 Proof of Proposition 5.8
In this case, we have \({d}(X_j,M) \le h_\infty ^2/\rho \), for every \(X_j\) in \(\mathbb {X}^{(\widehat{k})}\). The proof of Proposition 5.8 follows from the same calculation as in the proof of Proposition 5.5, replacing \(h_k^2/\rho \) by its upper bound \(h_{\infty }^2/\rho \) and taking \(h_{k+1} = h_{\infty }\) in the iteration step.
F Proof of the Main Reconstruction Results
We now prove main results Theorem 2.7 (additive noise model), and Theorems 2.8 and 2.9 (clutter noise model).
1.1 F.1 Additive Noise Model
Proof of Corollary 5.2
Let \( Q \in \mathscr {G}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\sigma }\). Write \(\varepsilon = c_{d,f_\mathrm{min},f_\mathrm{max}} (h \vee \rho \sigma /h)\) for \(c_{d,f_\mathrm{min},f_\mathrm{max}}\) large enough, an consider the event A defined by
Then from Proposition 5.1 and Lemma 9.1, \(\mathbb {P}_Q(A) \ge 1 - 5 \bigl ( \frac{1}{n} \bigr )^{2/d}\), and from the definition of \(\varepsilon \) and the construction of \(\mathbb {Y}_n\), for n large enough,
which yields the result. \(\square \)
Proof of Theorem 2.7
Following the above notation, we observe that on the event A, Theorem 4.4 holds for \(\varepsilon = c_{d,f_\mathrm{min},f_\mathrm{max}} (h \vee \rho \sigma /h)\), \(\theta = \varepsilon /(1140 \rho )\) (where we used the fact that \(\theta \le 2 \sin \theta \)) and \(\eta = \varepsilon ^2/(1140 \rho )\) with high probability, so that the first part of Theorem 2.7 is proven. Furthermore, for n large enough,
where for the last line we used the diameter bound of Lemma 2.2. \(\square \)
1.2 F.2 Clutter Noise Model
Proof of Corollary 5.6
Let \( P \in \mathscr {O}_{D,d,f_\mathrm{min},f_\mathrm{max},\rho ,\beta }\). For n large enough, write \(\varepsilon = c_{d,f_\mathrm{min}f_\mathrm{max}} h_{k_\delta }\) for \(c_{d,f_\mathrm{min}f_\mathrm{max}}\) large enough, and consider the event
From Proposition 5.5 and Lemma 9.1, \(\mathbb {P}_P\bigl (A^\delta \bigr ) \ge 1- 8 ( {1}/{n} )^{2/d}\)and from the definition of \(\varepsilon \) and the construction of \(\mathbb {Y}_n^\delta \), for n large enough,
which yields the result. \(\square \)
Proof of Theorem 2.8
Following the above notation, we observe that on the event \(A^\delta \), Theorem 4.4 holds for \(\varepsilon = c_{d,f_\mathrm{min}f_\mathrm{max}} h_{k_\delta }\), \(\theta = \varepsilon /(1140 \rho )\) and \(\eta = \varepsilon ^2/(1140 \rho )\), so that the first part of Theorem 2.8 is proven. As a consequence, for n large enough,
where for the second line we used the fact that \(M \cup \widehat{M}_\mathtt {TDC \delta } \subset \mathscr {B}_0\), a ball of radius \(K_0 = K_0(d,f_\mathrm{min},\rho )\). \(\square \)
Finally, Theorem 2.9 is obtained similarly using Proposition 5.8.
Rights and permissions
About this article
Cite this article
Aamari, E., Levrard, C. Stability and Minimax Optimality of Tangential Delaunay Complexes for Manifold Reconstruction . Discrete Comput Geom 59, 923–971 (2018). https://doi.org/10.1007/s00454-017-9962-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00454-017-9962-z