Abstract
We consider non-uniform random sampling in a signal space with finite rate of innovation \(V^{2}(\varLambda,\varPhi) \subset{\mathrm {L}}^{2}(\mathbb {R}^{d})\) generated by a series of functions \(\varPhi=(\phi_{\lambda})_{\lambda \in\varLambda}\). A subset \(V_{R,\delta}^{2}(\varLambda,\varPhi)\) of \(V^{2}(\varLambda,\varPhi)\) is consisting of functions concentrates at least \(1-\delta\) of the whole energy in a cube with side lengths \(R\). Under mild assumptions on the generators and the probability distribution, we show that for \(R\) sufficiently large, taking \(O(R^{d} \log(R^{d}))\) many samples with such the non-uniform distribution yields a sampling set for \(V_{R,\delta}^{2}(\varLambda,\varPhi)\) with high probability. We impose compact support on the generators as an additional constraint for obtaining a reconstruction algorithm from non-uniform random sampling with high probability.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
The space of signals with finite degree of freedom per unit of time is called the space with finite rate of innovation (FRI) [2, 5, 12, 15, 17, 19, 23]. The concept is introduced by Vetterli et al. [23] at first. The FRI model is ubiquitous and has wide scientific applications such as radar imaging [15], compression of electrocardiogram signals [2], curve fitting [12]. The model of signals with finite rate of innovation can cover following cases: (i) band-limited signals [3], (ii) signals in some shift-invariant spaces [1, 7,8,9, 24, 25], (iii) non-uniform splines [16, 20, 21], (iv) stream of pulses \(\sum_{k} a_{k}p(t-t_{k})\) where \(p\) is some pulse signal shape, applied in GPS applications and cellular radio, (v) sum of some of the signals above [19], and so on.
In this paper, we consider the signal spaces of finite rate of innovation \(V^{2}(\varLambda,\varPhi)\subset{\mathrm{L}}^{2}(\mathbb {R}^{d})\), defined as the closed linear span of a tuple of generators \(\{\phi_{\lambda}\} _{\lambda\in \varLambda} \subset{\mathrm{L}}^{2}(\mathbb {R}^{d})\), which is introduced in [17, 19]:
where \(\varLambda\) is a relatively-seperated subset of \(\mathbb{R}^{d}\).
Realistically we can learn about \(f\in V^{2}(\varLambda,\varPhi)\) only if the samples are taken in the set where most of the \(L^{2}\)-norm is localized. So motivated by [3, 8], we focus on the sampling problem on the following subset of \(V^{2}(\varLambda,\varPhi)\):
where \(C_{R}=[-R/2, R/2]^{n}\) and \(0< \delta<1\). Thus \(V^{2}_{R,\delta}(\varLambda, \varPhi)\) represents the subset consisting of those functions whose energy is largely concentrated on \(C_{R}\).
Usually, in order to guarantee the success of reconstruction, we demand the sampling inequality as below fulfilled:
where \(A\) and \(B\) are some positive constants. However, owing to the randomness in choosing the sampling points, we can not guarantee the sampling inequality fulfilled surely. Instead, our goal is to chase the following probability estimate:
where \(\epsilon>0\) can be taken as arbitrarily small. This problem is studied for the cases of band-limited signal spaces, shift-invariant signal spaces and finitedly generated shift-invariant signal spaces, respectively [3, 8, 10, 25].
In former research [3, 4, 6, 8, 13], the probability distribution of random sampling points is taken as the uniform distribution in \(C_{R}\). It is often beneficial in statistical sampling and function integration to sample uniformly over the applicable parameter space. However, non-uniform distributions also has application in real world [11, 13, 14]. A sequence of points which is uniformly distributed can be mapped into another sequence that reflects a desired non-uniform join probability distribution function [13] using different techniques like the rejection method or weighted sampling [11, 14]. In this paper, we will consider a more generalized probability distribution: we only demand that the probability density function exists, and has positive lower bound in \(C_{R}\). We will show that the uniform random sampling in \(C_{R}\) is a special case in our results about non-uniform random sampling.
In this paper, we will show the probability estimate (1) for non-uniform random sampling in signal spaces of finite rate of innovation. This is one of the main results in this paper.
In [9], a reconstruction method from nonuniform sampling is presented in shift-invariant spaces. But the sampling is NOT random sampling in [9]. In [4], a linear algorithm is obtained for the random sampling from regular languages. But how to give a reconstruction algorithm from uniform random sampling is still an open question in bandlimited functions spaces [3]. In this paper, we try to show a reconstruction algorithm from non-uniform random sampling in spaces with finite rate of innovation. This is one direct application of our main results.
Many problems arising in industry, statistics, physics, and finance require to compute multivariate integration \(\int_{C_{R}} f(x)dx\). The sample mean method has been recommended to give an approximation to \(\int_{C_{R}} f(x) dx\). The Koksma–Hlawka inequality gives an upper bound for the approximation error
where \(\varGamma=(x_{i})_{i=1}^{s}\subset{C_{R}}\) is the set of sampling points with uniform distribution, \(C(\varGamma)\) is the discrepancy of \(\varGamma\) and \(P (f)\) is a measure of the variation of \(f\) [6, 13]. A spontaneous question is that, can we use the mean of the squares of those to estimate \(\int_{\mathbb{R}^{d}} |f(x)|^{2} dx\)? Or what is the maximum error between \(\int_{\mathbb{R}^{d}} |f(x)|^{2} dx\) and \(\frac{1}{s}\sum_{i=1}^{s}|f(x_{i})|^{2}\)? In this paper, we provide that the estimate error is direct proportional to the \(\mathrm{L}^{2}\) norm of the function. If \(\varGamma=\{x_{i}\}_{i=1}^{s}\subset{\mathbb{R}^{d}}\) with non-uniform distribution fulfills some conditions with probability at least \(1-\epsilon\) for sufficiently small \(\epsilon>0\) , then we can get that the estimate error is a functional only dependent on \(f\), as shown below:
This is one direct application of our main results too.
The paper is organized as follows. In Sect. 2, we state the theorem of random sampling in a space of finite rate of innovation, and give notations and preliminaries. In Sect. 3, some examples are shown about the generators and probability density function. In Sect. 4, spectrum estimation based on Hilbert–Schmidt operator is described. In Sect. 5, we discuss non-uniform random sampling in finite sums of eigenspaces. The proof of main results are provided in Sect. 6. And in the last section, we give a reconstruction algorithm for a special case, and give some numerical testes to verify them. Some properties of signal spaces \(V^{2}(\varLambda,\varPhi)\) are presented in appendix.
2 Statement of the Main Result
2.1 Notations and Preliminaries
We first define relatively-seperated point sets, the relative covering index and the absolute covering index.
We say a point set \(X={x_{j}} \subset \mathbb {R}^{d}\) is relatively-seperated [17], if the relative covering index
is finite. And we can define the absolute covering index as below:
We can easily see that \(N_{0}(X)\leq D(X)\) and \(\sup_{x_{0}\in[0,1)^{d}} N_{0}(X+x_{0})=D(X)\) for any relatively-seperated set \(X\). The two definitions are used in different usages.
Throughout the paper, we consider a space of finite rate of innovation \(V^{2}(\varLambda,\varPhi)\subset{\mathrm{L}}^{2}(\mathbb {R}^{d})\), defined as the closed linear span of a tuple of generators \(\{\phi_{\lambda}\}_{\lambda\in\varLambda} \subset{\mathrm{L}}^{2}(\mathbb {R}^{d})\):
where \(\varLambda\) is a relatively-seperated subset of \(\mathbb {R}^{d}\) with relative covering number \(D(\varLambda)\).
In this paper, we use the polynomial weight function \(u\) below [17]:
where \(\alpha\geq0\). Usually we demand that the below norm of \(\varPhi\) is finite for some \(q,p \in[1,\infty]\):
The definition is a bit different with that in [17] and [19]. However, the definition in [17] and that in this paper are equivalent. Some properties of the signal space \(V^{2}(\varLambda ,\varPhi )\) will be shown in Appendix.
We pay attention to the subset \(V_{R,\delta}^{2}(\varPhi,\varLambda )\) of \(V^{2}(\varLambda,\varPhi) \):
What’s more, we will adopt a measurable typical or classical probability density function (P.D.F.) \(w_{1}:\mathbb{R}^{d} \to\mathbb{R}\) which satisfies \(\|w_{1}\|_{\mathrm{L}^{1}(\mathbb {R}^{d})}=1\) and \(w_{1}(x)\geq0\), \(\forall x \in \mathbb {R}^{d}\). And the density function of the probability distribution we use to sample is defined as:
We denote \(v_{R}=\sqrt{w_{R}}\), \(\| x\|_{\infty}=\max(|x_{1}|, \ldots, |x_{d}|)\), and \(C_{R}=[-\frac {R}{2},\frac{R}{2}]^{d}\). Throughout this paper, we will frequently refer to the constants as in Table 1.
2.2 Assumptions and Main Results
We collect our assumptions on the generators in the following list:
-
(A.0)
\(\varPhi\) is a Riesz basis of \(V^{2}(\varLambda,\varPhi)\).
-
(A.1)
\(\|\varPhi\|_{\infty,1,u_{0}}<\infty\).
-
(A.2)
\(\|\varPhi\|_{2,2,u_{\alpha}}<\infty\) for some \(\alpha>0\).
By (A.1), we have \(\forall1\leq p,q \leq\infty\), \(\|\varPhi\| _{q,p,u_{0}}\leq\|\varPhi\|_{\infty,1,u_{0}}<\infty\).
By (A.0) and [19, Theorem 4.1], we have that \(\forall1\leq p,q \leq\infty\), \(\|\widetilde{\varPhi}\|_{q,p,u_{0}}<\infty\).
The assumptions on the probability density function are shown in the following:
-
(B.0)
\(w_{1}\) is essentially bounded, or \(K_{w}:=\|w_{1}\| _{\mathrm{L}^{\infty}(\mathbb {R}^{d})}<\infty\).
-
(B.1)
The essential infimum of \(w_{1}\) in \(C_{1}\) is positive, that is, there exists \(\rho>0\) such that \(\mathrm{ess}\,\mathrm{inf}_{x \in C_{1}}w_{1}(x)\geq\rho K_{w}\), where \(K_{w}\) is defined in (B.0).
According to (6) and the assumptions of \(w_{1}\), we get that \(\|w_{R}\|_{\mathrm{L}^{\infty}(\mathbb {R}^{d})}=R^{-d}K_{w}\) by (B.0), and \(\mathrm{ess}\,\mathrm{sup}_{x \in C_{R}}w_{R}(x)>\rho R^{-d} K_{w}\) by (B.1).
Theorem 1
Assume that \(\varPhi\)satisfies assumptions (A.0)–(A.2), and \(w_{1}\)satisfies assumptions (B.0)–(B.1). Let \((x_{j})_{j \in\mathbb{N}} \subset \mathbb {R}^{d}\)be a sequence of i.i.d random variables with the probability density function \(w_{R}\). Assume that \(R \ge R_{0}\)and \(\delta, \nu\in(0,1)\)are sufficiently small to guarantee that
Let \(0 < \epsilon< 1\). If the number \(s\)of samples satisfies
then the sampling inequality
holds with probability at least \(1-\epsilon\).
Remark 2
The result above shows the sampling rate of \(O(R^{d} \log R)\) holds in the space of finite rate of innovation when fulfills the assumptions above and \(d\) is fixed, even if the distribution of random sampling is not uniform. If we have one more condition that \(u_{\alpha}\) is 2-admissible, or \(\alpha>d/2\) (see [17]), then \(\alpha'=d\) and the sampling rate could be actually \(O(R^{d} \log{R^{d}})\) when \(d\) is not fixed.
Remark 3
Theorem 1 is the main result in [3, 8, 25] in bandlimited signal spaces and shift-invariant spaces for uniform random sampling, respectively.
Corollary 4
Suppose that \(\mathrm{supp}{\;\phi_{\lambda}}\subset C_{R}\)for any \(\lambda\in\varLambda\), thus \(V_{R,\delta }^{2}(\varLambda,\varPhi)=V^{2}(\varLambda,\varPhi)\)for any small \(\delta>0\). For a sampling set \(\varGamma=(x_{j})_{j=1}^{s}\subset \mathbb {R}^{d}\), define the sampling operator \(S:V^{2}(\varLambda ,\varPhi) \to \mathbb {C}^{s}\)by \(f \mapsto(f(x_{j}))_{j=1}^{s}\). Denoting \(\ell\)as the sampling result \((f(x_{j}))_{j=1}^{s}\), we have that \(S^{*}S\)is invertible under the conditions in Theorem 1with probability at least \(1-\epsilon\), and thus \(f=(S^{*}S)^{-1} S^{*} \ell\)in this case. A schematic of random sampling and reconstruction is shown in Fig. 1.
Corollary 5
Denote the conditional number of the operator \(S^{*}S\)in Corollary 4by \(\kappa(S^{*}S)\). Then the conditional number satisfies
under the conditions in Theorem 1. The upper bound above conditional number is the quotient of the upper bound and lower bound of the sampling inequality.
Corollary 6
For any \(f\in V_{R,\delta}^{2}(\varLambda,\varPhi)\), we have that
when the conditions in Theorem 1are fulfilled.
Remark 7
We expect that \(D(s)\) in the below inequality tends to zero when \(s\to \infty\):
However, this can not reach in the FRI space. Thus the estimation in Corollary 6 is very coarse.
3 Some Examples of Generators and Probability Density Function
Two examples of generators of a space of finite rate of innovation fulfilling the assumptions of (A.0)–(A.2).
Example 8
The space \(S_{m}\) of all \(\mathrm{L}^{2}\) polynomial splines of order \(m\) fulfilling \(C^{m-2}\) continuity conditions at each knot \(t_{i}\), where \(m \in \mathbb {N}^{+}\) and \(T=(t_{i})_{i=-\infty}^{+\infty}\) is a bi-infinite increasing sequence satisfying
The space can be written as
where \(B_{i}\) is the normalized B-spline associated with the knots \(t_{i},\ldots,t_{i+m}\) [16]. We have \(B_{i}(x)=0\) when \(x\leq t_{i}\) or \(x\geq t_{i+m}\). Let \(\varLambda=T\) and \(\phi_{t_{i}}=B_{i}\), then the support of \(B_{i}\) is a subset of \([t_{i},t_{i+m}]\). Obviously, \(D(T)\leq\frac {m}{T_{0}}+m\), so \(T\) is relatively seperated.
Let \(q=\infty,p=1\), and take any weight function \(u(x)\) that monotonicly increases with \(|x|\). Then we have that
by using that \(\|B_{i}\|_{\infty}\leq1\) according to [16]. So the assumption (A.1) holds.
At the same time, this implies that \(\|\varPhi\|_{q,p,u}\,{\leq}\,\|\varPhi \|_{\infty,1,u}\,{\leq}\,(T_{1}+2+(\frac{1}{T_{0}}+2)m)u(T_{1}){\,{<}\,\infty}\) for any \(1\leq q,p \leq\infty\). So the assumption (A.2) holds.
Moreover, by [16, Theorem 4.41], we have that for any \(f=\sum_{i=-\infty}^{+\infty}c(i)B_{i}\),
In the other hand, we have
Thus \(\varPhi=\{B_{i}\}_{i=-\infty}^{+\infty}\) is a Riesz basis of \(S_{m}\). Therefore the assumption (A.0) is fulfilled.
Example 9
The space of trigonometric polynomials segmented in \(C_{R}\). This is a very trivial example. We will consider the following signal spaces:
We can take \(\varLambda\) arbitrarily and give the correspondence between \(\varLambda\) and the basis arbitrarily, and denote the basis still by \(\varPhi\). Note that \(\varLambda\) is finite, hence is relatively-seperated.
If we take \(R=mt\) where \(m\in \mathbb {Z}^{+}\), it is a Riesz basis obviously since the basis is orthogonal. Moreover, for any weight function \(u(x)\) that monotonicly increases with \(|x|\), we can easily get that \(\|\phi_{\lambda}(\cdot)u(\cdot-\lambda)\|_{\mathrm {L}^{\infty}(k+[0,1]^{d})}\leq u(R/2+1+\max_{\lambda\in\varLambda}\{\| \lambda\|_{\infty}\})\) is finite for \(|k|\leq\lceil R/2 \rceil\) while \(\|\phi_{\lambda}(\cdot)u(\cdot-\lambda)\|_{\mathrm{L}^{\infty}(k+[0,1]^{d})}=0\) for \(|k|> \lceil R/2 \rceil\). Then \(\|\varPhi\| _{\infty,1,u}\leq\max\{(2k_{0}+1)^{d},(R+2)^{d}\}\cdot u(R/2+1+\max_{\lambda\in\varLambda}\{\|\lambda\|_{\infty}\})\) holds.
Two examples of probability density function fulfilling the assumptions (B.0)–(B.1).
Example 10
Beta distribution of one dimension
where \(\alpha,\beta\geq1\), \(a<-\frac{1}{2}\), \(b>\frac{1}{2}\) and \(B(\alpha,\beta)=\frac{\varGamma(\alpha)\varGamma(\beta )}{\varGamma(\alpha+\beta)}\) is the Beta function. We can check that \(K_{w}=B_{a,b,\alpha,\beta}(\frac{\alpha-1}{\alpha +\beta-2})<\infty\) and \(\rho=\frac{\min\{B_{a,b,\alpha,\beta }(-\frac{1}{2}),B_{a,b,\alpha,\beta}(\frac{1}{2})\}}{K_{w}}>0\). Specially, we take \(\alpha=\beta=1\), \(a\to-\frac{1}{2}-\) and \(b\to \frac{1}{2}+\), then \(w_{R}\) degenerates into the uniform distribution on \(C_{R}\), just like the distribution used in [3] and [8]. Figure 2 is the plot of the P.D.F. of a typical Beta distribution with \(a=-1\), \(b=1\), \(\alpha=1\), \(\beta=2\).
Example 11
Normal distribution centered at 0:
where \(\sigma>0\). We have that \(K_{w}=N_{\sigma}(0)=\frac{1}{\sqrt{2\pi\sigma ^{2}}}<\infty\) and \(\rho=\frac{N_{\sigma}(\frac{1}{2})}{N_{\sigma}(0)}=\exp{(-\frac{1}{8\sigma^{2}})}>0\). For fixed \(\delta\in(0,\frac{1}{2+24D_{3}})\), in order to optimize the lower bound \(A\) of the sampling inequality, the optimal choice \(\sigma\) satisfies
It is easy to check that there exists a unique \(\sigma\) satisfying Eq. (12). We can guarantee that \(A\) is positive when \(\sigma\) take the optimal choice and \(\nu\) is small enough by the assumption \(0<\delta<\frac{1}{2+24D_{3}}\). Figure 3 is the plot of the P.D.F. of standard normal distribution.
4 Spectrum Estimation Based on Hilbert–Schmidt Operator
In the following, we will mostly work with the notation
where unconditional convergence in the strong operator topology is provided by the frame properties for all \(f \in{\mathrm{L}}^{2}(\mathbb{R}^{n})\). The tensor product notation \(v \otimes w\) refers to the rank-one operator \((v \otimes w) f \mapsto\langle f, w \rangle v\). By our assumptions (A.0)–(A.2), it is obvious that \(P_{\varPhi}\) is the orthogonal projection onto \(V^{2}(\varLambda,\varPhi)\).
Given \(R>0\), we write \(Q_{R} : \mathrm{L}^{2}(\mathbb {R}^{d}) \to{\mathrm {L}}^{2}(\mathbb {R}^{d})\) for the orthogonal projection operator \(f \mapsto f \cdot\chi_{C_{R}}\). We introduce the localization operator
We will show that \(A_{R}\) is self-adjoint and Hilbert–Schmidt, and thus has a basis of eigenvectors with associated \(\ell^{2}\) spectrum. We denote the span of the eigenvectors associated to the largest \(N\) eigenvalues by \(\mathcal{P}_{N}\), and then establish a random sampling theorem for that space. Next we will built the connection between the sampling of the elements in \(V_{R,\delta}^{2}(\varLambda,\varPhi)\) and that in \(\mathcal{P}_{N}\). How to choose the proper value of \(N\) depends on the decay of the spectrum of \(A_{R}\). The localization of the above operator is the closest connection with the energy condition of \(V_{R,\delta}^{2}(\varLambda,\varPhi)\).
Lemma 12
\(A_{R}\)is a positive-semidefinite Hilbert–Schmidt operator.
Proof
First of all, \(A_{R}\) is positive-semidefinite since
by using the self-adjointness of \(Q_{R}\).
We next prove that \(Q_{R} P_{\varPhi}\) is Hilbert–Schmidt operator, which will imply that \(A_{R}\) is Hilbert–Schmidt operator.
Let \(\forall\lambda\in\varLambda\), we have
By (13) and continuity of \(Q_{R}\), we have
with unconditional convergence in the strong operator topology.
Next we will prove that
In fact, we have
Thus, we have
with convergence in the strong operator topology. Similarly to the proof of Eq. (14), we see that
Combing with (14) and the Cauchy–Schwarz inequality, we can get that
Hence \(Q_{R} P_{\varPhi}\) is Hilbert–Schmidt operator. □
It follows that \(A_{R}\) has an orthogonal normalized basis of eigenvectors, whose associated eigenvalues are non-negative and square-summable. So the cluster points of the spectrum of \(A_{R}\) are either \(\{0\}\) or \(\varnothing\), and thus they can be sorted in decreasing order. Denoting the non-zero eigenvalues by \((\mu_{n})_{n \in I}\) sorted in decreasing order and the associated eigenfunctions by \((\psi_{n})_{n \in I}\), with \(I\) being either \(\mathbb {N}_{+}\) or \(\{1, \ldots, M\}\) for some integer \(M\). Then \(A_{R}\) is given as the sum
We can easily check that \(\psi_{n} \in V^{2}(\varPhi,\varLambda)\), as well as \(P_{\varPhi}(\psi_{n}) = \psi_{n}\). Since \(A_{R}\) is a composition of orthogonal projections, we have \(\mu _{1} \le\| A_{R} \|_{op} \le1\). Let
For the sampling problem we will need some information on the spectrum of \(A_{R}\). Let
and \(N(R)=0\) when \(\mu_{1} < 1/2\). Then whenever \(N(R)>0\), we have \(\mu_{N(R)} \ge1/2 > \mu_{N(R)+1}\).
An upper bound for \(N(R)\) is provided in the following lemma.
Lemma 13
Let \(\beta\)and \(R_{0}\)be that defined in Table 1. Then for all \(R>R_{0}\), the inequalities \(0 < N(R) \le D(\varLambda )\beta^{d} R^{d^{2}/\alpha'}\)hold, where \(D(\varLambda)\)stands for the relative covering index of \(\varLambda\), defined in (2).
Proof
We use the minimax formula for a given order of eigenvalue
Now fix \(S\geq R+2\), and denote \(\mathcal{H}_{S} := \mathrm{span} \{ \widetilde{\phi}_{\lambda}: \| \lambda\|_{\infty}\le S/2 \}\). It follows that
For any unit vector \(f\) in \(\mathcal{H}_{S}^{\bot}\), we obtain
We denote \(c_{\lambda}= \langle f, \widetilde{\phi}_{\lambda}\rangle\), \(c=(c_{\lambda})_{\lambda\in\varLambda}\),
and an infinite, positive semidefinite matrix
by recalling the assumption that \(f \bot\widetilde{\phi}_{\lambda}\) for \(\| \lambda\|_{\infty}\le S/2\). In particular, we get
by using Proposition 22 in Appendix. To estimate the Hilbert–Schmidt norm of the matrix, we have
where Proposition 22 in Appendix is used in the above argument again.
Now we will estimate \(\sum_{\lambda,\|\lambda\|_{\infty}>S/2}\| Q_{R}\phi_{\lambda}\|_{2}^{2}\) by the following:
Therefore we have
For \(S = (\beta-\frac{3}{2}) R^{d/\alpha'}+2\) with \(\beta\) in Table 1, one finds that \(R>R_{0}\geq2\) and the definition of \(\alpha '\) yield that \(R \le R^{d/\alpha'}\), and hence
Hence we have shown
for all unit vectors \(f \in\mathcal{H}_{S}^{\bot}\).
If \(R>2\), then
and the minimax estimate yields \(\mu_{N} < 1/2\).
On the other hand, if \(R>R_{0}\geq2\|\lambda^{*}\|_{\infty}+2\), then
So for all \(R>R_{0}\) the following inequality holds
□
5 Non-uniform Random Sampling in Finite Sums of Eigenspaces \(\mathcal{P}_{N}\)
Recall from the previous section that \(\lambda_{1} \ge\lambda_{2} \ge\)... are the eigenvalues of \(A_{R}\), with corresponding eigenfunctions \(\psi_{1},\psi_{2},\ldots\). The span of the first \(N\) eigenfunctions is denoted by \(\mathcal{P}_{N}\). We let \(\Delta_{N} = \mathrm{diag}(\lambda_{1},\ldots,\lambda_{N})\).
The goal of this section is to show a non-uniform random sampling statement for \(\mathcal{P}_{N}\). We will show it by applying a matrix Bernstein inequality which will be stated in Theorem 14, using the following notation: For \(A \in \mathbb{C}^{N \times N}\), we let \(\| A \|\) denote the operator norm with respect to the Euclidean norm. Further, the inequality \(A \le B\) for two matrices \(A,B\) of equal size means that \(B-A\) is positive semidefinite.
Theorem 14
(Matrix Bernstein inequality [22, Theorem 1.4]) Let \(X_{j}\)be a sequence of independent, random self-adjoint \(N\times N\)-matrices. Suppose that \(\mathbb{E}X_{j} =0\), and \(\| X_{j}\|\leq B\)a.s. Then for all \(t>0\),
holds where \(\lambda_{\max}(U)\)is the largest singular value of a matrix \(U\)so that \(\|U\|=\lambda_{\max}(U^{*}U)^{1/2}\)is the operator norm, and \(\sigma^{2}= \Vert \sum_{j=1}^{s}\mathbb{E}(X_{j}^{2}) \Vert \).
The random matrices under consideration are constructed as follows: For each \(j \in\mathbb{N}\) and \(k,l \in\{ 1, \ldots, N \}\), we introduce the \(N \times N\) rank-one random matrix \(M_{j}\) defined by
Here the \(x_{j}\)’s denote i.i.d. random variables, and are distributed in \(\mathbb {R}^{d}\) with probability density function \(w_{R}\). Let
We can now formulate and prove the non-uniform random sampling statement for \(\mathcal{P}_{N}\) by using Theorem 14.
Theorem 15
Let \((x_{j})_{j \in\mathbb{N}}\subset \mathbb {R}^{d}\)be a sequence of independent and identically distributed random variables with probability density function \(w_{R}\). Then, for all \(\nu\ge0\)and \(s \in \mathbb {N}\):
holds where \(v_{R}\)denotes the square root of \(w_{R}\)and \(\| v_{R} f \|_{2}^{2}\)is the expectation of \(|f(x_{j})|^{2}\).
Proof
By the definition of \(M_{j}\) defined by (15), we have
Furthermore, for any unit vector \(f \in\mathcal{P}_{N}\) we can rewrite it by \(f = \sum_{n=1}^{N} c_{n} \psi_{n}\) with a unit vector \((c_{n})_{n}\) of coefficients, and
and similarly
which implies that both \(M_{j}\) and \(\mathbb{E}M_{j}\) are positive semi-definite.
Thus we have
where \(\lambda_{\mathrm{min}}\) represents the smallest eigenvalue of a self-adjoint matrix.
Next we estimate \(\|X_{j}\|\). For all \(\|c\|_{2}=1\) and \(f=\sum_{k=1}^{s} c_{k} \psi_{k}\), we have
Since we can ignore a zero-measure subset from the existence of probability density function of \(x_{j}\), as well as
Thus we have
by Proposition 24 in Appendix. Therefore \(\|X_{j}\| \leq D_{2}\) holds.
Finally we will estimate \(\sigma^{2}\). In fact, we have that
and
Thus \(M_{j}^{2} = \sum_{\ell=1}^{N} |\psi_{\ell}(x_{j})|^{2} (M_{j})_{\mathit{km}}\). Moreover, we have the following estimation that
almost everywhere by Proposition 24 in Appendix.
Therefore
as well as
By the arguments above, we can get that
Now the statement of the theorem follows from the matrix Bernstein inequality formulated in Theorem 14, with proper estimation constants provided above. □
6 Proofs of the Main Results
It remains to transfer the random sampling statements from the spaces \(\mathcal{P}_{N}\) to the set \(V_{R,\delta}(\varPhi)\). The following lemma is a first step in this direction, by providing a norm estimate for the projection onto \(\mathcal{P}_{N}\).
Lemma 16
Let \(N \in\mathbb{N}\)and \(\gamma\in\mathbb{R}\)with \(\lambda_{N} \ge\gamma\ge\lambda_{N+1}\). Denote the orthogonal projection onto \(\mathcal{P}_{N}\)by \(E_{N}\), and the orthogonal projection onto the orthogonal complement by \(F_{N}\). Then for all \(f \in V_{R,\delta }(\varPhi)\), we have
If \(N=N(R) \neq0\), then these estimation can be simplified into
Proof
Let \(f \in V_{R,\delta}(\varPhi)\), w.l.o.g. \(\| f \|_{2} = 1\). Since \(f = P_{\varPhi}f\), we obtain
Let \(c_{j} = \langle f, \psi_{j} \rangle\), and define
and \(B = 1- A = \| F_{N} f \|_{2}^{2}\). Then \(\sum_{j=N+1}^{\infty}|c_{j}|^{2} \le\| F_{N} f \|_{2}^{2} = 1-A\). Using \(\gamma\ge\lambda _{N+1} \ge\lambda_{N+2} > \cdots\) and \(\lambda_{j} \le1\), we find
Solving this inequality for \(A\) yields \(A \ge 1-\frac{\delta}{1-\gamma}\), which implies \(B \le \frac{\delta}{1-\gamma}\). Finally, \(\gamma\le\lambda_{N}\) holds, which implies
For \(N=N(R) \neq0\), we may pick \(\gamma= 1/2\), which results in the estimates given for this case. □
To connect the sampling theorem between \(V_{R,\delta}^{2}(\varLambda ,\varPhi)\) and the eigenspace \(\mathcal{P}_{N}\), we need a statement below.
Lemma 17
Let \(X=\{x_{j}\}_{j=1}^{s}\)be a finite subset of \(\mathbb {R}^{d}\)and \(\alpha\in[\mu_{N+1},\mu_{N}]\). Assume that the inequality
holds for all \(g \in\mathcal{P}_{N}\). Then the inequality
holds for all \(f \in V_{R,\delta}^{2}(\varLambda,\varPhi)\), where
Proof
Denoting the orthogonal projection of a square-integrable function \(f\) to \(\mathcal{P}_{N}\) by \(\mathit{Ef}\). By Lemma 16 above and Proposition 23 in Appendix, we have
Then we get the conclusion. □
Before proving the main result, we have a following statement about absolute covering index.
Lemma 18
Suppose \(X=\{x_{j}\}_{j=1}^{s}\)are independent and identically distributed random variables in \(\mathbb {R}^{d}\). Let \(P\geq\sup_{k \in \mathbb {Z}^{d}}\mathbb {P}(x_{j} \in k+[0,1]^{d} )\)and \(a>e P\)with \(as\geq1\). Then we have
where \(N_{0}(X)\)represents the absolute covering index of \(X\)stated in (3).
Proof
Let \(P_{k}=\mathbb {P}(x_{j} \in k+[0,1]^{d} )\) for \(k \in \mathbb {Z}^{d}\). Combing with Chebyshev’s inequality, then we have that
Taking the optimal choice \(b_{k}=\log (\frac{a}{P_{k}} )\), we obtain that
□
Remark 19
For a distribution with probability density function \(w(x)\), we can take \(P=\|w\|_{\infty}\) so that for \(a\) and \(s\) satisfying the conditions of Lemma 18. So we have
Remark 20
If we take a special example that the distribution is uniform distribution on \(C_{R}\), then \(P\) can be taken as \(\lfloor R \rfloor ^{-d}\). So we have
It provide a different estimate from [3, Lemma 8]. Note that we demand \(a\) more strictly conditions than that in [3], but our estimate might be better than that in [3]. And our result is suitable for a large kinds of non-uniform probability distribution.
Then we can see the following statement.
Theorem 21
Let \((x_{j})_{j \in\mathbb{N}}\subset \mathbb {R}^{d}\)be a sequence of independent random variables with probability density function \(w_{R}\). Assume that \(R>R_{0}\), and furthermore
Then, for any \(s \in\mathbb{N}\)and \(s \geq\frac{R^{d}}{3 K_{w}}\),
is strictly positive, and the sampling estimate
holds with probability at least
Proof
Define the random variable \(N_{0}\) as the absolute covering index of \(x_{1},\ldots,x_{s}\). Fix \({N = N(R)}\), and consider the events
and
Following Lemma 17 and taking \(\alpha=\frac{1}{2}\), we see that for all \((x_{1},\ldots,x_{s})\) in the complement of \(V_{1} \cup V_{2}\) and all \(f \in V_{R,\delta}(\varPhi)\),
Theorem 15 combined with Lemma 13 imply that \(V_{1}\) occurs with probability at most
Furthermore, Lemma 18 yields that \(V_{2}\) occurs with probability at most
Thus the lower estimate in (18) occurs at least with the probability given in (19), whereas the upper estimate follows Proposition 24 in Appendix. □
Proof of Theorem 1
By the definitions of \(M\) and \(\beta\) in Table 1, we have \(D(\varLambda)R^{d^{2}/\alpha'}\beta^{d}\leq D(\varLambda )R^{d^{2}/\alpha'}M\) as well as \(\frac{R^{d}}{K_{w}}\leq D(\varLambda )R^{d^{2}/\alpha'}M\). And by the assumptions in (7), we have \(( \frac{e}{3} )^{3sR^{-d}K_{w}} \leq\exp ( -\frac{\nu^{2} s}{2 D_{2} R^{d} (K_{w}+\frac{v}{3} )} )\). Hence, as soon as
holds we have that \(D(\varLambda)R^{d^{2}/\alpha'}M \exp ( -\frac {\nu^{2} s}{2 D_{2} R^{d} (K_{w}+\frac{v}{3} )} ) \le\epsilon/2\). The theorem is proved. □
Proof of Corollary 4 and 5
Obviously \(\ell=Sf\), therefore \(S^{*}\ell=S^{*}Sf\). Now we estimate the spectrum of \(S^{*}S\). For all \(f \in V^{2}(\varLambda,\varPhi)\), we can see that
By Theorem 1, we have that
hold under the conditions in Theorem 1 with probability at least \(1-\epsilon\). This implies that \(\|S^{*}S\|\leq sD_{2}\), as well as \(S^{*}S\) is invertible and \(\|(S^{*}S)^{-1}\|\leq(sAR^{-d})^{-1}\) in this case. Then Corollary 4 is proved.
In this case, the conditional number
Let \(\delta\rightarrow0\), then Corollary 5 is proved. □
Proof of Corollary 6
By the conclusions of Theorem 1, we have that \(AR^{-d}\|f\| _{2}^{2} \leq\frac{1}{s}\sum_{i=1}^{s}|f(x_{i})|^{2} \leq D_{2}\|f\| _{2}^{2}\) for all \(f \in V_{R,\delta}^{2}(\varLambda,\varPhi)\) hold with probability \(1-\epsilon\) under the conditions in Theorem 1. Therefore we have that
in this case. The corollary is proved. □
7 Reconstruction Algorithm from Non-uniformly Random Sampling
Realistically one can sample \(f\) only on a bounded set; furthermore, every function vanishes at infinity, thus samples far out do not contribute anything significant to sampling and reconstruction. We can learn about \(f\) only if the samples are taken in the “essential support” of \(f\). Let \(\varLambda\) is a finite set so we can write \(\varPhi\) as \(\{\phi _{1},\phi_{2},\ldots,\phi_{i},\ldots,\phi_{n}\}\) where \(n=|\varLambda|\). Suppose the generators are bounded and the support of every \(\phi_{i}\) is in \(C_{R}\). So that for any small \(\delta>0\), we have \(V_{R,\delta}^{2}(\varLambda,\varPhi )=V^{2}(\varLambda,\varPhi)\) since all the energy of these functions is all in \(C_{R}\). In addition, we assume that \(\varPhi\) is a Riesz basis of \(V^{2}(\varLambda,\varPhi)\).
For a signal \(f=\sum_{i=1}^{n}c_{i}\phi_{i} \in V^{2}(\varLambda ,\varPhi)\), denote
to be our sample value,
to be the coefficients, and the matrix
Then we can easily see that \(F=CU\).
Note that \(\varPhi\) is a Riesz basis, we have that \(c_{i}=\langle f,\widetilde{\phi}_{i}\rangle\). So that by Proposition 22 in Appendix, we have
On the other hand, use the same method as Proposition 22 in Appendix, we have
By Corollary 4, for \(\nu\) small enough and \(s\) large enough, we have that
hold with probability at least \(1-\epsilon\) where \(A_{1} = \frac {s}{R^{d}} (\frac{K_{w} \rho}{2} - \nu )\) and \(B_{1} = s D_{2}\). Combining with the arguments above, we have that
where \(A_{2}=\|\widetilde{\varPhi}\|_{2,1,u_{0}}^{-2}\) and \(B_{2}=\| \varPhi\|_{2,1,u_{0}}^{2}\). Therefore
hold with probability at least \(1-\epsilon\). Similar to the arguments which prove Corollary 4 and Corollary 5, we get that \(UU^{*}\) is invertible and the conditional number of \(UU^{*}\) which is denoted by \(\kappa (UU^{*})\) would not larger than \(\frac{B_{1} B_{2}}{A_{1} A_{2}}=\frac {2 R^{d} D_{2} \|\varPhi\|_{2,1,u_{0}}^{2} \|\widetilde{\varPhi}\| _{2,1,u_{0}}^{2}}{K_{w} \rho- 2\nu}\) with probability at least \(1-\epsilon\). That is,
Then one can reconstruct the original function from the sample by
with probability with at least \(1-\epsilon\).
The reconstruction algorithm from non-uniform random sampling is shown in Table 2.
7.1 Numerical Test
Recall the spline functions mentioned in Example 8. Let us consider it as a numerical model for the generators in \(V_{R,\delta }^{2}(\varLambda,\varPhi)\). Let the spline knots be integer points and \(m=3\). We restrict the splines in \(C_{R}\), where \(R\) is an even positive integer. We will review that space by the following definition. Let \(d=1\) and \(R\) be an even positive integer. We use the spline function
where “∗” represents the convolution product. The generators are presented by
Let \(\varLambda=\{\lambda:\phi_{\lambda}\not\equiv0\}\). We can easily check that \(\varLambda=(\frac{1}{2}+\mathbb {Z})\cap[-\frac{R}{2}-1,\frac {R}{2}+1]\) and \({\lambda^{*}=1/2}\).
We will test two kinds of probability distributions in sampling. One is the normal distribution with mean 0 and variance \(\sigma\) like Example 10. In this case, we can get that the optimal choice of \(\sigma\) to optimize the lower bound of sampling inequality is \(1/2\) taking \(\delta \rightarrow0\), so that \(K_{w}=\sqrt{\frac{2}{\pi}}\) and \(\rho=\exp (-\frac{1}{2})\). Note that we actually abandon those sampling points out of \(C_{R}\) in practice. Another distribution is the uniform distribution in \(C_{R}\), just like that used in [3] and [8]. In this case, we have \(K_{w}=1\) and \(\rho=1\). Figure 4 plots the examples of different distributions of sampling points with \(R=8\).
Then we can give a totally practical trial to get the actual success rate. We use the algorithm in Table 2, using the numerical pseudo-inverse in Step 5, just like the “PseudoInverse” function in Wolfram Mathematica or “pinv” function in MATLAB. We take the space \(\varPhi\) as above while the distribution of sampling is normal distribution with a specific \(\sigma\) defined in Example 10. More specific, we take \(c(\lambda)\) periodically of period 8. The signal is shown in Fig. 4 too.
For a test, we will run the algorithm and calculate the Euclidean distance between the \(C\) we calculated numerically in Step 7 and the real \(C\) of the original function \(f\) which is regarded as the reconstruction error. One test is regard as successful test when the reconstruction error is less than \(10^{-9}\).
Firstly, we take into two parts: one is using the normal distribution mentioned above with \(\sigma=1/2\), taking \(R=8,16,24\) and \(s=20,30,\ldots,100\) respectively; another is using the uniform distribution in \(C_{R}\) taking the same parameters. We take 10000 trials for each case, then test the actual success rate in those cases. Figure 5 plots the success rate respectively to every value of \(s\) in each case.
From Fig. 5, we can see that: (i) the success rate increases with \(s\) increases; (ii) the success rate decreases with \(R\) increases and other conditions unchanged; (iii) due to the nonuniformness of the distribution, we need more sampling points or less \(R\) to reach the same success rate comparing with the uniform distribution. The results of the numerical tests match with Theorem 1 and Corollary 5.
Secondly, we fix \(R=8\), take \(s=40,60\) as well as \(\sigma =1/80,2/80,\ldots,120/80\) respectively and take 10000 trials for each case with the same normal distribution, then test the actual success rate in those cases. Figure 6 shows the success rate with the \(\sigma\)’s.
From Fig. 6, we can see that for a certain \(s\), the success rate first increases then decreases with \(\sigma\) increasing. We get the best choice that optimize the actual success rate is that \(\sigma=40/80=1/2\) by the data of \(s=60\). At the same time, we can improve the success rate by increasing sampling points under the same \(\sigma\). The results of the numerical tests match with Theorem 1 and Corollary 5.
Next, we explore the “threshold” of \(s\), or the smallest \(s\) to have the actual sampling success rate reached \(95\%\) for different \(R\)’s. We fix \(\sigma=1/2\) and \(R=8,16,32,64\) respectively, take different values of sampling rates around the threshold we pre-estimated while taking 10000 trials for each case of sampling rate, and look for the “threshold” of \(s\). Table 3 is the thresholds with those \(R\)’s.
We can see that every value of the threshold is a large twice times than the front value from Table 3. So the \(O(R^{d} \log R^{d})\) sampling rate in Theorem 1 is match with numerical test.
8 Conclusions
In this paper, we considered the non-uniform random sampling and reconstruction problem on signal spaces with finite rate of innovation. Under very mild assumptions on the generators and the probability distribution function, we show that for \(R\) sufficiently large, taking \(O(R^{d} \log(R^{d}))\) many random samples yields a sampling set for \(V^{2}_{R,\delta}(\varPhi)\) with high probability. As one of applications of our main results, we provided a reconstruction algorithm from non-uniform random sampling with high probability and implemented it when the generators are compact-supported.
References
Aldroubi, A., Gröchenig, K.: Nonuniform sampling and reconstruction in shift-invariant spaces. SIAM Rev. 43(4), 585–620 (2001)
Baechler, G., Scholefield, A., Baboulaz, L., Vetterli, M.: Sampling and exact reconstruction of pulses with variable width. IEEE Trans. Signal Process. 65(10), 2629–2644 (2017)
Bass, R.F., Gröchenig, K.: Relevant sampling of band-limited functions. Ill. J. Math. 57(1), 43–58 (2013)
Bernardi, O., Giménez, O.: A linear algorithm for the random sampling from regular languages. Algorithmica 62(1–2), 130–145 (2012)
Bi, N., Nashed, M.Z., Sun, Q.Y.: Reconstructing signals with finite rate of innovation from noisy samples. Acta Appl. Math. 107(1–3), 339–372 (2009)
Fang, K.T., Ma, C.X., Winker, P.: Centered \(L_{2}\)-discrepancy of random sampling and Latin hypercube design, and construction of uniform designs. Math. Comput. 71(237), 275–296 (2002)
Führ, H., Xian, J.: Quantifying invariance properties of shift-invariant spaces. Appl. Comput. Harmon. Anal. 36, 514–521 (2014)
Führ, H., Xian, J.: Relevant sampling in finitedly generated shift-invariant spaces. J. Approx. Theory 240, 1–15 (2019)
Gröchenig, K., Schwab, H.: Fast local reconstruction methods for nonuniform sampling in shift-invariant spaces. SIAM J. Matrix Anal. Appl. 24(4), 899–913 (2003)
Li, Y.X., Wen, J.M., Xian, J.: Reconstruction from convolution random sampling in local shift invariant spaces. Inverse Problems (2019). https://iopscience.iop.org/article/10.1088/1361-6420/ab40f7/meta
Moskowitz, B., Caflisch, R.E.: Smoothness and dimension reduction in quasi-Monte Carlo methods. Math. Comput. Model. 23(8), 37–54 (1996)
Mulleti, S., Seelamantula, C.S.: Ellipse fitting using the finite rate of innovation sampling principle. IEEE Trans. Image Process. 25(3), 1451–1464 (2016)
Pedergnana, M., García, S.G.: Smart sampling and incremental function learning for very large high dimensional data. Neural Netw. 78, 75–87 (2016)
Romero, V.J., Burkardt, J.V., Gunzburger, M.D., Peterson, J.S.: Comparison of pure and “Latinized” centroidal Voronoi tessellation against various other statistical sampling methods. Reliab. Eng. Syst. Saf. 91(10–11), 1266–1280 (2006)
Rudresh, S., Seelamantula, C.S.: Finite-rate-of-innovation-based super-resolution radar imaging. IEEE Trans. Signal Process. 65(19), 5021–5033 (2017)
Schumaker, L.L.: Spline Functions: Basic Theory. John Wiley & Sons, New York (1981)
Sun, Q.Y.: Nonuniform average sampling and reconstruction of signals with finite rate of innovation. SIAM J. Math. Anal. 38(5), 1389–1422 (2006)
Sun, Q.Y.: Wiener’s lemma for infinite matrices. Trans. Am. Math. Soc. 359(7), 3099–3123 (2007)
Sun, Q.Y.: Frames in spaces with finite rate of innovation. Adv. Comput. Math. 28(4), 301–329 (2008)
Sun, W.C.: Local sampling theorems for spaces generated by splines with arbitrary knots. Math. Comput. 78(265), 225–239 (2009)
Sun, W.C., Zhou, X.W.: Characterization of local sampling sequences for spline subspaces. Adv. Comput. Math. 30(2), 153–175 (2009)
Tropp, J.A.: User-friendly tail bounds for sums of random matrices. Found. Comput. Math. 12(4), 389–434 (2012)
Vetterli, M., Marziliano, P., Blu, T.: Sampling signals with finite rate of innovation. IEEE Trans. Signal Process. 50(6), 1417–1428 (2002)
Xian, J., Li, S.: Sampling set conditions in weighted finitely generated shift-invariant spaces and their applications. Appl. Comput. Harmon. Anal. 23(2), 171–180 (2007)
Yang, J.B.: Random sampling and reconstruction in multiply generated shift-invariant spaces. Anal. Appl. 17(2) 323–347 (2019)
Acknowledgements
The authors would like to thank the reviewers for their valuable comments and suggestions that led to the improvement of this paper. The corresponding author Jun Xian is partially supported by the National Natural Science Foundation of China (11422102, 11631015, 11871481); the Guangdong Provincial Government of China through the Computational Science Innovative Research Team program, China; and the Guangdong Province Key Laboratory of Computational Science, China.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix. Some Properties of the Signal Spaces \(V^{2}(\varLambda,\varPhi)\)
Appendix. Some Properties of the Signal Spaces \(V^{2}(\varLambda,\varPhi)\)
We assume that \(\varPhi:=\{\phi_{\lambda}\}_{\lambda\in\varLambda}\) is a Riesz basis of \(V^{2}(\varLambda,\varPhi)\), therefore there exists a dual Riesz basis \(\widetilde{\varPhi}=\{\widetilde{\phi }_{\lambda}\}_{\lambda\in\varLambda}\). Thus we have that for all \(f \in V^{2}(\varLambda,\varPhi)\), such that
We can easily see that for all \(1\leq q' \leq q \leq\infty\), \(\infty \geq p' \geq p \geq1\) and \(u'(x)\leq u(x), \forall x \in \mathbb {R}^{d}\), we have \(\|\varPhi\|_{q',p',u'}\leq\|\varPhi\|_{q,p,u}\). Moreover, by [19, Theorem 4.1], we have that if \({\|\varPhi\| _{q,p,u}<\infty}\), then so is the same norm of the dual Riesz basis, or to say \(\|\widetilde{\varPhi}\|_{q,p,u}<\infty\).
For convenience of the main proof, we will show some properties of the space \(V^{2}(\varLambda,\varPhi)\) under the assumptions of the generators in Sect. 2.2.
Proposition 22
If \(f \in{\mathrm{L}}^{2}(\mathbb {R}^{d})\), \(\varPhi=\{\phi_{\lambda}\} \)satisfying (A.0) and \(\|\varPhi\|_{2,1,u_{0}}<\infty\), then we have:
and
Proof
We firstly prove (22). In fact, we have
A similar method is used in the proof of [18, Theorem 2.4(iii)]. Similar to the proof above, we can prove (23) as well. □
Proposition 23
For every \(f \in V^{2}(\varLambda,\varPhi)\)and every subset \(\varGamma\subset \mathbb {R}^{d}\)where the absolute covering index \(N_{0}(\varGamma)\)defined in (3) is finite, we have
Proof
First we consider \(\varGamma\) with \(N_{0}(\varGamma)=1\), that is, for any \(k\in \mathbb {Z}^{d}\), there is at most one point in \(\varGamma\cap (k+[0,1]^{d})\). Let \(c(\lambda)=\langle f,\widetilde{\phi}_{\lambda}\rangle\), then \(f=\sum_{\lambda\in\varLambda}c(\lambda)\phi _{\lambda}\). Similar to the proof of Proposition 22, we have
where the last inequality is derived from the conclusion of Proposition 22.
For a general subset \(\varGamma\), we can split it into at most \(N_{0}(\varGamma)\) non-intersect parts each of which has absolute covering index 1. Then sum the function value of each part respectively, sum all of them all together and we get the final conclusion. □
Proposition 24
\(V^{2}(\varLambda,\varPhi)\)is a reproducing kernel space, or to say there exists a family \((v_{x})_{x\in \mathbb {R}^{d}}\subset V^{2}(\varLambda,\varPhi)\)such that \(f(x)=\langle f,v_{x}\rangle ,\forall f\in V^{2}(\varLambda,\varPhi)\)and \(\|v_{x}\|_{2}^{2}\leq D_{2}\)almost everywhere. Moreover, we have \(\|f\|_{\infty}^{2}\leq D_{2}\|f\|_{2}^{2}\)for all \(f\in V^{2}(\varLambda,\varPhi)\).
Proof
We define that \(v_{x}(y)=\sum_{\lambda\in\varLambda}\widetilde{\phi }_{\lambda}(y)\overline{\phi_{\lambda}(x)}\), and \(w_{x}(y)=\sum_{\lambda \in\varLambda}\phi_{\lambda}(y)\overline{\widetilde{\phi}_{\lambda}(x)}\). For all \(f\in V^{2}(\varLambda,\varPhi)\), we have that
Similarly, we have
Therefore, \(\langle f,w_{x}-v_{x}\rangle=0\) holds for all \(f\in V^{2}(\varLambda,\varPhi)\). However, both \(v_{x}\) and \(w_{x}\) are in \(V^{2}(\varLambda,\varPhi)\) by definition, then \(w_{x}-v_{x}\in V^{2}(\varLambda,\varPhi)\). So \(\langle w_{x}-v_{x},w_{x}-v_{x}\rangle=0\), or \({w_{x}-v_{x}=0}\). Therefore \(w_{x}=v_{x}\), which implies \(v_{x}(y)=\sum_{\lambda\in\varLambda }\phi_{\lambda}(y)\overline{\widetilde{\phi}_{\lambda}(x)}\) as well.
We now show that \(v(x)_{x\in \mathbb {R}^{d}}\) is a reproducing kernel. For fixed \(x\in \mathbb {R}^{d}\), we have
By \(\|\varPhi\|_{\infty,1,u_{0}}< \infty\), we can change the order of the sum and the integration. So we can obtain that
Thus \(\mathrm{ess}\,\mathrm{sup}_{x\in \mathbb {R}^{d}}\|v_{x}\|_{2}^{2}\leq D_{2}<\infty\). Therefore \((v_{x})_{x\in \mathbb {R}^{d}}\) is a reproducing kernel of \(V^{2}(\varPhi,\varLambda)\). Moreover, we have
□
Rights and permissions
About this article
Cite this article
Lu, Y., Xian, J. Non-uniform Random Sampling and Reconstruction in Signal Spaces with Finite Rate of Innovation. Acta Appl Math 169, 247–277 (2020). https://doi.org/10.1007/s10440-019-00298-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10440-019-00298-6
Keywords
- Random sampling
- Non-uniform sampling
- Spaces with finite rate of innovation
- Non-uniform distribution
- Reconstruction algorithm