Abstract
We construct minimax optimal non-asymptotic confidence sets for low rank matrix recovery algorithms such as the Matrix Lasso or Dantzig selector. These are employed to devise adaptive sequential sampling procedures that guarantee recovery of the true matrix in Frobenius norm after a data-driven stopping time \(\hat n\) for the number of measurements that have to be taken. With high probability, this stopping time is minimax optimal. We detail applications to quantum tomography problems where measurements arise from Pauli observables. We also give a theoretical construction of a confidence set for the density matrix of a quantum state that has optimal diameter in nuclear norm. The non-asymptotic properties of our confidence sets are further investigated in a simulation study.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
18.1 Introduction
18.1.1 Uncertainty Quantification in Compressed Sensing
Compressed sensing and related convex relaxation algorithms have had a profound impact on high-dimensional statistical modelling in recent years. They provide efficient recovery of low-dimensional objects that sit within a high-dimensional ill-posed system of linear equations. Prototypical low-dimensional structures are described by sparsity and low rank hypotheses. The statistical analysis of such algorithms has mostly been concerned with recovery rates, or with the closely related question of how many measurements are sufficient to reach a prescribed recovery level—some key references are Refs. [3, 5,6,7, 16, 24, 25, 33].
A statistical question of fundamental importance that has escaped a clear answer so far is the question of uncertainty quantification: Can we tell from the observations how well the algorithm has worked? In technical terms: can we report confidence sets for the unknown parameter? Or, in the sequential sampling setting, can we give data-driven rules that ensure recovery of the true parameter at a given precision? Answers to this question are of great importance in various applications of compressed sensing. For one-dimensional subproblems, such as projection onto a fixed coordinate of the parameter vector, recent advances have provided some useful confidence intervals (see Refs. [8, 9, 22, 36]), but our understanding of valid inference procedures for the entire parameter remains limited.
Whereas the ‘estimation theory’ for compressed sensing is quite similar for sparsity and low rank constraints, this is not so for the theory of confidence sets. On the one hand, if one is interested in inference on the full parameter, sparsity conditions induce information theoretic barriers, as shown in Ref. [30]: unless one is willing to make additional signal strength assumptions (inspired by the literature on nonparametric confidence sets, such as Refs. [13, 19]), a uniformly valid confidence set for the unknown parameter vector θ cannot have a better performance than \(1/\sqrt {n}\) in quadratic loss. This significantly falls short of the optimal recovery rates \((k \log p)/n\) for the most interesting sparsity levels k. On the other hand, and perhaps surprisingly, we will show in this article that the low-rank constraint is naturally compatible with certain risk estimation approaches to confidence sets. This will be seen to be true for general sub-Gaussian sensing matrices, but also for sensing matrices arising from Pauli observables, as is specifically relevant in quantum state tomography problems (see the next section). In the latter case it will be helpful to enforce the additional ‘quantum state shape constraint’ on the unknown matrix to obtain optimal results. One can conclude that, in contrast to ‘sparse models’, no signal strength assumptions are necessary for the existence of adaptive confidence statements in low rank recovery problems. Our findings are confirmed in a simple simulation study, see Sect. 18.4.
The honest non-asymptotic confidence sets we will derive below can be used for the construction of adaptive sampling procedures: An experimenter wants to know—at least with a prescribed probability of 1 − α—that the matrix recovery algorithm, the ‘estimator’, has produced an output \(\tilde \theta \) that is close to the ‘true state’ θ. The sequential protocol advocated here—which is related to ‘active learning algorithms’ from machine learning, e.g., Ref. [29]—should tell the experimenter whether a new batch of measurements has to be taken to decrease the recovery error, or whether the collected observations are already sufficient. The data-driven stopping time for this protocol should not exceed the minimax optimal stopping time, again with high probability. We shall show that for Pauli and sub-Gaussian sensing ensembles, such algorithms exist under mild assumptions on the true matrix θ. These assumptions are in particular always satisfied under the ‘quantum shape constraint’ that naturally arises in quantum tomography problems.
Our results depend on the choice of the Frobenius norm and the Hilbert space geometry induced by it. For other natural matrix norms, such as for instance the trace-(nuclear) norm, the theory is more difficult. We show as a first step that at least theoretically a trace-norm optimal confidence set can be constructed for the unknown quantum state (Theorem 18.4)—this suggests interesting directions for future research.
18.1.2 Application to Quantum State Estimation
This work was partly motivated by a problem arising in present-day physics experiments that aim at estimating quantum states. Conceptually, a quantum mechanical experiment involves two stages (c.f. Fig. 18.1): A source (or preparation procedure) that emits quantum mechanical systems with unknown properties, and a measurement device that interacts with incoming quantum systems and produces real-valued measurement outcomes, e.g. by pointing a dial to a value on a scale. Quantum mechanics stipulates that both stages are completely described by certain matrices.
The properties of the source are represented by a positive semi-definite unit trace matrix θ, the quantum state, also referred to as density matrix. In turn, the measurement device is modelled by a Hermitian matrix X, which is referred to as an observable in physics jargon. A key axiom of the quantum mechanical formalism states that if the measurement X is repeatedly performed on systems emitted by the source that is preparing θ, then the real-valued measurement outcomes will fluctuate randomly with expected value
(referred to as expectation value in the quantum physics literature). The precise way in which physical properties are represented by these matrices is immaterial to our discussion (cf. any textbook, e.g. Ref. [32]). We merely note that, while in principle any Hermitian X can be measured by some physical apparatus, the required experimental procedures are prohibitively complicated for all but a few highly structured matrices. This motivates the introduction of Pauli designs below, which correspond to fairly tractable ‘spin parity measurements’.
The quantum state estimation or quantum state tomography Footnote 1 problem is to estimate an unknown density matrix θ from the measurement of a collection of observables X 1, …, X n. This task is of particular importance to the young field of quantum information science [31]. There, the sources might be a carefully engineered component used for technological applications such as quantum key distribution or quantum computing. In this context, quantum state estimation is the process of characterising the components one has built—clearly an important capability for any technology.
A major challenge lies in the fact that relevant instances are described by d × d-matrices for fairly large dimensions d ranging from 100 to 10,000 in presently performed experiments [18]. Such high-dimensional estimation problems can benefit substantially from structural properties of the objects to be recovered. Fortunately, the density matrices occurring in quantum information experiments are typically well-approximated by matrices of low rank r ≪ d. In fact, in the practically most important applications, one usually even aims at preparing a state of rank one—a so-called pure quantum state. While environmental noise will drive the actual state away from the perfect rank-one case, the error will usually be small.
As a result, quantum physicists have early on shown an interest in low-rank matrix recovery methods [12, 15,16,17, 28]. Initial works [15, 16] focused on the minimal number n of observables X 1, …, X n required for reconstructing a rank-r density matrix θ in the noiseless case, i.e. under the idealised assumption that the expectation values tr(θX i) are known exactly. The practically highly relevant problem of quantifying the uncertainty of an estimate \(\hat \theta \) arising from noisy observations on low-rank states was addressed only later [12] and remains less well understood.
More concretely, the basic approach taken in Ref. [12] for uncertainty quantification is similar to the one pursued in the present paper. In a first step, one uses a Matrix Lasso or Dantzig Selector to construct an estimate. Then, a confidence region is obtained by comparing predictions derived from the initial estimate to new samples. However, Ref. [12] suffers from two demerits. First, and most importantly, the performance analysis of the scheme relies on a bound on the rank r of the unknown true θ. Such a bound is not available in practice. Second, the dependence of the rate on r is not tight. Both of these demerits will be addressed here.
We close this section pointing to more broadly related works. Uncertainty quantification in quantum state tomography in general has been treated by numerous authors—a highly incomplete list is Refs. [2, 4, 10, 34, 35]. However, the concept of dimension reduction for low-rank states does not feature explicitly in these papers. This contrasts with Ref. [17], where the authors propose model selection techniques based on information criteria to arrive at low-rank estimates. The use of general-purpose methods—like maximum likelihood estimation and the Akaike Information Criterion—in Ref. [17] means that it is applicable to very general experimental designs. In contrast to this, the present paper relies on compressed sensing ideas to arrive at rigorous a priori guarantees on statistical and computational performance. Also, it remains non-obvious how such model selection steps can be transformed into ‘post-model selection’ confidence sets—typically such constructions result in sub-optimal signal strength conditions that ensure model selection consistency (see Ref. [26] and also the discussion after Theorem 2 in Ref. [30]). Our confidence procedures never estimate the unknown rank of the quantum state—not even implicitly. Rather, they estimate the performance of a dimension-reduction technique directly based on sample splitting.
18.2 Matrix Compressed Sensing
We consider inference on a d × d matrix θ that is symmetric, or, if it consists of possibly complex entries, assumed to be Hermitian (that is θ = θ ∗ where θ ∗ is the conjugate transpose of θ). Denote by \(\mathbb M_d(\mathbb K)\) the space of d × d matrices with entries in \(\mathbb K = \mathbb C\) or \(\mathbb K=\mathbb R\). We write ∥⋅∥F for the usual Frobenius norm on \(\mathbb M_d(\mathbb K)\) arising from the inner product tr(AB) = 〈A, B〉F. Moreover let \(\mathbb H_d(\mathbb C)\) be the set of all Hermitian matrices, and \(\mathbb H_d(\mathbb R)\) for the set of all symmetric d × d matrices with real entries. The norm symbol ∥⋅∥ without subindex denotes the standard Euclidean norm on \(\mathbb R^n\) or on \(\mathbb C^n\) arising from the Euclidean inner product 〈⋅, ⋅〉.
We denote the usual operator norm on \(\mathbb M_d(\mathbb C)\) by ∥⋅∥op. For \(M\in \mathbb M_d(\mathbb C)\) let \((\lambda _k^2: k =1, \dots , d)\) be the eigenvalues of M TM (which are all real-valued and positive). The l 1-Schatten, trace, or nuclear norm of M is defined as
Note that for any matrix M of rank 1 ≤ r ≤ d the following inequalities are easily shown,
We will consider parameter subspaces of \(\mathbb H_d(\mathbb C)\) described by low rank constraints on θ, and denote by R(k) the space of all Hermitian d × d matrices that have rank at most k, k ≤ d. In quantum tomography applications, we may assume an additional ‘shape constraint’, namely that θ is a density matrix of a quantum state, and hence contained in state space
where θ ≽ 0 means that θ is positive semi-definite. In fact, in most situations, we will only require the bound \(\|\theta \|{ }_{S_1} \le 1\) which trivially holds for any θ in Θ+.
We have at hand measurements arising from inner products 〈X i, θ〉F = tr(X iθ), i = 1, …, n, of θ with d × d (random) matrices X i. This measurement process is further subject to independent additive noise ε. Formally, the measurement model is
where the ε i’s and X i’s are independent of each other. We write Y = (Y 1, …, Y n)T, and for probability statements under the law of Y, X, ε given fixed θ we will use the symbol \(\mathbb P_\theta \). Unless mentioned otherwise we will make the basic assumption of Gaussian noise
where σ > 0 is known. See Remark 18.6 for some discussion of the unknown variance case. In the context of quantum mechanics, the inner product tr(X iθ) gives the expected value of the observable X i when measured on a system in state θ (cf. Sect. 18.1.2). A class of physically realistic measurements (correlations among spin-1∕2 particles) is described by X i’s drawn from the Pauli basis. Our main results also hold for measurement processes of this type. Before we describe this in Sect. 18.2.2, let us first discuss our assumptions on the matrices X i.
18.2.1 Sensing Matrices and the RIP
When \(\theta \in \mathbb M_d(\mathbb R)\), we shall restrict to design matrices X i that have real-valued entries, too, and when \(\theta \in \mathbb H_d(\mathbb C)\) we shall consider designs where \(X^i \in \mathbb H_d(\mathbb C)\). This way, in either case, the measurements tr(X iθ)’s and hence the Y i’s are all real-valued. More concretely, the sensing matrices X i that we shall consider are described in the following assumption, which encompasses both a prototypical compressed sensing setting—where we can think of the matrices X i as i.i.d. draws from a Gaussian ensemble (X m,k) ∼iidN(0, 1)—as well as the ‘random sampling from a basis of \(\mathbb M_d(\mathbb C)\)’ scenario. The systematic study of the latter has been initiated by quantum physicists [15, 28], as it contains, in particular, the case of Pauli basis measurements [12, 16] frequently employed in quantum tomography problems. Note that in Part (a) the design matrices are not Hermitian but our results can easily be generalised to symmetrised sub-Gaussian ensembles (as those considered in Ref. [24]).
Condition 18.1
-
(a)
\(\theta \in \mathbb H_d(\mathbb R)\), ‘isotropic’ sub-Gaussian design: The random variables \((X^i_{m,k})\) , 1 ≤ m, k ≤ d, i = 1, …, n, generating the entries of the random matrix X i are i.i.d. distributed across all indices i, m, k with mean zero and unit variance. Moreover, for every \(\theta \in \mathbb M_d(\mathbb R)\) such that ∥θ∥F ≤ 1, the real random variables Z i = tr(X iθ) are sub-Gaussian: for some fixed constants τ i > 0 independent of θ,
$$\displaystyle \begin{aligned}\mathbb E e^{\lambda Z_i} \le \tau_1 e^{\lambda^2 \tau_2^2} ~\forall \lambda \in \mathbb R.\end{aligned}$$ -
(b)
\(\theta \in \mathbb H_d(\mathbb C)\), random sampling from a basis (‘Pauli design’): Let \(\{E_1, \dots , E_{d^2}\} \subset \mathbb H_d(\mathbb C)\) be a basis of \(\mathbb M_d(\mathbb C)\) that is orthonormal for the scalar product 〈⋅, ⋅〉F and such that the operator norms satisfy, for all i = 1, …, d 2 ,
$$\displaystyle \begin{aligned}\|E_i\|{}_{op} \le \frac{K}{\sqrt{d}},\end{aligned}$$for some universal ‘coherence’ constant K. [In the Pauli basis case we have K = 1.] Assume the X i, i = 1, …, n, are draws from the finite family \(\mathcal E = \{d E_i: i=1, \dots , d^2\}\) sampled uniformly at random.
The above examples all obey the matrix restricted isometry property, that we describe now. Note first that if \(\mathcal X: \mathbb R^{d\times d} \to \mathbb R^n \) is the linear ‘sampling’ operator
so that we can write the model equation (18.3) as \(Y=\mathcal X \theta + \varepsilon \), then in the above examples we have the ‘expected isometry’
Indeed, in the isotropic design case we have
and in the ‘basis case’ we have, from Parseval’s identity and since the X i’s are sampled uniformly at random from the basis,
The restricted isometry property (RIP) then requires that this ‘expected isometry’ actually holds, up to constants and with probability ≥ 1 − δ, for a given realisation of the sampling operator, and for all d × d matrices θ of rank at most k:
where τ n(k) are some constants that may depend, among other things, on the rank k and the ‘exceptional probability’ δ. For the above examples of isotropic and Pauli basis design inequality (18.7) can be shown to hold with
where
for some η > 0 denotes a ‘polylog function’, and where c = c(δ) is a constant. See Refs. [6, 28] for these results, where it is also shown that c(δ) can be taken to be at least O(1∕δ 2) as δ → 0 (sufficient for our purposes below).
18.2.2 Quantum Measurements
Here, we introduce a paradigmatic set of quantum measurements that is frequently used in both theoretical and practical treatments of quantum state estimation (e.g. [16, 18]). For a more general account, we refer to standard textbooks [20, 31]. The purpose of this section is to motivate the ‘Pauli design’ case (Condition 18.1(b) of the main theorem, as well as the approximate Gaussian noise model. Beyond this, the technical details presented here will not be used.
18.2.2.1 Pauli Spin Measurements on Multiple Particles
We start by describing ‘spin measurements’ on a single ‘spin-1∕2 particle’. Such a measurement corresponds to the situation of having d = 2. Without worrying about the physical significance, we accept as fact that on such particles, one may measure one of three properties, referred to as the ‘spin along the x, y, or z-axis’ of \(\mathbb R^3\). Each of these measurements may yield one of two outcomes, denoted by + 1 and − 1 respectively.
The mathematical description of these measurements is derived from the Pauli matrices
in the following way. Recall that the Pauli matrices have eigenvalues ± 1. For x ∈{1, 2, 3} and j ∈{+1, −1}, we write \(\psi _j^x\) for the normalised eigenvector of σ x with eigenvalue j. The spectral decomposition of each Pauli spin matrix can hence be expressed as
with
denoting the projectors onto the eigenspaces. Now, a physical measurement of the ‘spin along direction x’ on a system in state θ will give rise to a {−1, 1}-valued random variable C x with
where \(\theta \in \mathbb H_2(\mathbb C)\). Using Eq. (18.10), this is equivalent to stating that the expected value of C x is given by
Next, we consider the case of joint spin measurements on a collection of N particles. For each, one has to decide on an axis for the spin measurement. Thus, the joint measurement setting is now described by a word x = (x 1, …, x N) ∈{1, 2, 3}N. The axioms of quantum mechanics posit that the joint state θ of the N particles acts on the tensor product space \((\mathbb C^2)^{\otimes N}\), so that \(\theta \in \mathbb H_{2^N}(\mathbb C)\).
Likewise, the measurement outcome is a word j = (j 1, …, j N) ∈{1, −1}N, with j i the value of the spin along axis x i of particle i = 1, …, N. As above, this prescription gives rise to a {1, −1}N-valued random variable C x. Again, the axioms of quantum mechanics imply that the distribution of C x is given by
Note that the components of the random vector C x are not necessarily independent, as θ will generally not factorise
It is often convenient to express the information in Eq. (18.14) in a way that involves tensor products of Pauli matrices, rather than their spectral projections. In other words, we seek a generalisation of Eq. (18.13) to N particles. As a first step toward this goal, let
be the parity function. Then one easily verifies
In this sense, the tensor product \(\sigma ^{x_1}\otimes \dots \otimes \sigma ^{x_N}\) describes a measurement of the parity of the spins along the respective directions given by x.
In fact, the entire distribution of C x can be expressed in terms of tensor products of Pauli matrices and suitable parity functions. To this end, we extend the definitions above. Write
for the identity matrix in \(\mathbb M_2(\mathbb C)\). For every subset S of {1, …, N}, define the ‘parity function restricted to S’ via
Lastly, for S ⊂{1, …, N} and x ∈{1, 2, 3}N, the restriction of x to S is
Then for every such x, S one verifies the identity
In other words, the distribution of C x contains enough information to compute the expectation value of all observables \((\sigma ^{x^S_1} \otimes \dots \otimes \sigma ^{x^S_N})\) that can be obtained by replacing the Pauli matrices on an arbitrary subset S of particles by the identity σ 0. The converse is also true: the set of all such expectation values allows one to recover the distribution of C x. The explicit formula reads
and can be verified by direct computation.Footnote 2
In this sense, the information obtainable from joint spin measurements on N particles can be encoded in the 4N real numbers
Indeed, every such y arises as y = x S for some (generally non-unique) combination of x and S. This representation is particularly convenient from a mathematical point of view, as the collection of matrices
forms an ortho-normal basis with respect to the 〈⋅, ⋅〉F inner product. Thus the terms in Eq. (18.22) are just the coefficients of a basis expansion of the density matrix θ.Footnote 3
From now on, we will use Eq. (18.22) as our model for quantum tomographic measurements. Note that the E y satisfy Condition 18.1(b) with coherence constant K = 1 and d = 2N.
18.2.2.2 Bernoulli Errors and Pauli Observables
In the model (18.3) under Condition 18.1(b) we wish to approximate d ⋅tr(E yθ) for a fixed observable E y (we fix the random values of the X i’s here) and for d = 2N. If y = x S for some setting x and subset S, then the parity function B y := χ S(C x) has expected value \(2^{N/2} \cdot \mathrm {tr}(E^y \theta )=\sqrt {d} \cdot tr(E^y \theta )\) (see Eqs. (18.20) and (18.23)), and itself is a Bernoulli variable taking values {1, −1} with
Note that
so indeed p ∈ [0, 1] and the variance satisfies
This is the error model considered in Ref. [12].
In order to estimate all Y i, i = 1, …, n, for given E i := E y, a total number nT of identical preparations of the quantum state θ are being performed, divided into batches of T Bernoulli variables \(B_{i,j} := B^y_j, j=1, \dots , T\). The measurements of the sampling model Eq. (18.3) are thus
where
is the effective error arising from the measurement procedure making use of T preparations to estimate each quantum mechanical expectation value. Now note that
We see that since the ε i’s are themselves sums of independent random variables, an approximate Gaussian error model with variance σ 2 will be roughly appropriate. If T ≥ n then \(\sigma ^2 = \mathbb E \varepsilon _1^2\) is no greater than d∕n, and if in addition T ≥ d 2 then all results in Sect. 18.3 below can be proved for this Bernoulli noise model too, see Remarks 18.5 and 18.6 for details.
18.2.3 Minimax Estimation Under the RIP
Assuming the matrix RIP to hold and Gaussian noise ε, one can show that the minimax risk for recovering a Hermitian rank k matrix is
where ≃ denotes two-sided inequality up to universal constants.
For the upper bound one can use the nuclear norm minimisation procedure or matrix Dantzig selector from Candès and Plan [6], and needs n to be large enough so that the matrix RIP holds with τ n(k) < c 0 where c 0 is a small enough numerical constant. Such an estimator \(\tilde \theta \) then satisfies, for every θ ∈ R(k) and those \(n \in \mathbb N\) for which τ n(k) < c 0,
with probability greater than 1 − 2δ, and with the constant D(δ) depending on δ and also on c 0 (suppressed in the notation). Note that the results in Ref. [6] use a different scaling in sample size in their Theorem 2.4, but eq. (II.7) in that reference explains that this is just a question of renormalisation. The same result holds for randomly sampled ‘Pauli bases’, see Ref. [28] (and take note of the slightly different normalisation in the notation there, too), and also for the Bernoulli noise model from Sect. 18.2.2.2, see Ref. [12].
A key interpretation for quantum tomography applications is that, instead of having to measure all n = d 2 basis coefficients tr(E iθ), i = 1, …, d 2, a number
of randomly chosen basis measurements is sufficient to reconstruct θ in Frobenius norm loss (up to a small error). In situations where d is large compared to k such a gain can be crucial.
Remark 18.1 (Uniqueness)
It is worth noting that in the absence of errors, so when \(Y_0=\mathcal X \theta _0\) in terms of the sampling operator of Eq. (18.4), the quantum shape constraint ensures that under a suitable RIP condition, only the single matrix θ 0 is compatible with the data. More specifically, let \(Y_0=\mathcal X\theta _0\) for some θ 0 ∈ Θ+ of rank k, and assume that \(\mathcal X\) satisfies RIP with \(\tau _n(4k)<\sqrt {2}-1\). Then
This is a direct consequence of Theorem 3.2 in Ref. [33], which states that if RIP is satisfied with \(\tau _n(4k)<\sqrt {2}-1\) and \(Y_0=\mathcal X\theta _0\), the unique solution of
is given by θ 0. If θ 0 ∈ Θ+, then the minimisation can be replaced by (compare also Ref. [23]).
giving rise to the above remark. This observation further signifies the role played by the quantum shape constraint.
18.3 Uncertainty Quantification for Low-Rank Matrix Recovery
We now turn to the problem of quantifying the uncertainty of estimators \(\tilde \theta \) that satisfy the risk bound (18.26). In fact the confidence sets we construct could be used for any estimator of θ, but the conclusions are most interesting when used for minimax optimal estimators \(\tilde \theta \). For the main flow of ideas we shall assume ε = (ε 1, …, ε n)T ∼ N(0, σ 2I n) but the results hold for the Bernoulli measurement model from Sect. 18.2.2.2 as well—this is summarised in Remark 18.5.
From a statistical point of view, we phrase the problem at hand as the one of constructing a confidence set for θ: a data-driven subset C n of \(\mathbb M_d(\mathbb C)\) that is ‘centred’ at \(\tilde \theta \) and that satisfies
for a chosen ‘coverage’ or significance level 1 − α, and such that the Frobenius norm diameter |C n|F reflects the accuracy of estimation, that is, it satisfies, with high probability,
In particular such a confidence set provides, through its diameter |C n|F, a data-driven estimate of how well the algorithm has recovered the true matrix θ in Frobenius-norm loss, and in this sense provides a quantification of the uncertainty in the estimate.
In the situation of an experimentalist this can be used to decide sequentially whether more measurements should be taken (to improve the recovery rate), or whether a satisfactory performance has been reached. Concretely, if for some 𝜖 > 0 a recovery level \(\|\tilde \theta - \theta \|{ }_F \le \epsilon \) is desired for an estimator \(\tilde \theta \), then assuming \(\tilde \theta \) satisfies the minimax optimal risk bound dk∕n from (18.26), we expect to need, ignoring constants,
measurements. Note that we also need the RIP to hold with τ n(k) from (18.8) less than a small constant c 0, which requires the same number of measurements, increased by a further poly-log factor of d (and independently of σ).
Since the rank k of θ remains unknown after estimation we cannot obviously guarantee that the recovery level 𝜖 has been reached after a given number of measurements. A confidence set C n for \(\tilde \theta \) provides such certificates with high probability, by checking whether |C n|F ≤ 𝜖, and by continuing to take further measurements if not. The main goal is then to prove that a sequential procedure based on C n does not require more than approximately
samples (with high probability). We construct confidence procedures in the following subsections that work with at most as many measurements, for the designs from Condition 18.1.
18.3.1 Adaptive Sequential Sampling
Before we describe our confidence procedures, let us make the following definition, where we recall that R(k) denotes the set of d × d Hermitian matrices of rank at most k ≤ d.
Definition 18.1
Let 𝜖 > 0, δ > 0 be given constants. An algorithm \(\mathcal A\) returning a d × d matrix \(\hat \theta \) after \(\hat n \in \mathbb N\) measurements in model (18.3) is called an (𝜖, δ)—adaptive sampling procedure if, with \(\mathbb P_\theta \)-probability greater than 1 − δ, the following properties hold for every θ ∈ R(k) and every 1 ≤ k ≤ d:
and, for positive constants C(δ), γ, the stopping time \(\hat n\) satisfies
Such an algorithm provides recovery at given accuracy level 𝜖 with \(\hat n\) measurements of minimax optimal order of magnitude (up to a poly-log factor), and with probability greater than 1 − δ. The sampling algorithm is adaptive since it does not require the knowledge of k, and since the number of measurements required depends only on k and not on the ‘worst case’ rank d.
The construction of non-asymptotic confidence sets C n for θ at any sample size n in the next subsections will imply that such algorithms exist for low rank matrix recovery problems. The main idea is to check sequentially, for a geometrically increasing number 2m of samples, m = 1, 2, …, if the diameter \(|C_{2^m}|{ }_F\) of a confidence set exceeds 𝜖. If this is not the case, the algorithm terminates. Otherwise one takes 2m+1 additional measurements and evaluates the diameter \(|C_{2^{m+1}}|{ }_F\). A precise description of the algorithm is given in the proof of the following theorem, which we detail for the case of ‘Pauli’ designs. The isotropic design case is discussed in Remark 18.9.
Theorem 18.1
Consider observations in the model (18.3) under Condition 18.1(b) with θ ∈ Θ +. Then an adaptive sampling algorithm in the sense of Definition 18.1 exists for any 𝜖, δ > 0.
Remark 18.2 (Dependence in σ of Definition 18.1 and Theorem 18.1)
Definition 18.1 and Theorem 18.1 are stated for the case where the standard deviation of the noise σ is assumed to be bounded by an absolute constant. It is straight-forward to modify the proofs to obtain a version where the dependency of the constants on the variance is explicit. Indeed, under Condition 1(a), Theorem 18.1 continues to hold if Eq. (18.31) is replaced by
For the ‘Pauli design case’—Condition 1(b)—Eq. (18.31) can be modified to
Remark 18.3 (Necessity of the Quantum Shape Constraint)
Note that the assumption θ ∈ Θ+ in the previous theorem is necessary (in the case of Pauli design): Else the example of θ = 0 or θ = E i—where E i is an arbitrary element of the Pauli basis—demonstrates that the number of measurements has to be at least of order d 2: otherwise with positive probability, E i is not drawn at a fixed sample size. On this event, both the measurements and \(\hat \theta \) coincide under the laws \(\mathbb P_0\) and \(\mathbb P_{E_i}\), so we cannot have \(\|\hat \theta -0\|{ }_F < \epsilon \) and \(\|\hat \theta - E_i\|{ }_F<\epsilon \) simultaneously for every 𝜖 > 0, disproving existence of an adaptive sampling algorithm. In fact, the crucial condition for Theorem 18.1 to work is that the nuclear norms \(\|\theta \|{ }_{S_1}\) are bounded by an absolute constant (here = 1), which is violated by \(\|E_i\|{ }_{S_1} = \sqrt {d}\).
18.3.2 A Non-asymptotic Confidence Set Based on Unbiased Risk Estimation and Sample-Splitting
We suppose that we have two samples at hand, the first being used to construct an estimator \(\tilde \theta \), such as the one from (18.26). We freeze \(\tilde \theta \) and the first sample in what follows and all probabilistic statements are under the distribution \(\mathbb P_\theta \) of the second sample Y, X of size \(n \in \mathbb N\), conditional on the value of \(\tilde \theta \). We define the following residual sum of squares statistic (recalling that σ 2 is known):
which satisfies \(\mathbb E _\theta \hat r_n = \|\theta -\tilde \theta \|{ }_F^2\) as is easily seen (see the proof of Theorem 18.2 below). Given α > 0, let ξ α,σ be quantile constants such that
(these constants converge to the quantiles of a fixed normal distribution as n →∞), let \(z_\alpha =\log (3/\alpha )\) and, for z ≥ 0 a fixed constant to be chosen, define the confidence set
where
Note that in the ‘quantum shape constraint’ case we can always bound \(\|v-\tilde \theta \|{ }_F \le 2\) which gives a confidence set that is easier to compute and of only marginally larger overall diameter. In many important situations, however, the quantity \(\bar z/\sqrt {n}\) is of smaller order than \(1/\sqrt {n}\), and the more complicated expression above is preferable.
It is not difficult to see (using that \(x^2 \lesssim y+x/\sqrt {n}\) implies \(x^2 \lesssim y +1/n\)) that the square Frobenius norm diameter of this confidence set is, with high probability, of order
Whenever \(d \ge \sqrt {n}\)—so as long as at most n ≤ d 2 measurements have been taken—the deviation terms are of smaller order than kd∕n, and hence C n has minimax optimal expected squared diameter whenever the estimator \(\tilde \theta \) is minimax optimal as in (18.26). Improvements for \(d < \sqrt {n}\), corresponding to n > d 2 measurements, will be discussed in the next subsections.
The following result shows that C n is an honest confidence set for arbitrary d × d matrices (without any rank constraint). Note that the result is non-asymptotic—it holds for every \(n \in \mathbb N\).
Theorem 18.2
Let \(\theta \in \mathbb H_d(\mathbb C)\) be arbitrary and let \(\mathbb P_\theta \) be the distribution of Y, X from model (18.3).
-
(a)
Assume Condition 18.1(a) and let C n be given by (18.33) with z = 0. We then have for every \(n \in \mathbb N\) that
$$\displaystyle \begin{aligned}\mathbb P_\theta(\theta \in C_n) \ge 1-\frac{2\alpha}{3} - 2 e^{-c n}\end{aligned}$$where c is a numerical constant. In the case of standard Gaussian design, c = 1∕24 is admissible.
-
(b)
Assume Condition 18.1(b), let C n be given by (18.33) with z > 0 and assume also that θ ∈ Θ + and \(\tilde \theta \in \Theta _+\) (that is, both satisfy the ‘quantum shape constraint’). Then for every \(n \in \mathbb N\) ,
$$\displaystyle \begin{aligned}\mathbb P_\theta(\theta \in C_n) \ge 1-\frac{2\alpha}{3} - 2 e^{-C(K) z}\end{aligned}$$where, for K the coherence constant of the basis,
$$\displaystyle \begin{aligned}C(K) = \frac{1}{(16+8/3)K^2}.\end{aligned}$$
In Part (a), if we want to control the coverage probability at level 1 − α, n needs to be large enough so that the third deviation term is controlled at level α∕3. In the Gaussian design case with α = 0.05, n ≥ 100 is sufficient, for smaller sample sizes one can reduce the coverage level. The bound in (b) is entirely non-asymptotic (using the quantum constraint) for suitable choices of z. Also note that the quantile constants z, z α, ξ α all scale at least as \(O(\log (1/\alpha ))\) in the desired coverage level α → 0.
Remark 18.4 (Dependence of the Confidence Set’s Diameter on K (Pauli Design) and σ)
Note that in the case of the Pauli design from Condition 1(b), the confidence set’s diameter depends on K only through the potential dependence of \(\|\theta - \tilde \theta \|{ }_F^2\) on K—the constants involved in the construction of \(\tilde C_n\) and on the bound on its diameter do not depend on K. On the other hand, the coverage probability of the confidence set depends on K, see Theorem 18.2, (b).
In this paper we assume that σ is a universal constant, and so as such it does not appear in Eqs. (18.33) and (18.34). It can however be interesting to investigate the dependence in σ. In the case of isotropic design from Condition 1(a), we could set
(where σ 2 could be replaced by twice the plug-in estimator of σ 2, using \(\hat \theta \)) and one would get
and Theorem 18.2 also holds by introducing minor changes in the proof. In the case of the Pauli design from Condition 1(b), we could set
(where σ 2 could be replaced by twice the plug-in estimator of σ 2, using \(\hat \theta \)) and one would get
and Theorem 18.2 also holds by introducing minor changes in the proof. In this case we do not get a full dependence in σ as in the isotropic design case from Condition 1(a). However if \(k^2d \lesssim n\), we could also obtain a result similar to the one for the Gaussian design, using part (c) of Lemma 18.1.
Remark 18.5 (Bernoulli Noise)
Theorem 18.2(b) holds as well for the Bernoulli measurement model from Sect. 18.2.2.2 with T ≥ d 2, with slightly different constants in the construction of C n and the coverage probabilities. See Remark 18.10 after the proof of Theorem 18.2(b) below. The modified quantile constants z, z α, ξ α still scale as \(O(\sqrt {1/\alpha })\) in the desired coverage level α → 0, and hence the adaptive sampling Theorem 18.1 holds for such noise too, if the number T of preparations of the quantum state exceeds d 2.
Remark 18.6 (Unknown Variance)
The above confidence set C n can be constructed with \(\tilde r_n= \frac {1}{n}\|Y-\mathcal X \tilde \theta \|{ }^2\) replacing \(\hat r_n\)—so without requiring knowledge of σ—if an a priori bound σ 2 ≤ vd∕n is available, with v a known constant. An example of such a situation was discussed at the end of Sect. 18.2.2.2 above in quantum tomography problems: when T ≥ n, the constant z should be increased by v in the construction of C n, and the coverage proof goes through as well by compensating for the centring at \(\mathbb E \varepsilon _i^2=\sigma ^2\) by the additional deviation constant v.
Remark 18.7 (Anisotropic Design Instead of Condition 1(a))
It is also interesting to consider the case of anisotropic design. This case is not very different, when it comes to confidence sets, than isotropic design, as long as the variance-covariance matrix of the anisotropic sub-Gaussian design is such that the ratio of its largest eigenvalue with the smallest eigenvalue is bounded. Lemma 18.1(a), which quantifies the effect of the design, would change as follows: There exist constants c −, c +, c > 0 that depend only on the variance-covariance matrix of the anisotropic sub-Gaussian design and that are such that
Using this instead of the inequality in Lemma 18.1(a) in the proof of Theorem 18.2, part (a) leads to a similar result as Theorem 18.2, part (a).
18.3.3 Improvements When \(d \le \sqrt {n}\)
The confidence set from Theorem 18.2 is optimal whenever the desired performance of \(\|\theta -\tilde \theta \|{ }_F^2\) is no better than of order \(1/\sqrt {n}\). From a minimax point of view we expect \(\|\theta -\tilde \theta \|{ }_F^2\) to be of order kd∕n for low rank θ ∈ R(k). In absence of knowledge about k ≥ 1 the confidence set from Theorem 18.2 can hence be guaranteed to be optimal whenever \(d \ge \sqrt {n}\), corresponding to the important regime n ≤ d 2 for sequential sampling algorithms. Refinements for measurement scales n ≥ d 2 are also of interest—we present two optimal approaches in this subsection for the designs from Condition 18.1.
18.3.3.1 Isotropic Design and U-Statistics
Consider first isotropic i.i.d design from Condition 18.1(a), and an estimator \(\tilde \theta \) based on an initial sample of size n (all statements that follow are conditional on that sample). Collect another n samples to perform the uncertainty quantification step. Define the U-statistic
whose \(\mathbb E _\theta \)-expectation, conditional on \(\tilde \theta \), equals \(\|\theta -\tilde \theta \|{ }_F^2\) in view of
Define
where
and \(C_1 \ge \zeta _1 \|\theta \|{ }_F,~C_2 \ge \zeta _2 \|\theta \|{ }_F^2\) with ζ i constants depending on α, σ. Note that if θ ∈ Θ+ then ∥θ∥F ≤ 1 can be used as an upper bound. In practice the constants ζ i can be calibrated by Monte Carlo simulations (see the implementation section below), or chosen based on concentration inequalities for U-statistics (see Ref. [14], Theorem 4.4.8). This confidence set has expected diameter
and hence is compatible with any minimax recover rate \(\|\tilde \theta - \theta \|{ }_F^2 \lesssim kd/n\) from (18.26), where k ≥ 1 is now arbitrary. For suitable choices of ζ i we now show that C n also has non-asymptotic coverage.
Theorem 18.3
Assume Condition 18.1(a), and let C n be as in (18.36). For every α > 0 we can choose \(\zeta _i(\alpha )=O(\sqrt {1/\alpha }), i=1,2,\) large enough so that for every \(n \in \mathbb N\) we have
Remark 18.8 (Dependence of the Confidence Set’s Diameter on σ)
As what was noted in Remark 18.4, Theorem 18.3 does not make explicit the dependence on σ, which is assumed to be (bounded by) an universal constant. In order to take the dependence on σ into account, we could replace z α,n in Eq. (18.36) by \( \frac {C_1 \|\theta -\tilde \theta \|{ }_F}{\sqrt {n}} + \sigma ^2\frac {C_2 d}{n}\) (where σ 2 could be replaced by twice the plug-in estimator of σ 2, using \(\hat \theta \)), and we would get
and Theorem 18.3 also holds by introducing minor changes in the proof.
18.3.3.2 Re-averaging Basis Elements When \(d \le \sqrt {n}\)
Consider the setting of Condition 18.1(b) where we sample uniformly at random from a (scaled) basis \(\{dE_1, \dots , dE_{d^2}\}\) of \(\mathbb M_d(\mathbb C)\). When \(d \le \sqrt {n}\) we are taking n ≥ d 2 measurements, and there is no need to sample at random from the basis as we can measure each individual coefficient, possibly even multiple times. Repeatedly sampling a basis coefficient tr(E kθ) leads to a reduction of the variance of the measurement by averaging. More precisely, when taking n = md 2 measurements for some (for simplicity integer) m ≥ 1, and if (Y k,l : l = 1, …, m) are the measurements Y i corresponding to the basis element E k, k ∈{1, …, d 2}, we can form averaged measurements
We can then define the new measurement vector \(\tilde Z = (\tilde Z_1, \dots , \tilde Z_{d^2})^T\) (using also m = n∕d 2)
and the statistic
which estimates \(\|\theta - \tilde \theta \|{ }_F^2\) with precision
Hence, for z α the quantiles of a N(0, 1) distribution and ξ α,σ as in (18.32) with d 2 replacing n there, we can define a confidence set
which has non-asymptotic coverage
for every \(n \in \mathbb N\), by similar (in fact, since Lemma 18.1 is not needed, simpler) arguments as in the proof of Theorem 18.2 below. The expected diameter of \(\bar C_n\) is by construction
now compatible with any rate of recovery kd∕n, 1 ≤ k ≤ d.
18.3.4 A Confidence Set in Trace Norm Under Quantum Shape Constraints
The confidence sets from the previous subsections are all valid in the sense that they contain information about the recovery of θ by \(\tilde \theta \) in Frobenius norm ∥⋅∥F. It is of interest to obtain results in stronger norms, such as for instance the nuclear norm \(\|\cdot \|{ }_{S_1}\), which is particularly meaningful for quantum tomography problems since it then corresponds to the total variation distance on the set of ‘probability density matrices’. In fact, since
the nuclear norm has a clear interpretation in terms of the maximum probability with which two quantum states can be distinguished by arbitrary measurements.
The absence of the ‘Hilbert space geometry’ induced by the relationship of the Frobenius norm to the inner product 〈⋅, ⋅〉F makes this problem significantly harder, both technically and from an information-theoretic point of view. In particular it appears that the quantum shape constraint θ ∈ Θ+ is crucial to obtain any results whatsoever, and for the theoretical results presented here it will be more convenient to perform an asymptotic analysis where \(\min (n,d) \to \infty \) (with o, O-notation to be understood accordingly).
Instead of Condition 18.1 we shall now consider any design (X 1, …, X n) in model (18.3) that satisfies the matrix RIP (18.7) with
As discussed above, this covers in particular the designs from Condition 18.1. We shall still use the convention discussed before Condition 18.1 that θ and the matrices X i are such that tr(X iθ) is always real-valued.
In contrast to the results from the previous section we shall now assume a minimal low rank constraint on the parameter space:
Condition 18.2
θ ∈ R +(k) := R(k) ∩ Θ + for some k satisfying
This in particular implies that the RIP holds with τ n(k) = o(1). Given this minimal rank constraint θ ∈ R +(k), we now show that it is possible to construct a confidence set C n that adapts to any low rank 1 ≤ k 0 < k. Here we may choose k = d but note that this forces n ≫ d 2 (for Condition 18.2 to hold with k = d).
We assume that there exists an estimator \(\tilde \theta _{\mathrm {Pilot}}\) that satisfies, uniformly in R(k 0) for any k 0 ≤ k and for n large enough,
where D = D(δ) depends on δ, and where so-defined r n will be used frequently below. Such estimators exist as has already been discussed before (18.26). We shall in fact require a little more, namely the following oracle inequality: for any k and any matrix S of rank k ≤ d, with high probability and for n large enough,
which in fact implies (18.41). Such inequalities exist assuming the RIP and Condition 18.2, see, e.g., Theorem 2.8 in Ref. [6]. Starting from \(\tilde \theta _{\mathrm {Pilot}}\) one can construct (see Theorem 18.5 below) an estimator that recovers θ ∈ R(k) in nuclear norm at rate \(k \sqrt {d/n}\), which is again optimal from a minimax point of view, even under the quantum constraint (as discussed, e.g., in Ref. [24]). We now construct an adaptive confidence set for θ centred at a suitable projection of \(\tilde \theta _{\mathrm {Pilot}}\) onto Θ+.
In the proof of Theorem 18.4 below we will construct estimated eigenvalues \((\hat \lambda _j, j=1, \dots , d)\) of θ (see after Lemma 18.3). Given those eigenvalues and \(\tilde \theta _{\mathrm {Pilot}}\), we choose \(\hat k\) to equal the smallest integer ≤ d such that there exists a rank \(\hat k\) matrix \(\tilde \theta '\) for which
is satisfied. Such \(\hat k\) exists with high probability (since the inequalities are satisfied for the true θ and λ j’s, as our proofs imply). Define next \(\hat \vartheta \) to be the 〈⋅, ⋅〉F-projection of \(\tilde \theta _{\mathrm {Pilot}}\) onto
and note that, since \(2 \hat k \ge \hat k\),
Finally define, for C a constant chosen below,
Theorem 18.4
Assume Condition 18.2 for some 1 ≤ k ≤ d, and let δ > 0 be given. Assume that with probability greater than 1 − 2δ∕3, (a) the RIP (18.7) holds with τ n(k) as in (18.40) and (b) there exists an estimator \(\tilde \theta _{\mathrm {Pilot}}\) for which (18.42) holds. Then we can choose C = C(δ) large enough so that, for C n as in the last display,
Moreover, uniformly in R +(k 0), 1 ≤ k 0 ≤ k, and with \(\mathbb P_\theta \)-probability greater than 1 − δ,
Theorem 18.4 should mainly serve the purpose of illustrating that the quantum shape constraint allows for the construction of an optimal trace norm confidence set that adapts to the unknown low rank structure. Implementation of C n is not straightforward so Theorem 18.4 is mostly of theoretical interest. Let us also observe that in full generality a result like Theorem 18.4 cannot be proved without the quantum shape constraint. This follows from a careful study of certain hypothesis testing problems (combined with lower bound techniques for confidence sets as in Refs. [19, 30]). Precise results are subject of current research and will be reported elsewhere.
18.4 Simulation Experiments
In order to illustrate the methods from this paper, we present some numerical simulations. The setting of the experiments is as follows: A random matrix \(\eta \in \mathbb M_d(\mathbb C)\) of norm ∥η∥F = R 1∕2 is generated according to two distinct procedures that we will specify later, and the observations are
where the ε i are i.i.d. Gaussian of mean 0 and variance 1. The observations are reparametrised so that η represents the ‘estimation error’ \(\theta - \hat \theta \), and we investigate how well the statistics
estimate the ‘accuracy of estimation’ \(\|\eta \|{ }_F^2= \|\theta -\hat \theta \|{ }_F^2\), conditional on the value of \(\hat \theta \). We will choose η in order to illustrate two extreme cases: a first one where the nuclear norm \(\|\eta \|{ }_{S_1}\) is ‘small’, corresponding to a situation where the quantum constraint is fulfilled; and a second one where the nuclear norm is large, corresponding to a situation where the quantum constraint is not fulfilled. More precisely we generate the parameter η in two ways:
-
‘Random Dirac’ case: set a single entry (with position chosen at random on the diagonal) of η to R 1∕2, and all the other coordinates equal to 0.
-
‘Random Pauli’ case: Set η equal to a Pauli basis element chosen uniformly at random and then multiplied by R 1∕2.
The designs that we consider are the Gaussian design, and the Pauli design, described in Condition 1. We perform experiments with d = 32, R ∈{0.1, 1} and
Note that d 2 = 1024, so that the first four choices of n correspond to the important regime n < d 2. Our results are plotted as a function of the number n of samples in Figs. 18.2, 18.3, 18.4, and 18.5. The solid red and blue curves are the median errors of the normalised estimation errors
after 1000 iterations, and the dotted lines are respectively, the (two-sided) 90% quantiles. We also report (see Tables 18.1, 18.2, 18.3, and 18.4) how well the confidence sets based on these estimates of the norm perform in terms of coverage probabilities, and of diameters. The diameters are computed as
for the U-Statistic approach and
for the RSS approach, where we have chosen C UStat = 2.5, C RSS = 1 and \(C_{\mathrm {UStat}}^{\prime } = C_{\mathrm {RSS}} = 6\) for all experiments—calibrated to a 95% coverage level.
From these numerical results, several observations can be made:
-
In Gaussian random designs, the results are insensitive to the nature of η (see Figs. 18.2 and 18.3 and Tables 18.1 and 18.2). This is not surprising since the Gaussian design is ‘isotropic’.
-
For Pauli designs with the quantum constraint (see Fig. 18.4 and Table 18.3) the RSS method works quite well even for small sample sizes. But the U-Stat method is not very reliable—indeed we see no empirical evidence that Theorem 18.3 should also hold true for Pauli design.
-
For Pauli design and when the quantum shape constraint is not satisfied our methods cease to provide reliable results (see Fig. 18.5 and in particular Table 18.4). Indeed, when the matrix η is chosen itself as a random Pauli (which is the hardest signal to detect under Pauli design) both the RSS and the U-Stat approach perform poorly. The confidence set are not honest anymore, which is in line with the theoretical limitations we observe in Theorem 18.2. Figure 18.5 illustrates that the methods do not detect the signal, since the norm of η is largely under-evaluated for small sample sizes. These limitations are less pronounced when n ≥ d 2. In this case one could use alternatively the re-averaging approach from Sect. 18.3.3.2 (not investigated in the simulations) to obtain honest results without the quantum shape constraint.
18.5 Proofs
18.5.1 Proof of Theorem 18.1
Proof
Before we define the algorithm and prove the result, a few preparatory remarks are required: Our sequential procedure will be implemented in m = 1, 2, …, T potential steps, in each of which 2 ⋅ 2m = 2m+1 measurements are taken. The arguments below will show that we can restrict the search to at most
steps. We also note that from the discussion after (18.7)—in particular since c = c(δ) from (18.8) is O(1∕δ 2)—a simple union bound over m ≤ T implies that the RIP holds with probability ≥ 1 − δ′, some δ′ > 0, simultaneously for every m ≤ T satisfying \(2^m \ge c'kd \overline {\log }d\), and with \(\tau _{2^m}(k)<c_0\), where c′ is a constant that depends on δ′, c 0 only. The maximum over \(T=O(\log (d/\epsilon ))\) terms is absorbed in a slightly enlarged poly-log term. Hence, simultaneously for all such sample sizes 2m, m ≤ T, a nuclear norm regulariser exists that achieves the optimal rate from (18.26) with n = 2m and for every k ≤ d, with probability greater than 1 − δ∕3. Projecting this estimator onto Θ+ changes the Frobenius error only by a universal multiplicative constant (arguing as in (18.43) below), and we denote by \(\tilde \theta _{2^m} \in \Theta _+\) the resulting estimator computed from a sample of size 2m.
We now describe the algorithm at the m-th step: Split the 2m+1 observations into two halves and use the first subsample to construct \(\tilde \theta _{2^m} \in \Theta _+\) satisfying (18.26) with \(\mathbb P_\theta \)-probability ≥ 1 − δ∕3. Then use the other 2m observations to construct a confidence set \(C_{2^m}\) for θ centred at \(\tilde \theta _{2^m}\): if 2m < d 2 we take \(C_{2^m}\) from (18.33) and if 2m ≥ d 2 we take \(C_{2^m}\) from (18.37)—in both cases of non-asymptotic coverage at least 1 − α, α = δ∕(3T). If \(|C_{2^m}|{ }_F \le \epsilon \) we terminate the procedure (\(m=\hat m\), \(\hat n = 2^{\hat m+1}\), \(\hat \theta = \tilde \theta _{2^{\hat m}}\)), but if \(|C_{2^m}|{ }_F>\epsilon \) we repeat the above procedure with 2 ⋅ 2m+1 = 2m+1+1 new measurements, etc., until the algorithm terminates, in which case we have used
measurements in total.
To analyse this algorithm, recall that the quantile constants z, z α, ξ α appearing in the confidence sets (18.33) and (18.37) for our choice of α = δ∕(3T) grow at most as \(O(\log (1/\alpha ))=O(\log T) = o(\overline {\log }d)\). In particular in view of (18.26) and (18.34) or (18.38) the algorithm necessarily stops at a ‘maximal sample size’ n = 2T+1 in which the squared Frobenius risk of the maximal model (k = d) is controlled at level 𝜖. Such \(T \in \mathbb N\) is \(O(\log (d/\epsilon ))\) and depends on σ, d, 𝜖, δ, hence can be chosen by the experimenter.
To prove that this algorithms works we show that the event
has probability at most 2δ∕3 for large enough C(δ), γ. By the union bound it suffices to bound the probability of each event separately by δ∕3. For the first: Since \(\hat n\) has been selected we know \(|C_{\hat n}|{ }_F\le \epsilon \) and since \(\hat \theta = \tilde \theta _{\hat n}\) the event A 1 can only happen when \(\theta \notin C_{\hat n}\). Therefore
For A 2, whenever θ ∈ R(k) and for all m ≤ T for which \(2^m \ge c'kd \overline {\log } d\), we have, as discussed above, from (18.34) or (18.38) and (18.26) that
where D′ is a constant. In the last inequality the expectation is taken under the distribution of the sample used for the construction of \(C_{2^m}\), and it holds on the event on which \(\tilde \theta _{2^m}\) realises the risk bound (18.26). Then let C(δ), γ be large enough so that \(C(\delta ) kd (\log d)^\gamma / \epsilon ^2 \ge c'kd\overline {\log }d\) and let \(m_0 \in \mathbb N\) be the smallest integer such that
Then, for C(δ) large enough and since \(T=O(\log (d/\epsilon )\),
by Markov’s inequality, completing the proof. \(\blacksquare \)
Remark 18.9 (Isotropic Sampling)
The proof above works analogously for isotropic designs as defined in Condition 18.1a). When 2m ≥ d 2, we replace the confidence set (18.37) in the above proof by the confidence set from (18.36). Assuming also that ∥θ∥F ≤ M for some fixed constant M, we can construct a similar upper bound for T and the above proof applies directly (with T of slighter larger but still small enough order). Instead of assuming an upper bound on ∥θ∥F one can simply continue using the confidence set (18.33) also when 2m ≥ d 2, in which case one has the slightly worse bound
for the number of measurements required.
18.5.2 Proof of Theorem 18.2
Proof
By Lemma 18.1 below with \(\vartheta =\tilde \theta - \theta \) the \(\mathbb P_\theta \)-probability of the complement of the event
is bounded by the deviation terms 2e −cn and 2e −C(K)z, respectively (note z = 0 in Case (a)). We restrict to this event in what follows. We can decompose
Since \(\mathbb P(Y+Z<0) \le \mathbb P(Y<0) + \mathbb P(Z<0)\) for any random variables Y, Z we can bound the probability
by the sum of the following probabilities
The first probability I is bounded by
About term II: Conditional on \(\mathcal X\) the variable \(\frac {1}{\sqrt {n}} \langle \varepsilon , \mathcal X(\theta -\tilde \theta )\rangle \) is centred Gaussian with variance \((\sigma ^2/n)\|\mathcal X(\theta -\tilde \theta )\|{ }^2\). The standard Gaussian tail bound then gives by definition of \(\bar z\), and conditional on \(\mathcal X\),
since, on the event \(\mathcal E\),
The overall bound for II follows from integrating the last but one inequality over the distribution of X. Term III is bounded by α∕3 by definition of ξ α,σ. \(\blacksquare \)
Remark 18.10 (Modification of the Proof for Bernoulli Errors)
If instead of Gaussian errors we work with the error model from Sect. 18.2.2.2, we require a modified treatment of the terms II, III in the above proof. For the pure noise term III we modify the quantile constants slightly to \(\xi _{\alpha , \sigma } = \sqrt {(1/\alpha )}\). If the number T of preparations satisfies T ≥ 4d 2 then Chebyshev’s inequality and (18.24) give
For the ‘cross term’ we have likewise with \(z_{\alpha }=\sqrt {1/\alpha }\) and \(a_i = (\mathcal X(\theta -\tilde \theta ))_i\) that, on the event \(\mathcal E\),
just as at the end of the proof of Theorem 18.2, so that coverage follows from integrating the last inequality w.r.t. the distribution of X. The scaling T ≈ d 2 is similar to the one discussed in Theorem 3 in Ref. [12].
Lemma 18.1
-
(a)
For isotropic design from Condition 18.1(a) and any fixed matrix \(\vartheta \in \mathbb H_d(\mathbb C)\) we have, for every \(n \in \mathbb N\) ,
$$\displaystyle \begin{aligned}\Pr \left(\left|\frac{1}{n}\|\mathcal X\vartheta\|{}^2 - \|\vartheta\|{}_F^2\right| > \frac{\|\vartheta\|{}_F^2}{2}\right) \le 2 e^{-c n}.\end{aligned}$$In the standard Gaussian design case we can take c = 1∕24.
-
(b)
In the ‘Pauli basis’ case from Condition 18.1(b) we have for any fixed matrix \(\vartheta \in \mathbb H_d(\mathbb C)\) satisfying the Schatten-1-norm bound \(\|\vartheta \|{ }_{S_1} \le 2\) and every \(n \in \mathbb N\) ,
$$\displaystyle \begin{aligned}\Pr \left(\left|\frac{1}{n}\|\mathcal X\vartheta\|{}^2 - \|\vartheta\|{}_F^2\right| > \max\left(\frac{\|\vartheta\|{}_F^2}{2}, z\frac{d}{n} \right) \right) \le 2 \exp \left\{-C(K) z \right\}\end{aligned}$$where C(K) = 1∕[(16 + 8∕3)K 2], and where K is the coherence constant of the basis.
-
(c)
In the ‘Pauli basis’ case from Condition 18.1(b) we have for any fixed matrix \(\vartheta \in \mathbb H_d(\mathbb C)\) such that the rank of 𝜗 is smaller than 2k and every \(n \in \mathbb N\) ,
$$\displaystyle \begin{aligned}\Pr \left(\left|\frac{1}{n}\|\mathcal X\vartheta\|{}^2 - \|\vartheta\|{}_F^2\right| > \max\left(\frac{\|\vartheta\|{}_F^2}{2}, z\frac{d}{n} \right) \right) \le 2 \exp \left\{-\frac{n}{17 K^2 k^2 d} \right\}.\end{aligned}$$
Proof
We first prove the isotropic case. From (18.5) we see
where the Z i∕∥𝜗∥F are sub-Gaussian random variables. Then the \(Z_i^2/\|\vartheta \|{ }_F^2\) are sub-exponential and we can apply Bernstein’s inequality (Prop. 4.1.8 in Ref. [14]) to the last probability. We give the details for the Gaussian case and derive explicit constants. In this case g i := Z i∕∥𝜗∥F ∼ N(0, 1) so the last probability is bounded, using Theorem 4.1.9 in Ref. [14], by
and the result follows.
Under Condition 18.1(b), if we write \(D= \max (n\|\vartheta \|{ }_F^2/2, z d)\) we can reduce likewise to bound the probability in question by
where the Y i = |tr(X i𝜗)|2 are i.i.d. bounded random variables. Using \(\|E_i\|{ }_{op} \le K/\sqrt {d}\) from Condition 18.1(b) and the quantum constraint \(\|\vartheta \|{ }_F \le \|\vartheta \|{ }_{S_1} \le 2\) we can bound
as well as
Bernstein’s inequality for bounded variables (e.g., Theorem 4.1.7 in Ref. [14]) applies to give the bound
after some basic computations, by distinguishing the two regimes of \(D=n\|\vartheta \|{ }_F^2/2 \ge zd\) and \(D=zd \ge n\|\vartheta \|{ }_F^2/2\).
Finally for (c), using the same reasoning as above and using \(\|E_i\|{ }_{op} \le K/\sqrt {d}\) from Condition 18.1(b) and the fact that the estimator is also of rank less than k, we have \(\|\vartheta \|{ }_F \le \|\vartheta \|{ }_{S_1} \le \sqrt {2k}\|\vartheta \|{ }_F\) we can bound
as well as
Bernstein’s inequality for bounded variables (e.g., Theorem 4.1.7 in Ref. [14]) applies to give the bound
after some basic computations.\(\blacksquare \)
18.5.3 Proof of Theorem 18.3
Proof
Since \(\mathbb E _\theta \hat R_n = \|\theta - \tilde \theta \|{ }_F^2\) we have from Chebyshev’s inequality
Now \(U_n =\hat R_n -\mathbb E _\theta \hat R_n\) is a centred U-statistic and has Hoeffding decomposition U n = 2L n + D n where
is the linear part and
the degenerate part. We note that L n and D n are orthogonal in \(L^2(\mathbb P_\theta )\).
The linear part can be decomposed into
where
and
Now by the i.i.d. assumption we have
Moreover, by transposing the indices m, k and m′, k′ in an arbitrary way into single indices M = 1, …, d 2, K = 1, …, d 2, d 2 = p, respectively, basic computations given before eq. (28) in Ref. [30] imply that the variance of the second term is bounded by
where c is a constant that depends only on \(\mathbb E X_{1,1}^4\) (which is finite since the X 1,1 are sub-Gaussian in view of Condition 18.1(a)). Moreover, the degenerate term satisfies
in view of standard U-statistic computations leading to eq. (6.6) in Ref. [21], with d 2 = p, and using the same transposition of indices as before. This proves coverage by choosing the constants in the definition of z α,n large enough.\(\blacksquare \)
18.5.4 Proof of Theorem 18.4
We prove the result for symmetric matrices with real entries—the case of Hermitian matrices requires only minor (mostly notational) adaptations.
Given the estimator \(\tilde \theta _{\mathrm {Pilot}}\), we can easily transform it into another estimator \(\tilde \theta \) for which the following is true.
Theorem 18.5
There exists an estimator \(\tilde \theta \) that satisfies, uniformly in θ ∈ R(k), for any k ≤ d and with \(\mathbb P_\theta \)-probability greater than 1 − 2δ∕3,
as well as,
and then also
Proof
Let \(\tilde \theta _{\mathrm {Pilot}}\) and let \(\tilde \theta \) be the element of R(d) with smallest rank k′ such that
Such \(\tilde \theta \) exists and has rank ≤ k, with probability ≥ 1 − 2δ∕3, since θ ∈ R(k) satisfies the above inequality in view of (18.41). The \(\|\cdot \|{ }^2_F\)-loss of \(\tilde \theta \) is no larger than r n(k) by the triangle inequality
and this completes the proof of the third claim in view of (18.2). \(\blacksquare \)
The rest of the proof consists of three steps: The first establishes some auxiliary empirical process type results, which are then used in the second step to construct a sufficiently good simultaneous estimate of the eigenvalues of θ. In Step III the coverage of the confidence set is established.
18.5.4.1 Step I
Let θ ∈ R +(k) = R(k) ∩ Θ+ and let \(\tilde \theta \) be the estimator from Theorem 18.5. Then with probability ≥ 1 − 2δ∕3, and if \(\eta = \tilde \theta - \theta \), we have
and that
For the rest of the proof we restrict in what follows to the event of probability greater than or equal to 1 − 2δ∕3 described by (a) and (b) in the hypothesis of the theorem.
Write \(Y_i^{\prime } = Y_i -\mathrm {tr}(X^i \tilde \theta )\) for the ‘new observations’
For any d × d′ matrix V we set
which estimates
Let now U be any unit vector in \(\mathbb R^d\). Then in the above notation (d′ = 1) we can write
If \(\mathbb U\) denotes the d × d matrix UU T, the last quantity can be written as
We can hence bound, for \(\mathcal S = \{U \in \mathbb R^d: \|U\|{ }_2=1\}\)
Lemma 18.2
The right hand side on the last inequality is, with probability greater than 1 − δ, of order
Proof
The first term in the bound corresponds to the first supremum on the right hand side of the last inequality, and follows directly from the matrix RIP (and Lemma 18.4). For the second term we argue conditionally on the values of \(\mathcal X\) and on the event for which the matrix RIP is satisfied. We bound the supremum of the Gaussian process
indexed by elements U of the unit sphere \(\mathcal S\) of \(\mathbb R^d\), which satisfies the metric entropy bound
by a standard covering argument. Moreover \(\mathbb U = UU^T \in R(1)\) and hence for any pair of vectors \(U, \bar U \in \mathcal S\) we have that \(\mathbb U - \bar {\mathbb U} \in R(2)\). From the RIP we deduce for every fixed \(U, \bar U \in \mathcal S\) that
since τ n(2) = O(1) and since
Hence any δ-covering of \(\mathcal S\) in ∥⋅∥ induces a δ∕C covering of \(\mathcal S\) in the intrinsic covariance \(d_{\mathbb G_\varepsilon }\) of the (conditional on \(\mathcal X\)) Gaussian process \(\mathbb G_\varepsilon \), i.e.,
with constants independent of X. By Dudley’s metric entropy bound (e.g., Ref. [14]) applied to the conditional Gaussian process we have for d > 0 some constant
and hence we deduce that
with constants independent of X, so that the result follows from applying Markov’s inequality. \(\blacksquare \)
18.5.4.2 Step II
Define the estimator
Then we can write, using \(U^T \tilde \gamma _\eta (I_d)U = \tilde \gamma _\eta (U)\),
and from the previous lemma we conclude, for any unit vector U that with probability ≥ 1 − δ,
Let now \(\hat \theta \) be any symmetric positive definite matrix such that
Such a matrix exists, for instance θ ∈ R +(k), and by the triangle inequality we also have
Lemma 18.3
Let M be a symmetric positive definite d × d matrix with eigenvalues λ j’s ordered such that λ 1 ≥ λ 2 ≥… ≥ λ d. For any j ≤ d consider an arbitrary collection of j orthonormal vectors \(\mathcal V_j = (V^\iota : 1 \le \iota \le j)\) in \(\mathbb R^d\). Then we have
and
Let \(\hat R\) be the rotation that diagonalises \(\hat \theta \) such that \(\hat R^T \hat \theta \hat R = diag(\hat \lambda _j: j =1, \dots , d)\) ordered such that \(\hat \lambda _j \ge \hat \lambda _{j+1}\) ∀j. Moreover let R be the rotation that does the same for θ and its eigenvalues λ j. We apply the previous lemma with \(M= \hat \theta \) and \(\mathcal V\) equal to the column vectors r ι : ι ≤ l − 1 of R to obtain, for any fixed l ≤ j ≤ d,
and also that
From (18.47) we deduce, that
as well as
with probability ≥ 1 − δ. Combining these bounds we obtain
18.5.4.3 Step III
We show that the confidence sets covers the true parameter on the event of probability ≥ 1 − δ on which Steps I and II are valid, and for the constant C chosen large enough.
Let \(\Pi = \Pi _{R^+(2\hat k)}\) be the projection operator onto \(R^+(2 \hat k)\). We have
We have, using (18.50) and Lemma 18.5 below
for C large enough.
Moreover, using the oracle inequality (18.42) with S = Πθ and (18.43),
We finally deal with the approximation error: Note
By (18.50) we know that
Hence out of the λ l’s with indices \(l >\hat k\) there have to be less than \(\hat k\) coefficients which exceed 4v n. Since the eigenvalues are ordered this implies that the λ l’s with indices \(l>2 \hat k\) are all less than or equal to 4v n, and hence the quantity in the last but one display is bounded by (since \(\hat k < 2 \hat k\)), using again (18.50) and the definition of \(\hat k\),
Overall we get the bound
for C large enough, which completes the proof of coverage of C n by collecting the above bounds. The diameter bound follows from \(\hat k \le k\) (in view of the defining inequalities of \(\hat k\) being satisfied, for instance, for \(\tilde \theta ' = \theta \), whenever θ ∈ R +(k 0).)
18.6 Auxiliary Results
18.6.1 Proof of Lemma 18.3
-
(a)
Consider the subspaces E = span((V ι)ι≤j)⊥ and F = span((e ι)ι≤j+1) of \(\mathbb R^d\), where the e ι’s are the eigenvectors of the d × d matrix M corresponding to eigenvalues λ j. Since dim(E) + dim(F) = (d − j) + j + 1 = d + 1, we know that \(E \bigcap F\) is not empty and there is a vectorial sub-space of dimension 1 in the intersection. Take \(U \in E \bigcap F\) such that ∥U∥ = 1. Since U ∈ F, it can be written as
$$\displaystyle \begin{aligned}U = \sum_{\iota=1}^{j+1} u_{\iota} e_{\iota}\end{aligned}$$for some coefficients u ι. Since the e ι’s are orthogonal eigenvectors of the symmetric matrix M we necessarily have
$$\displaystyle \begin{aligned}MU = \sum_{\iota=1}^{j+1} \lambda_{\iota} u_{\iota} e_{\iota},\end{aligned}$$and thus
$$\displaystyle \begin{aligned}U^T MU = \sum_{\iota=1}^{j+1} \lambda_{\iota} u_{\iota}^2.\end{aligned}$$Since the λ ι’s are all non-negative and ordered in decreasing absolute value, one has
$$\displaystyle \begin{aligned}U^T MU = \sum_{\iota=1}^{j+1} \lambda_{\iota} u_{\iota}^2 \geq \lambda_{j+1} \sum_{\iota=1}^{j+1} u_{\iota}^2 = \lambda_{j+1} \|U\|{}^2 = \lambda_{j+1}.\end{aligned}$$Taking the supremum in U yields the result.
-
(b)
For each ι ≤ j, let us write the decomposition of V ι on the basis of eigenvectors (e l : l ≤ d) of M as
$$\displaystyle \begin{aligned}V^{\iota} = \sum_{l \leq d} v^{\iota}_l e_{l}.\end{aligned}$$Since the (e l) are the eigenvectors of M we have
$$\displaystyle \begin{aligned}\sum_{\iota \leq j} (V^{\iota})^T M V^{\iota} = \sum_{\iota \leq j} \sum_{l=1}^d \lambda_{l} (v^{\iota}_l)^2,\end{aligned}$$where \(\sum _{l=1}^d (v^{\iota }_l)^2 = 1\) and \(\sum _{\iota \leq j} (v^{\iota }_l)^2 \leq 1\), since the V ι are orthonormal. The last expression is maximised in \((v^{\iota }_l)_{\iota \leq j, 1 \le l \leq d}\) and under these constraints, when \(v^{\iota }_{\iota } = 1\) and \(v^{\iota }_{l} = 0\) if ι ≠ l (since the (λ ι) are in decreasing order), and this gives
$$\displaystyle \begin{aligned}\sum_{\iota \leq j} (V^{\iota})^T M V^{\iota} \leq \sum_{\iota \leq j} \lambda_{\iota}.\end{aligned}$$
18.6.2 Some Further Lemmas
Lemma 18.4
Under the RIP (18.7) we have for every 1 ≤ k ≤ d that, with probability at least 1 − δ,
Proof
The matrix RIP can be written as
for a suitable \(M\in \mathbb H_{d^2}(\mathbb C)\). The above bound then follows from applying the Cauchy-Schwarz inequality to
\(\blacksquare \)
The following lemma can be proved by basic linear algebra, and is left to the reader.
Lemma 18.5
Let M ≥ 0 with positive eigenvalues (λ j)j ordered in decreasing order. Denote with \(\Pi _{R^+(j-1)}\) the projection onto R +(j − 1) = R(j − 1) ∩ Θ +. Then for any 2 ≤ j ≤ d we have
Notes
- 1.
The term ‘tomography’ goes back to the use of Radon transforms in early schemes for estimating quantum states of electromagnetic fields [1, 27]. It has become synonymous with ‘quantum density matrix estimation’, even though current methods applied to quantum systems with a finite dimension d have no technical connection to classical tomographic reconstruction algorithms.
- 2.
A more insightful way of proving the first identity is to realise that \(\mathbb E\big (\chi _S(C^x)\big )\) is effectively a Fourier coefficient (over the group \(\mathbb {Z}_2^N\)) of the distribution function of the {−1, 1}N-valued random variable C x (e.g., [11]). Equation (18.21) is then nothing but an inverse Fourier transform.
- 3.
We note that quantum mechanics allows to design measurement devices that directly probe the observable of \(\sigma ^{y_1} \otimes \dots \otimes \sigma ^{y_N}\), without first measuring the spin of every particle and then computing a parity function. In fact, the ability to perform such correlation measurements is crucial for quantum error correction protocols [31]. For practical reasons these setups are used less commonly in tomography experiments, though.
References
L. Artiles, R. Gill, M. Guta, An invitation to quantum tomography. J. R. Stat. Soc. 67, 109 (2005)
K. Audenaert, S. Scheel, Quantum tomographic reconstruction with error bars: a Kalman filter approach. New J. Phys. 11(2), 023028 (2009)
P.J. Bickel, Y. Ritov, A.B. Tsybakov, Simultaneous analysis of lasso and Dantzig selector. Ann. Stat. 37(4), 1705–1732 (2009)
R. Blume-Kohout, Robust error bars for quantum tomography (2012). arXiv:1202.5270
P. Bühlmann, S. van de Geer, Statistics for High-Dimensional Data: Methods, Theory and Applications (Springer, Berlin, 2011)
E.J. Candès, Y. Plan, Tight oracle inequalities for low-rank matrix recovery from a minimal number of noisy random measurements. IEEE Trans. Inform. Theory 57(4), 2342–2359 (2011)
E.J. Candès, T. Tao, The Dantzig selector: statistical estimation when p is much larger than n. Ann. Stat. 35(6), 2313–2351 (2007)
A. Carpentier, A. Kim, An iterative hard thresholding estimator for low rank matrix recovery with explicit limiting distribution (2015). arxiv preprint 1502.04654
A. Carpentier, O. Klopp, M. Löffler, R. Nickl, et al., Adaptive confidence sets for matrix completion. Bernoulli 24(4A), 2429–2460 (2018)
M. Christandl, R. Renner, Reliable quantum state tomography. Phys. Rev. Lett. 109, 120403 (2012)
R. De Wolf, A brief introduction to Fourier analysis on the Boolean cube. Theory Comput. Libr. Grad. Surv. 1, 1–20 (2008)
S.T. Flammia, D. Gross, Y.-K. Liu, J. Eisert, Quantum tomography via compressed sensing: error bounds, sample complexity and efficient estimators. New J. Phys. 14(9), 095022 (2012)
E. Giné, R. Nickl, Confidence bands in density estimation. Ann. Stat. 38, 1122–1170 (2010)
E. Giné, R. Nickl, Mathematical Foundations of Infinite-Dimensional Statistical Models (Cambridge University Press, Cambridge, 2015)
D. Gross, Recovering low-rank matrices from few coefficients in any basis. IEEE Trans. Inf. Theory 57(3), 1548–1566 (2011)
D. Gross, Y.-K. Liu, S.T. Flammia, S. Becker, J. Eisert, Quantum state tomography via compressed sensing. Phys. Rev. Lett. 105(15), 150401 (2010)
M. Guta, T. Kypraios, I. Dryden, Rank-based model selection for multiple ions quantum tomography. New J. Phys. 14, 105002 (2012)
H. Haeffner, W. Haensel, C.F. Roos, J. Benhelm, D.C. al kar, M. Chwalla, T. Koerber, U.D. Rapol, M. Riebe, P.O. Schmidt, C. Becher, O. Gühne, W. Dür, R. Blatt, Scalable multi-particle entanglement of trapped ions. Nature 438, 643 (2005)
M. Hoffmann, R. Nickl, On adaptive inference and confidence bands. Ann. Stat. 39, 2382–2409 (2011)
A.S. Holevo, Statistical Structure of Quantum Theory (Springer, Berlin, 2001)
Y.I. Ingster, A.B. Tsybakov, N. Verzelen, Detection boundary in sparse regression. Electron. J. Stat. 4, 1476–1526 (2010)
A. Javanmard, A. Montanari, Confidence intervals and hypothesis testing for high-dimensional regression. J. Mach. Learn. Res. 15(1), 2869–2909 (2014)
A. Kalev, R.L. Kosut, I.H. Deutsch, Informationally complete measurements from compressed sensing methodology. arXiv:1502.00536
V. Koltchinskii, Von Neumann entropy penalization and low-rank matrix estimation. Ann. Stat. 39(6), 2936–2973 (2011)
V. Koltchinskii, K. Lounici, A.B. Tsybakov, Nuclear-norm penalization and optimal rates for noisy low-rank matrix completion. Ann. Stat. 39(5), 2302–2329 (2011)
H. Leeb, B.M. Pötscher, Can one estimate the conditional distribution of post-model-selection estimators? Ann. Stat. 34(5), 2554–2591 (2006)
U. Leonhardt, Measuring the Quantum State of Light (Cambridge University Press, Cambridge, 2005)
Y.-K. Liu, Universal low-rank matrix recovery from Pauli measurements. Adv. Neural Inf. Process. Syst. 24, 1638–1646 (2011)
V. Mnih, C. Szepesvári, J.Y. Audibert, Empirical bernstein stopping, in Proceedings of the 25th International Conference on Machine Learning (ACM, New York, 2008), pp. 672–679
R. Nickl, S. van de Geer, Confidence sets in sparse regression. Ann. Stat. 41(6), 2852–2876 (2013)
M.A. Nielsen, I.L. Chuang, Quantum Computation and Quantum Information (Cambridge University Press, Cambridge, 2000)
A. Peres, Quantum Theory (Springer, Berlin, 1995)
B. Recht, M. Fazel, P.A. Parrilo, Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Rev. 52, 471 (2010)
J. Shang, H.K. Ng, A. Sehrawat, X. Li, B.-G. Englert, Optimal error regions for quantum state estimation. New J. Phys. 15(12), 123026 (2013)
K. Temme, F. Verstraete, Quantum chi-squared and goodness of fit testing. J. Math. Phys. 56(1), 012202 (2015)
S. van de Geer, P. Bühlmann, Y. Ritov, R. Dezeure, On asymptotically optimal confidence regions and tests for high-dimensional models. Ann. Stat. 42(3), 1166–1202 (2014)
Acknowledgements
The work of A. Carpentier was done when she was in Cambridge and is partially supported by the Deutsche Forschungsgemeinschaft (DFG) Emmy Noether grant MuSyAD (CA 1488/1-1), by the DFG-314838170, GRK 2297 MathCoRe, by the DFG GRK 2433 DAEDALUS (384950143/GRK2433), and by the DFG CRC 1294 ‘Data Assimilation’, Project A03, and by the UFA-DFH through the French-German Doktorandenkolleg CDFA 01-18. D. Gross acknowledges support by the DFG (SPP1798 CoSIP), Germany’s Excellence Strategy—Cluster of Excellence Matter and Light for Quantum Computing (ML4Q) EXC 2004/1—390534769, and the ARO under contract W911NF-14-1-0098 (Quantum Characterization, Verification, and Validation). J. Eisert was supported by the DFG, CRC 183, Project B01, by the DFG GRK DAEDALUS, the DFG SPP1798 CoSIP, the ERC, and the Templeton Foundation.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Carpentier, A., Eisert, J., Gross, D., Nickl, R. (2019). Uncertainty Quantification for Matrix Compressed Sensing and Quantum Tomography Problems. In: Gozlan, N., Latała, R., Lounici, K., Madiman, M. (eds) High Dimensional Probability VIII. Progress in Probability, vol 74. Birkhäuser, Cham. https://doi.org/10.1007/978-3-030-26391-1_18
Download citation
DOI: https://doi.org/10.1007/978-3-030-26391-1_18
Published:
Publisher Name: Birkhäuser, Cham
Print ISBN: 978-3-030-26390-4
Online ISBN: 978-3-030-26391-1
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)