Abstract
This article presents new results concerning the recovery of a signal from the magnitude only measurements where the signal is not sparse in an orthonormal basis but in a redundant dictionary, which we call it phase retrieval with redundant dictionary for short. To solve this phaseless problem, we analyze the \( \ell _1 \)-analysis model. Firstly we investigate the noiseless case with presenting a null space property of the measurement matrix under which the \( \ell _1 \)-analysis model provides an exact recovery. Secondly we introduce a new property (S-DRIP) of the measurement matrix. By solving the \( \ell _1 \)-analysis model, we prove that this property can guarantee a stable recovery of real signals that are nearly sparse in overcomplete dictionaries.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
1.1 Phase Retrieval
Phase retrieval is the process of recovering signals from phaseless measurements. It is of fundamental importance in numerous areas of applied physics and engineering [11, 14]. In general form, phase retrieval problem is to estimate the original signal \( x_0\in \mathbb {H}^n \) (\(\mathbb {H}={\mathbb C}\) or \( {\mathbb R}\)) from
where \(A = [a_1,\ldots , a_m]^*\in {\mathbb H}^{m\times n}\) is the measurement matrix and \( e=[e_1,\cdots , e_m]\in {\mathbb H}^m \) is an error term. While only the magnitude of \( Ax_0 \) is available, it is important to note that the setup naturally leads to ambiguous solutions. For example, if \( \hat{x}\in {\mathbb H}^n \) is a solution to (1.1), then any multiplication of \( \hat{x} \) and a scalar \( c\in {\mathbb H}\) (\( |c|=1 \)) is also a solution to (1.1). Hence, these global ambiguities are considered acceptable for this problem. In this paper, we recover the signal \( x_0 \) actually means that we reconstruct \( x_0 \) up to a unimodular constant.
It is known that, when \( {\mathbb H}={\mathbb R}\), at least \( 2n-1 \) measurements are needed to recover a signal \( x\in {\mathbb R}^n\) [3]. For the complex case, the minimum number of measurements are proved to be at least \( 4n-4 \) when n is in the form of \( n=2^k+1, k\in \mathbb {Z_+} \) [9]. However, for a general dimension n , the same question is still open. About the minimum number of observations, more details can be found in [4, 20]. To reduce the measurement number, priori information must be given, such as sparsity, which means that only few elements in the target signal \( x_0 \) is nonzero. In view of such sparse signals, phase retrieval is also known as compressive phase retrieval, which have many applications in data acquisition [15, 18]. The compressive phase retrieval problem is in fact the magnitude-only compressive sensing problem. For compressive phase retrieval, Wang and Xu explored the minimum number of measurements and extended the null space property in compressed sensing to compressive phase retrieval [20]. In [19], Voroniski and Xu gave the definition of strong restricted isometry property (Definition 2.2) and then many conclusions in compressed sensing can be extended to compressive phase retrieval, such as instance optimality [12].
1.2 Phase Retrieval with Redundant Dictionary
The above conclusions in compressive phase retrieval hold just for signals which are sparse in the standard coordinate basis. However, there are many examples in which a signal of interest is not sparse in an orthonormal basis but sparse in an overcomplete dictionary, such as radar images [13]. We refer to such signals as dictionary-sparse signals. In recent years, many researchers laid special stress on analysing these dictionary-sparse signals in compressed sensing [1, 7, 16]. However, the phase retrieval literature is lacking on this subject. Motivated by the wide application of redundant dictionaries and frames in signal processing and data analysis, we aim to build up a framework for the recovery of dictionary-sparse signals in phase retrieval, which we call it phase retrieval with redundant dictionary.
Suppose \( D\in {\mathbb H}^{n\times N} \) is an overcomplete dictionary (\( n<N \)) or a redundant dictionary. When \( n\ll N \), we say the dictionary D is highly overcomplete or highly redundant. Suppose the signal \( x_0\in {\mathbb H}^n \) is sparse in the overcomplete dictionary \( D\in {\mathbb H}^{n\times N } \). I.e., there exists a sparse vector \( z_0\in {\mathbb H}^N \), such that \( x_0=Dz_0 \). Thus the phase retrieval with redundant dictionary can be interpreted as recovering a signal \( x_0=Dz_0 \) from the measurements \( |ADz_0| \), where \( z_0 \) is sparse. That is to recover \( Dz_0 \) from
1.3 The \( \ell _1 \)-Analysis Model
Suppose the signal \( x_0\in {\mathbb H}^n \) can be expressed as \( x_0=Dz_0 \), where \( D\in {\mathbb H}^{n\times N } \) is a redundant dictionary and \( z_0\in {\mathbb H}^N \) is a sparse vector. When \( {\mathbb H}={\mathbb C}\), we use \( D^* \) to represent the adjoint conjugate of D . When \( {\mathbb H}= {\mathbb R}\), we use \( D^* \) to represent the transpose of D . In compressed sensing, to reconstruct the signal \( x_0 \), the most commonly used model is the \( \ell _1 \)-analysis model
where \( \epsilon \) is the upper bound of the noise. Due to the smaller dimension of the unknown, \( \ell _1 \)-analysis leads to a simple optimization problem, which is considerably easier to solve. That’s why the \( \ell _1 \)-analysis model is widely used. We refer interested readers to [1, 7, 10] for more superiorities of \( \ell _1 \)-analysis model. In [7], Candès et al. proved that when D is a tight frame and \( D^*x_0 \) is almost k -sparse, the \( \ell _1 \)-analysis (1.3) can guarantee a stable recovery provided that the measurement matrix is Gaussian random matrix with \( m=\mathcal {O}(k\log (n/k)) \).
For the phase retrieval with redundant dictionary (1.2), we also consider the corresponding \( \ell _1 \)-analysis model
where \( \epsilon \) is the upper bound of the noise level. In this paper, we aim to explore the conditions under which the \( \ell _1 \)-analysis model (1.4) can generate an accurate or a stable solution to (1.2). First for the noiseless case, we analyze the null space of the measurement matrix and give the conditions for exact recovery. Then for the noise case, we give a new property on the measurement matrix and prove that this property can guarantee a stable recovery.
Note that when \( D=I \), the phase retrieval with redundant dictionary is reduced to the traditional phase retrieval and the \( \ell _1 \)-analysis model is reduced to
For this case, when \( {\mathbb H}={\mathbb R}\), Gao et al. provided a detailed analysis of (1.5) in [12] and had the conclusion that a k -sparse signal can be stably recovered by \( \mathcal {O}(k\log (n/k)) \) Gaussian random measurements. Then a natural question that comes to mind is whether this conclusion still holds for a general frame D .
1.4 Organization
The rest of the paper is organized as follows. In Sect. 2, we give notations and recall some previous conclusions. In Sect. 3, for noiseless case (\( \epsilon =0 \)), we analyze the null space of the measurement matrix and give sufficient and necessary conditions for (1.4) to achieve an exact solution, which will be discussed in real and complex case separately. In general, it’s hard to check whether a matrix satisfies the null space property or not. So in Sect. , we introduce a new property (S-DRIP) (Definition 4.1) on the measurement matrix, which is a natural generalization of the DRIP (see [7] for more details). Using this property, we prove that when the measurement matrix is real Gaussian random matrix with \( m\ge \mathcal {O}(k\log (n/k)) \), the \( \ell _1 \)-analysis (1.4) can guarantee a stable recovery of real signals which are k -sparse in a redundant dictionary. In Sect. 5, we discuss the drawbacks of our results and file out some proper directions for the coming study. Lastly, some proofs are given in the Appendix.
2 Notations and Previous Results
We use \( \ell _0 \)-norm to measure the cardinality of non-zeros of a vector z . We call a signal z is k -sparse, if there are at most k non-zero elements in the signal, i.e., \( \Vert z\Vert _0\le k \). A set of vectors \(\{d_1,\cdots , d_N \} \) in \( {\mathbb H}^n \) is a frame of \( {\mathbb H}^n \) if there exist constants \( 0<s\le t<\infty \) such that for any \( f\in {\mathbb H}^n \),
If \( s=t \), the frame is a tight frame. We call \( D\in {\mathbb H}^{n\times N} \) a frame in the sense that the columns of D form a frame. Let
and
Suppose the target signal \(x_0\) is in the set \( D\Sigma _k^N \), which means that \(x_0\) can be represented as \(x_0=Dz_0\), where \(z_0\in \Sigma _k^N\).
The best k -term approximation error is defined as
For positive integers p, q with \( p\le q \), we use [p : q] to represent the set \( \{p,p+1,\ldots ,q-1,q \} \). Suppose \( T\subseteq [1:m] \) is a subset of [1 : m] . We use \( T^c \) to represent the complement set of T and |T| to denote the cardinal number of T . Let \( A_T:=[a_j, j\in T]^* \) denote the sub-matrix of A where only rows with indices in T are kept. Denote \( \mathcal {N}(A) \) as the null space of A .
Definition 2.1
(DRIP) [7] Fix a dictionary \( D\in {\mathbb R}^{n\times N} \) and a matrix \( A\in {\mathbb R}^{m\times n} \). The matrix A satisfies the DRIP with parameters \( \delta \) and k if
holds for all k -sparse vectors \( z\in {\mathbb R}^N \).
The paper [7] have shown that Gaussian random matrices and other random compressed sensing matrices satisfy the DRIP of order k provided the number of measurements m on the order of \( \mathcal {O}(k\log (n/k)) \).
Definition 2.2
(SRIP)[19] We say the matrix \(A=[a_1,\cdots ,a_m]^{\top }\in \mathbb {R}^{m\times n}\) has the Strong Restricted Isometry Property of order k and constants \(\theta _-,\ \theta _+\in (0, 2)\) if
holds for all k -sparse signals \(x\in \mathbb {R}^n\).
This property was first introduced in [19]. Voroninski and Xu also proved that the Gaussian random matrices satisfy SRIP with high probability.
Theorem 2.1
[19] Suppose that \(t>1\) and \( A\in \mathbb {R}^{m\times n} \) is a Gaussian random matrix with \(m=\mathcal {O}(tk\log (n/k))\). Then there exist \(\theta _-\), \(\theta _+\), with \(0<\theta _-<\theta _+<2\), such that A satisfies SRIP of order tk and constants \(\theta _-\), \(\theta _+\), with probability \(1-exp(-cm/2)\), where \(c>0\) is an absolute constant and \(\theta _-\), \(\theta _+\) are independent of t .
3 The Null Space Property
In this section, for any \( x_0\in D\Sigma _k^N \), we consider the noiseless case of (1.4),
Similarly as the traditional compressed sensing problem, we analyze the null space of the measurement matrix A to explore conditions under which (3.6) can obtain \( cx_0 \) (\( |c|=1 \)).
3.1 The Real Case
We first restrict the signals and measurements to the field of real numbers. The next theorem provides a sufficient and necessary condition for the exact recovery of (3.6).
Theorem 3.1
For a given matrix \( A\in {\mathbb R}^{m\times n} \) and a dictionary \( D\in {\mathbb R}^{n\times N} \), we claim that the following properties are equivalent.
-
(A)
For any \(x_0\in D\Sigma _k^N\),
$$\begin{aligned} argmin _{x\in {\mathbb R}^n}\{\Vert D^*x\Vert _1:|Ax|=|Ax_0|\}=\{\pm x_0\}. \end{aligned}$$ -
(B)
For any \(T\subseteq [1:m]\), it holds
$$\begin{aligned} \Vert D^*(u+v)\Vert _1<\Vert D^*(u-v)\Vert _1 \end{aligned}$$for all
$$\begin{aligned} u\in \mathcal {N}(A_T)\backslash \{0\} ,\quad v\in \mathcal {N}(A_{T^c})\backslash \{0\} \end{aligned}$$satisfying
$$\begin{aligned} u+v\in D\Sigma _k^N. \end{aligned}$$
Proof
(B)\(\Rightarrow \)(A). Assume (A) is false, namely, there exists a solution \( \hat{x}\ne \pm x_0 \) to (3.6). As \( \hat{x} \) is a solution, we have
and
Denote \(a_j^{\top }, j=1,\ldots ,m\) as the rows of A . Then (3.7) implies that there exists a subset \(T\subseteq [1:m]\) satisfying
i.e.,
Define
As \( \hat{x}\ne \pm x_0 \), we have \(u\in \mathcal {N}(A_T)\backslash \{0\}\), \(v\in \mathcal {N}(A_{T^c})\backslash \{0\}\) and \(u+v=2x_0\in D\Sigma _k^N\). Then from (B), we know
which contradicts with (3.8).
(A)\(\Rightarrow \)(B). Assume (B) is false, which means that there exists a subset \(T\subseteq [1:m]\),
such that
and
Let \( x_0:=u+v\in D\Sigma _k^N\) be the signal we want to recover. Set \(\tilde{x} :=u-v\) and we have \( \tilde{x}\ne \pm x_0 \). Then from (3.10) we have
Let \(a_j^{\top }, j=1,\ldots ,m\) denote the rows of A . Then from the definition of \( x_0 \) and \( \tilde{x} \), we have
By (3.9), the subset T satisfies
and
which implies
Putting (3.11) and (3.12) together, we know \( \tilde{x} \) is a solution to model (3.6). However, \(\tilde{x}\ne \pm x_0 \) contradicts with (A). \(\square \)
3.2 The Complex Case
We now consider the same problem in complex case which means that the signals and measurements are all in the complex number field. Let \( \mathcal {S}=\{S_1,\ldots , S_p\} \) be a partition of [1 : m] . The next theorem is a generalization of Theorem 3.1.
Theorem 3.2
For a given matrix \( A\in {\mathbb C}^{m\times n} \) and a dictionary \( D\in {\mathbb C}^{n\times N} \), we claim that the following two properties are equivalent.
-
(A)
For any given \(x_0\in D\Sigma _k^N\),
$$\begin{aligned} argmin _{x\in {\mathbb C}^n}\{\Vert D^*x\Vert _1:|Ax|=|Ax_0|\}=\{cx_0, c\in \mathbb {S}\}. \end{aligned}$$(B) Suppose \( \mathcal {S}=\{S_1,\ldots , S_p\}\) is a partition of [1 : m]. For any \(\eta _j\in \mathcal {N}(A_{S_j})\backslash \{0\}\), if
$$\begin{aligned} \frac{\eta _1-\eta _l}{c_1-c_l}=\frac{\eta _1-\eta _j}{c_1-c_j} \in D\Sigma _k^N\backslash \{0\},\,\,j,l \in [2:p],\,\,j\ne l \end{aligned}$$(3.13)holds for some pairwise distinct \(c_1,\ldots ,c_p\in \mathbb {S}\), we have
$$\begin{aligned} \Vert D^*(\eta _j-\eta _l)\Vert _1<\Vert D^*(c_l\eta _j-c_j\eta _l)\Vert _1. \end{aligned}$$
Proof
\((B)\Rightarrow (A)\). Suppose the statement (A) is false. That is to say, there exists a solution \( \hat{x}\notin \{cx_0, c\in \mathbb {S}\}\) to (3.6) which satisfies
and
Denote \(a_j^*, j=1,\ldots ,m\) as the rows of A . From (3.15) we have
with \( c_j\in \mathbb {S}, \,j=1,\ldots , m \). We can define an equivalence relation on [1 : m], namely \(j\sim l\), when \(c_j=c_l\). This equivalence relation leads to a partition \(\mathcal {S}=\{S_1,\ldots ,S_p\}\) of [1 : m]. For any \(S_j\), we have
Set \(\eta _j:=c_jx_0-\hat{x}\). Then we have \(\eta _j\in \mathcal {N}(A_{S_j})\backslash \{0\}\) and
According to the condition (B), we can get
i.e.,
That is equivalent to
which contradicts with (3.14).
\((A)\Rightarrow (B)\). Assume (B) is false, namely, there exists a partition \(\mathcal {S}=\{S_1,\ldots ,S_p\}\) of [1 : m], \(\eta _j\in \mathcal {N}(A_{S_j})\backslash \{0\}\), \( j\in [1:p] \) and some pairwise distinct \( c_1,\ldots ,c_p\in \mathbb {S} \) satisfying (3.13) but
holds for some distinct \(j_0, l_0\in [1:p]\). Set
Then we have
and
Let \(a_j^*, j=1,\ldots ,m\) denote the rows of A . From \( \eta _j\in \mathcal {N}(A_{S_j})\backslash \{0\} \), we obtain
The definition of \( x_0 \) and \( \tilde{x} \) implies
While for \(k\notin S_{l_0}\cup S_{j_0} \), we might as well suppose \(k\in S_t \) \( (t\ne l_0, j_0)\), i.e., \( \langle a_k, \eta _t\rangle = 0 \). From
we can obtain
here j, l, m, n are distinct integers. Set
Then we have
So \( \tilde{x} \) and \( x_0 \) can be rewritten as
Then \(\langle a_k,\eta _t\rangle =0\) implies
Using a similar argument, we can prove that the claim is also true for other subset \( S_j \). So we have
Combining (3.16) and (3.18), we know \(\tilde{x}\) is also a solution to (3.6). However, \( \tilde{x}\notin \{cx_0, c\in \mathbb {S}\} \) contradicts with (A). \(\square \)
Remark 3.1
If we choose \( D=I \), the null space property in Theorem 3.1 and Theorem 3.2 is consistent with the null space property which was introduced in paper [20].
According to the Theorems 3.1 and 3.2, if the measurement matrix satisfies the null space property, we can obtain an exact solution by solving model (3.6). But in general, condition (B) in Theorems 3.1 or 3.2 is difficult to be checked. So in Sect. 4, we provide another property (S-DRIP) of the measurement matrix which can also guarantee an exact recover of model (3.6) in noiseless case. In addition, we prove that this property can be satisfied by Gaussian random matrix.
4 S-DRIP and Stable Recovery
In compressed sensing, for any tight frame D , [7] had the conclusion that a signal \( x_0\in D\Sigma _k^N \) can be approximately reconstructed by \( \ell _1 \)-analysis (1.3) provided the measurement matrix satisfies DRIP and the best k -term approximation error of \( D^*x_0 \) is small. While in phase retrieval, when \( {\mathbb H}={\mathbb R}\), Gao et al. proved that if the measurement matrix satisfies SRIP, then the \( \ell _1 \)-analysis (1.5) can provide a stable solution to traditional phase retrieval problem [12]. For the phase retrieval with redundant dictionary, we combine the above two results to explore the conditions under which the \( \ell _1 \)-analysis model (1.4) can guarantee a stable recovery.
We first impose a natural property on the measurement matrix, which is a combination of DRIP and SRIP.
Definition 4.1
(S-DRIP) Let \( D\in {\mathbb R}^{n\times N} \) be a frame. We say the measurement matrix A obeys the S-DRIP of order k with constants \(\theta _-, \theta _+\in (0,2)\) if
holds for all k -sparse signals \(v\in {\mathbb R}^N\).
Thus a matrix \( A\in \mathbb {R}^{m\times n} \) satisfying S-DRIP means that any \( m'\times n \) submatrix of A , with \( m'\ge m/2 \) satisfies DRIP with appropriate parameters.
In fact any matrix \( A\in {\mathbb R}^{m\times n} \) obeying
(\( 0<c_-<c_+<2 \) and \( \gamma \) is a positive number constant) for fixed \( Dv\in {\mathbb R}^n \) will satisfy the S-DRIP with high probability. This can be seen by a standard covering argument (see the proof of Theorem 2.1 in [19]). In [19], Voroninski and Xu proved that Gaussian random matrix satisfies (4.19) in Lemma 4.4. So we have the following conclusion.
Corollary 4.1
For \( t>1 \), Gaussian random matrix \( A\in {\mathbb R}^{m\times n} \) with \( m=\mathcal {O}(tk\log (n/k)) \) satisfies the S-DRIP of order tk and constants \(\theta _-, \theta _+\in (0,2)\) with probability \( 1-2e^{-\gamma m} \), where \( \gamma \) is an absolute positive constant and \( \theta _-, \theta _+ \) are independent of t .
For any \( x_0\in D\Sigma _k^N \), we return to consider the solving model
where \( \epsilon \) is the error bound. Here signals and matrices are all restricted to the real number field. The next theorem tells under what conditions the solution to (4.20) is stable.
Theorem 4.1
Assume that \( D\in {\mathbb R}^{n\times N} \) is a tight frame and \( x_0\in D\Sigma _k^N \). The matrix \(A\in \mathbb {R}^{m\times n}\) satisfies the S-DRIP of order tk and level \( \theta _-, \theta _+ \in (0,2)\), with
Then the solution \( \hat{x} \) to (4.20) satisfies
where \( c_1=\frac{\sqrt{2(1+\delta )}}{1-\sqrt{t/(t-1)}\delta } \), \( c_2=\frac{\sqrt{2}\delta +\sqrt{t(\sqrt{(t-1)/t}-\delta )\delta }}{t(\sqrt{(t-1)/t}-\delta )}+1.\) Here \( \delta \) is a constant satisfying
We first give a more general lemma, which is the key to prove Theorem 4.1.
Lemma 4.1
Let \( D\in \mathbb {R}^{n\times N} \) be an arbitrary tight frame, \( x_0\in D\Sigma _k^N\) and \( \rho \ge 0\). Suppose that \( A\in \mathbb {R}^{m\times n} \) is a measurement matrix satisfying the DRIP with \( \delta = \delta _{tk}^A\le \sqrt{\frac{t-1}{t}} \) for some \( t>1 \). Then for any
we have
where \( c_1=\frac{\sqrt{2(1+\delta )}}{1-\sqrt{t/(t-1)}\delta } \), \( c_2=\frac{\sqrt{2}\delta +\sqrt{t(\sqrt{(t-1)/t}-\delta )\delta }}{t(\sqrt{(t-1)/t}-\delta )}+1.\)
We put the proof of this Lemma in the Appendix.
Remark 4.1
When \( D=I \), which corresponds to the case of standard compressive phase retrieval, Theorem 4.1 and Lemma 4.1 are consistent with Theorem 3.1 and Lemma 2.1 in [12], respectively.
Remark 4.2
The DRIP constant in Lemma 4.1 is better than the DRIP constants given in [2] and [7]. In [7], Candès et al. proved that the \( l_1 \)-analysis (1.3) can guarantee a stable recovery of signals which are k -sparse in the tight frame D provided the measurement matrix satisfying DRIP with \( \delta _{2k} < 0.08 \). Then Baker improved the result by increasing the DRIP constant to \( \delta _{2k}<\frac{2}{3}\) in [2]. Here we extended Baker’s approach to get a better bound \( \delta _{tk}\le \sqrt{\frac{t-1}{t}} \) for \( t>1 \). As [6] shows, in the special case \( D=I \), for any \( t\ge 4/3 \), the condition \( \delta _{tk}\le \sqrt{\frac{t-1}{t}} \) is sharp for stable recovery in the noisy case. So it is not difficult to conclude that for any tight frame D , the condition \( \delta _{tk}\le \sqrt{\frac{t-1}{t}} \) is also sharp when \( t\ge 4/3 \).
Proof of the Theorem 4.1
As \(\hat{x}\) is the solution to (4.20), we have
and
Denote \( a_j^{\top }, j\in \{1,\ldots ,m\} \) as the rows of A and divide \(\{1,\ldots ,m\}\) into two groups:
Then either \(|T|\ge m/2\) or \(|T^c|\ge m/2\). Without loss of generality, we suppose \(|T|\ge m/2\) .
Then (4.22) implies that
Combining (4.21) and (4.23), we have
Recall that A satisfies S-DRIP of order tk with constants \(\theta _-, \ \theta _+ \in (0,2)\). Here
So \(A_T\) satisfies DRIP of order tk with
Combining (4.24), (4.25) and Lemma 4.1, we obtain
where \(c_1\) and \(c_2\) are defined as before in the Theorem 4.1.
If \(|T^c|\ge \frac{m}{2}\), we can get the corresponding result
Then we have proved the theorem. \(\square \)
According to Theorem 4.1, when \( \epsilon =0 \) and \( D^*x_0 \) is k -sparse, the \( \ell _1 \)-analysis (4.20) can provide an exact recovery of the phase retrieval with redundant dictionary (1.2) provided the measurement matrix satisfies S-DRIP. Meanwhile, from Theorem 4.1 and Corollary 4.1, we conclude that the \( \ell _1 \)-analysis (4.20) can provide a stable solution to problem (1.2) if we use as many as \( \mathcal {O}(k \log (n/k)) \) Gaussian random measurements.
5 Discussion
To solve the phase retrieval with redundant dictionary (1.2), we analyze the \( \ell _1 \)-analysis model and give two conditions on the measurement matrix that each of them can guarantee an exact recovery in noiseless case. Theorems 3.1 and 3.2 give the null space property as a sufficient and necessary condition for exact recovery. For the \( \ell _1 \)-synthesis model, we can also use the same analysis to give a null space property of the measurement matrix. A more detailed description of the \( \ell _1 \)-synthesis model is provided in [8]. Theorem 4.1 shows that the \( \ell _1 \)-analysis model is accurate when the measurement matrix satisfies S-DRIP and \( \Vert D^*x_0\Vert _0\le k \). In theory, the \( \ell _1 \)-analysis model has a good performance on solving the phase retrieval with redundant dictionary (1.2). However, the \( \ell _1 \)-analysis is a non-convex optimization for phase retrieval with redundant dictionary due to the non-convex feasible solution set. When \( D=I \), the algorithms of this model have been studied in [15, 17, 22]. These algorithms all demonstrate empirical success, but the convergence issue remains a difficult problem. Extending these algorithms to a redundant dictionary D and giving a convergence analysis is one direction of our future research. Another key drawback of our results is that Theorem 4.1 only holds in the real number field. The point is that the phase changes continuously and there is no proper definition of SRIP in the complex number field. The extension of this result to complex number field is another direction of our future work.
References
Aldroubi, A., Chen, X., Powell, A.M.: Perturbations of measurement matrices and dictionaries in compressed sensing. Appl. Comput. Harmon. Anal. 33(2), 282–291 (2012)
Baker, C.A.: A note on sparsification by frames. arXiv:1308.5249 (2013)
Balan, R., Casazza, P., Edidin, D.: On signal reconstruction without phase. Appl. Comput. Harmon. Anal. 20(3), 345–356 (2006)
Bodmann, B.G., Hammen, N.: Stable phase retrieval with low-redundancy frame. Adv. Comput. Math. 41(2), 317–331 (2015)
Cai, T.T., Zhang, A.: Sharp RIP bound for sparse signal and low-rank matrix recovery. Appl. Comput. Harmon. Anal. 35(1), 74–93 (2013)
Cai, T.T., Zhang, A.: Sparse representation of a polytope and recovery of sparse signals and low-rank matrices. IEEE Trans. Inf. Theory 60(1), 122–132 (2014)
Candès, E.J., Eldar, Y.C., Needell, D., Randall, P.: Compressed sensing with coherent and redundant dictionaries. Appl. Comput. Harmon. Anal. 31(1), 59–73 (2011)
Chen, X.M., Wang, H.C., Wang, R.R.: A null space analysis of the \(\ell _1\)-synthesis method in dictionary-based compressed sensing. Appl. Comput. Harmon. Anal. 37(3), 492–515 (2014)
Conca, A., Edidin, D., Hering, M., Vinzant, C.: An algebraic characterization of injectivity in phase retrieval. Appl. Comput. Harmon. Anal. 38(2), 346–356 (2015)
Elad, M., Milanfar, P., Rubinstein, R.: Analysis versus synthesis in signal priors. Inverse Probl. 23(3), 947968 (2007)
Fienup, J.: Reconstruction of a complex-valued object from the modulus of its Fourier transform using a support constraint. J. Opt. Soc. Am. 4(1), 118–123 (1987)
Gao, B., Wang, Y., Xu, Z.Q.: Stable signal recovery from phaseless measurements. J. Fourier Anal. Appl. 22(4), 787–808 (2016)
Herman, M., Strohmer, T.: High resolution radar via compressed sensing. IEEE Trans. Signal Process. 57(6), 22752284 (2009)
Kohler, D., Mandel, L.: Source reconstruction from the modulus of the correlation function: a practical approach to the phase problem of optical coherence theory. J. Opt. Soc. Am. 63(2), 126–134 (1973)
Moravec, M., Romberg, J., Baraniuk, R.: Compressive phase retrieval. Proc. SPIE 6701, 670120–670120 (2007). doi:10.1117/12.736360
Rauhut, H., Schnass, K., Vandergheynst, P.: Compressed sensing and redundant dictionaries. IEEE Trans. Inf. Theory 54(5), 2210–2219 (2008)
Schniter, P., Rangan, S.: Compressive phase retrieval via generalized approximate message passing. IEEE Trans. Signal Process. 63(4), 1043–1055 (2015)
Shechtman, Y., Eldar, Y.C., Szameit, A., Segev, M.: Sparsity based sub-wavelength imaging with partially incoherent light via quadratic compressed sensing. Opt. Express 19(16), 14807–14822 (2011)
Voroninski, V., Xu, Z.Q.: A strong restricted isometry property, with an application to phaseless compressed sensing. Appl. Comput. Harmon. Anal. 40(2), 386–395 (2016)
Wang, Y., Xu, Z.Q.: Phase retrieval for sparse signals. Appl. Comput. Harmon. Anal. 37(3), 531–544 (2014)
Xu, G.W., Xu, Z.Q.: On the \(\ell _1\)-norm invariant convex \( k \)-sparse decomposition of signals. J. Oper. Res. Soc. China 1(4), 537–541 (2013)
Yang, Z., Zhang, C.S., Xie, L.H.,: Robust compressive phase retrieval via L1 minimization with application to image reconstruction. arXiv:1302.0081 (2013)
Acknowledgments
My deepest gratitude goes to Professor Zhiqiang Xu, my academic supervisor, for his guidance and many useful discussions.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Roman Vershynin.
Appendix
Appendix
The following two lemmas are useful in the proof of Lemma 4.1.
Lemma 6.1
(Sparse representation of a polytope [6, 21]): Suppose \(\alpha >0\) is a constant and \(s>0\) is an integer. Set
For any \(v\in \mathbb {R}^n\), set
Then \(v\in T(\alpha ,s)\) if and only if v is in the convex hull of \( U (\alpha ,s,v)\). In particular, any \(v\in T(\alpha ,s)\) can be expressed as
Lemma 6.2
(Lemma 5.3 in [5]): Suppose \(m\ge r \), \( a_1\ge a_2\ge \cdots \ge a_m\ge 0 \) and \( \sum _{i=1}^{r}a_i\ge \sum _{i=r+1}^{m}a_i \). Then for all \( \alpha \ge 1 \), we have
Now we are ready to prove Lemma 4.1.
Proof of the Lemma 4.1
We assume that the tight frame \( D\in {\mathbb R}^{n\times N} \) is normalized, i.e., \( DD^*=I \) and \( \Vert y\Vert _2=\Vert D^*y\Vert _2 \) for all \( y\in \mathbb {R}^n \). For a subset \( T\subseteq \{1,2,\ldots ,N\} \), we denote \( D_T \) as the matrix D restricted to the columns indexed by T (replacing other columns by zero vectors).
Set \(h:=\hat{x}-x_0\). Let \(T_0\) denote the index set of the largest k coefficients of \(D^*x_0\) in magnitude. Then
which implies
Suppose \( S_0 \) is the index set of the k largest entries in absolute value of \( D^*h \). We get
Set
We divide \( D^*_{S_0^c }h\) into two parts \( D^*_{S_0^c }h=h^{(1)}+h^{(2)} \), where
Then a simple observation is that \( \Vert h^{(1)}\Vert _1\le \Vert D^*_{S_0^c}h\Vert _1\le \alpha k \). Set
Since all non-zero entries of \( h^{(1)} \) have magnitude larger than \( \alpha /(t-1) \), we have
which implies \( \ell \le (t-1)k \).
Note that
Then in Lemma 6.1, by setting \( s:=k(t-1)-\ell \), we can express \( h^{(2)} \) as a weighted mean:
where \( 0\le \lambda _i\le 1 \), \( \sum _{i=1}^{M}\lambda _i=1\), \( \Vert u_i\Vert _0\le k(t-1)-\ell \), \(\Vert u_i\Vert _\infty \le \alpha /(t-1) \) and \(supp (u_i)\subseteq supp (h^{(2)}) \). Thus
Recall that \(\alpha =\frac{\Vert D^*_{S_0}h\Vert _1+2\sigma _k(D^*x_0)_1+\rho }{k}\). Then
where \( z:=\Vert D^*_{S_0}h+h^{(1)}\Vert _2, \,\, R:=\frac{2\sigma _k(D^*x_0)_1+\rho }{\sqrt{k}}\).
Now we suppose \( 0\le \mu \le 1 \), \( d\ge 0\) are two constants to be determined. Set
Then for any fixed \(i\in [1:M]\),
For \(\sum _{i=1}^M\lambda _i=1\), we have the following identity
In (6.27), we chose \( d=1/2 \) and \( \mu =\sqrt{t(t-1)}-(t-1)< 1/2 \). Then (6.28) implies
We next estimate the three terms in (6.29). First we give the following useful relation:
Noting that \(\Vert D^*_{S_0}h\Vert _0\le k \), \( \Vert h^{(1)}\Vert _0= \ell \le (t-1)k \) and \(\Vert u_i\Vert _0\le s =k(t-1)-\ell \), we obtain
and
Here we assume \( t\cdot k \) as an integer first. Since A satisfies the DRIP of order \(t\cdot k\) with constant \(\delta \), we can obtain
and
Combining the above results with (6.26) and (6.29), we get
which is a quadratic inequality for z. Recall that \( \delta <\sqrt{(t-1)/t} \). So by solving the above inequality, we get
We know \( \Vert D^*_{S_0^c}h\Vert _1\le \Vert D^*_{S_0}h\Vert _1+R\sqrt{k} \). In the Lemma 6.2, if we set \( m=N \), \( r=k \), \( \lambda =R\sqrt{k}\ge 0 \) and \( \alpha =2 \), we can obtain
So
Substituting R into this inequality, we can get the conclusion. For the case where \(t\cdot k\) is not an integer, we set \(t^*:=\lceil tk\rceil / k\), then \(t^*>t\) and \(\delta _{t^*k}=\delta _{tk}<\sqrt{\frac{t-1}{t}}<\sqrt{\frac{t^*-1}{t^*}}\). We can prove the result by working on \(\delta _{t^*k}\). \(\square \)
Rights and permissions
About this article
Cite this article
Gao, B. The \( \ell _1 \)-Analysis in Phase Retrieval with Redundant Dictionary. J Fourier Anal Appl 23, 1097–1117 (2017). https://doi.org/10.1007/s00041-016-9500-z
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00041-016-9500-z