Abstract
In this paper, we discuss the following rank-constrained matrix approximation problem in the Frobenius norm: \(\Vert C-AX\Vert =\min \) subject to \( \text{ rk }\left( {C_1 - A_1 X} \right) = b \), where b is an appropriate chosen nonnegative integer. We solve the problem by applying the classical rank-constrained matrix approximation, the singular value decomposition, the quotient singular value decomposition and generalized inverses, and get two general forms of the least squares solutions.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
In this paper, we adopt the following notation. The symbol \({\mathbb {C}}^{m\times n}\) (\({\mathbb {U}}^{m\times m}\)) denotes the set of all \(m \times n\) complex matrices (\({m\times m}\) unitary matrices), \(I_k\) denotes the \(k\times k\) identity matrix, 0 denotes a zero matrix of appropriate size, and \(\Vert \cdot \Vert \) stands for the matrix Frobenius norm. For \(A\in {\mathbb {C}}^{m\times n}\), \(A^H\) and \(\mathrm{rk}(A)\) stand for the conjugate transpose and the rank of A, respectively.
The Moore-Penrose inverse of \(A\in {\mathbb {C}}^{m\times n}\) is defined as the unique matrix \(X\in {\mathbb {C}}^{n\times m}\) satisfying
and is usually denoted by \(X = A^{\dag }\) (see [1]). The symbol \(A\left\{ {i, \ldots ,j} \right\} \) is the set of matrices \(X \in {{\mathbb {C}}}^{n\times m}\) which satisfies equations \(\left( i \right) , \ldots ,\left( j \right) \) from among equations \(\left( 1\right) \)–\(\left( 4\right) \). A matrix \(X \in A\left\{ {i, \ldots ,j} \right\} \) is called an \(\left\{ {i, \ldots ,j} \right\} \)-inverse of A, and is denoted by \(A^{\left( {i, \ldots ,j} \right) }\). It is well known that \(AA^{(1,3)}=AA^{\dag }\). Furthermore, we denote
Given that \(A\in {\mathbb {C}}^{m\times n}\), \(B \in {\mathbb {C}}^{m\times k}\), and \(C \in {\mathbb {C}}^{l\times n}\), the column block matrix consisting of A and B is denoted by \(\left( \begin{matrix}A,&B \end{matrix}\right) \), and its rank is denoted by \(\mathrm{rk} \left( \begin{matrix}A,&B \end{matrix}\right) \); the row block matrix consisting of A and C is denoted by \(\left( {\begin{matrix}A \\ C \end{matrix}}\right) \), and its rank is denoted by \(\mathrm{rk} \left( {\begin{matrix}A \\ C \end{matrix}}\right) \). It is well known that \(\Vert \left( \begin{matrix}A,&B \end{matrix}\right) \Vert ^2= \Vert A\Vert ^2+\Vert B\Vert ^2\). The two known formulas for ranks of block matrices are given in [11],
In the literature, the minimum rank matrix approximations or rank-constrained matrix approximations have been widely studied [1,2,3, 5, 6, 8, 9, 12, 14,15,16,17, 20, 22, etc]. Recently, Friedland and Torokhti [6] studied the problem of finding least square solutions to the equation \( BXC=A \) subject to \( {\mathrm{rk}\left( X \right) \le k}\) in the Frobenius norm by applying SVD; Wei and Wang [21] studied the problem of finding rank-k Hermitian nonnegative definite least squares solutions to the equation \(BXB^H=A\) in the Frobenius norm and discussed the ranges of the rank k; Sou and Rantzer [15] studied the minimum rank matrix approximation problem in the spectral norm \( \min \limits _{X}\mathrm{rk}(X)\) subject to \(\Vert A-BXC\Vert _2<1\); Wei and Shen [22] studied a more general problem \(\min \limits _{X}\mathrm{rk}(X)\) subject to \(\Vert A-BXC\Vert _2<\xi \), where \(\xi \ge \theta \) and \(\theta =\min \limits _Y\Vert A-BYC\Vert _2\), by applying SVD and R-SVD; Tian and Wang [18] gave a least-squares solution of \(AXB = C\) subject to \(\mathrm{rk}\!\left( AXB-C\right) =\min \) in the Frobenius norm. On the other hand, the minimum rank matrix approximations or rank-constrained matrix approximations have found many applications in control theory [4, 15, 22], signal process [6] and numerical algebra [3, 8], etc.
Note that Golub et al. [8] studied the problem of finding rank-constrained least square solutions to the equation \( {\left( {{\begin{matrix} A,&X \end{matrix} }} \right) = \left( {{\begin{matrix} A,&B \end{matrix} }} \right) } \) subject to \( \mathrm{rk} \left( {{\begin{matrix} A , &{} X \\ \end{matrix} }} \right) \le k\) in all unitarily invariant norms by applying SVD and QR decomposition; Demmel [3] considered the least square solutions to \( \left( {{{\begin{matrix} A &{} B \\ C &{} X \\ \end{matrix}} }} \right) = \left( {{{\begin{matrix} A &{} B \\ C &{} D \\ \end{matrix}} }} \right) \) for X subject to \( \mathrm{rk}\!\left( {{{\begin{matrix} A &{} B \\ C &{} X \\ \end{matrix}} }} \right) \le k\) in the Frobenius norm and the 2-norm; Wang [19] studied a general problem of determining the least squares solution X of \( \left( {{{\begin{matrix} {X} &{} J \\ K &{} L \\ \end{matrix}} }} \right) = \left( {\begin{matrix} A &{} B \\ C &{} D \\ \end{matrix}} \right) \) subject to \( \mathrm{rk}\!\left( {\begin{matrix} X &{} J \\ K &{} L \\ \end{matrix}} \right) =k\) in the Frobenius norm.
In [3, 6, 8], a commonly assumption is that the rank is less than or equal to k. In fact, in most situation, the rank is equal to k. For instance, consider the descriptor linear system
Applying a full-state derivative feedback controller \(u(t)=-K\dot{x}(t)\) to system (1.2), we have the closed-loop system \( (E+BK)\dot{x}(t)=Ax(t)\). The dynamical order is defined to be \(\mathrm{rk} \left( E+BK\right) =p\). One of the minimum gain problems is characterize the set
Therefore, Duan [4] studied the problem of finding rank-k least square solutions to \(BX=A\); Liu et al. [10] considered the problem \(\min \limits _{\mathrm{rk} \left( X\right) =k}\Vert A-BXB^H\Vert \), in which A and X are (skew) Hermitian matrices. In this paper, we study a more general problem by applying SVD and Q-SVD. Assume that b is a prescribed nonnegative integer, \(A \in {{\mathbb {C}}}^{m\times n}\), \(A_1 \in {{\mathbb {C}}}^{w \times n}\), \(C\in {{\mathbb {C}}}^{m\times p}\) and \(C_1 \in {{\mathbb {C}}}^{w \times p}\) are given matrices. We now investigate the problem of determining the least squares solution X of the matrix equation \( AX=C \) subject to \( \mathrm{rk}\left( {C_1 - A_1 X} \right) = b\) in the Frobenius norm. This problem can be stated as follows.
Problem 1.1
Suppose that \(A \in {{\mathbb {C}}}^{m\times n}\), \(A_1 \in {{\mathbb {C}}}^{w \times n}\), \(C\in {{\mathbb {C}}}^{m\times p}\) and \(C_1 \in {{\mathbb {C}}}^{w \times p}\) are given matrices. For an appropriate chosen nonnegative integer b, characterize the set
2 Preliminaries
In this section, we mention the following results for our further discussions.
Lemma 2.1
[4] Given that \({\mathcal {X}}_1 \in {\mathbb {C}}^{s \times n_1 }\) and the integer \(k_1 \) satisfying \(0 \le k_1 \le \min \left\{ {m_1 ,n_1 } \right\} \), then there exists \({\mathcal {X}}_2 \in {\mathbb {C}}^{(m_1- s ) \times n_1 }\) satisfying
if and only if
Lemma 2.2
Suppose that \(H \in {\mathbb {C}}^{m\times n} \), \(\mathrm{rk}(H)=r\), l is a given nonnegative integer with \(l \le r\), the decomposition of H is
where \(G\in {\mathbb {C}}^{m_1\times n_1}\), \(\mathrm{rk}(G)=r\), \(k\le m_1\le m\), \(k\le n_1\le n\), U and V are unitary matrices of appropriate sizes. Then
where
and
The following result is the classical rank-constrained matrix approximation of Eckart and Young [5], and Mirsky [12].
Lemma 2.3
Suppose that \({\mathcal {C}} \in {\mathbb {C}}^{ s \times n_1} \) with \(\mathrm{rk}({\mathcal {C}})=c\), \(c_1\) is a given nonnegative integer with \({c_1} \le c\). Let the SVD [7] of \({\mathcal {C}}\) be
where \(\Lambda = \text{ diag }\left\{ {\lambda _1 ,\ldots ,\lambda _c} \right\} \), \(\lambda _1 \ge \cdots \ge \lambda _c > 0\), \({\mathcal {U}}\) and \({\mathcal {V}}\) are unitary matrices of appropriate sizes. Then
Furthermore, when \(\lambda _{c_1} > \lambda _{{c_1} + 1} \),
when \(p_2< {c_1} < {p_1} \le r\) and \(\lambda _{p_2}> \lambda _{p_2 + 1} = \cdots = \lambda _{p_1} > \lambda _{{p_1} + 1}\),
and \({\mathcal {Q}}\) is an arbitrary matrix satisfying \({\mathcal {Q}} \in {\mathbb {C}}^{\left( {{p_1}- {p_2}} \right) \times \left( {{c_1} - p_2} \right) }\) and \({\mathcal {Q}}^H{\mathcal {Q}} = I_{{c_1} - {p_2}}\).
Suppose that \({\mathcal {X}}_{1} \in { {\mathbb {C}}}^{s\times n_1}\), \({\mathcal {X}}_{2} \in { {\mathbb {C}}}^{(m_1- s )\times {n_1} }\), \({ \text{ rk }\left( { {{{\begin{matrix} {{\mathcal {X}}_1 } \\ {{\mathcal {X}}_2 } \, \end{matrix}} }}} \right) = k_1}\) and \(k_1\le \min \{m_1, n_1\}\) be a given nonnegative integer, it is easy to check that
Suppose that \({\mathcal {C}} \in {\mathbb {C}}^{ s \times n_1} \) with \(\mathrm{rk}({\mathcal {C}})=c\). Consider the rank-constrained matrix approximation
under the condition \(k_1 \le \min \{c{ + }(m_1- s ), n_1\}\). We have the following Lemma 2.4 by applying Lemma 2.1 and Lemma 2.3 to (2.2).
Lemma 2.4
Suppose that \({\mathcal {C}} \in {\mathbb {C}}^{ s \times n_1} \) with \(\mathrm{rk}({\mathcal {C}})=c\), \(k_1\) is a given nonnegative integer with \(0 \le k_1 \le \min \left\{ c{ + }(m_1- s ), n_1\right\} \), and the SVD of \({\mathcal {C}}\) be given as in Lemma 2.3, then,
- (a)
if \(c \le k_1 \le \min \left\{ c{ + }(m_1- s ), n_1\right\} \),
$$\begin{aligned} \mathop {\min } \limits _{{\mathcal {X}}_1, { \text{ rk }\left( { {{{\begin{matrix} {{\mathcal {X}}_1 } \\ {{\mathcal {X}}_2 } \, \end{matrix}} }}} \right) = k_1}} \left\| {{\mathcal {C}} - {\mathcal {X}}_1 }\right\| = 0, \end{aligned}$$and
$$\begin{aligned} \left( {{\begin{matrix} {{\mathcal {X}}_1 } \\ {{\mathcal {X}}_2 } \\ \end{matrix} }} \right) = \left( {{\begin{matrix} {{\mathcal {C}} } \\ {\left( {{\begin{matrix} {{\mathcal {X}}_{21} } , &{} {{\mathcal {X}}_{22} } \\ \end{matrix} }} \right) {\mathcal {V}}^H} \\ \end{matrix} }} \right) , \end{aligned}$$where \({\mathcal {X}}_{21} \in { {\mathbb {C}}}^{(m_1- s )\times c}\), \({\mathcal {X}}_{22} \in { {\mathbb {C}}}^{(m_1- s )\times \left( {n_1- c}\right) }\) and \( \text{ rk }\left( {{\mathcal {X}}_{22} } \right) = k_1 - c\).
- (b)
if \(0\le k_1 < c\),
$$\begin{aligned} \mathop {\min } \limits _{{\mathcal {X}}_1, { \text{ rk } \left( { {{{\begin{matrix} {{\mathcal {X}}_1 } \\ {{\mathcal {X}}_2 } \\ \end{matrix}} }}} \right) = k_1}} \left\| {{\mathcal {C}} - {\mathcal {X}}_1 }\right\| = \left( {\sum \limits _{i = k_1 +1}^c {\lambda _i^2 } }\right) ^{\frac{1}{2}}, \end{aligned}$$and
$$\begin{aligned} \left( {{\begin{matrix} {{\mathcal {X}}_1 } \\ {{\mathcal {X}}_2 } \\ \end{matrix} }} \right) = \left( \begin{matrix} {\mathcal {U}} &{}\quad 0 \\ 0 &{}\quad I_{m_1- s} \\ \end{matrix} \right) \left( \begin{matrix} {\Lambda _1 } &{}\quad 0 \\ 0 &{}\quad 0 \\ {{\mathcal {X}}_{21} } &{}\quad 0 \\ \end{matrix} \right) {\mathcal {V}}^H, \end{aligned}$$where \({\mathcal {X}}_{21} \in { {\mathbb {C}}}^{(m_1- s )\times c}\), when \(\lambda _{k_1 } > \lambda _{k_1 + 1} \),
$$\begin{aligned} \Lambda _1 = \mathrm{diag}\left\{ {\lambda _1 ,\ldots ,\lambda _{k_1 } } \right\} ; \end{aligned}$$when \(q_2< k_1 < q_1 \le r\) and \(\lambda _{q_2 }>\lambda _{q_2 + 1} = \ldots = \lambda _{q_1 } > \lambda _{q_1 + 1} \),
$$\begin{aligned} \Lambda _1 = \mathrm{diag}\left\{ {\lambda _1 ,\ldots ,\lambda _{q_2 } ,\lambda _{k_1 } {\mathcal {QQ}}^H}\right\} , \end{aligned}$$in which \({\mathcal {Q}}\) is an arbitrary matrix satisfying \({\mathcal {Q}} \in {\mathbb {C}}^{\left( {q_1 -q_2 } \right) \times \left( {k_1 - q_2 }\right) }\) and \({\mathcal {Q}}^H{\mathcal {Q}} = I_{k_1 - q_2 } \).
Lemma 2.5
[13] Suppose that \(A \in {{\mathbb {C}}}^{m\times n}\), \(A_1 \in {{\mathbb {C}}}^{w \times n}\), \(D^H=\left( \begin{matrix} A^H ,&{} A_1^H \\ \end{matrix} \right) \) and \(k=\mathrm{rk}(D)\), then there exist \(U \in {{\mathbb {U}}}^{m\times m} \) and \(V \in {{\mathbb {U}}}^{w\times w}\) and a nonsingular matrix \(W \in {{\mathbb {C}}}^{n\times n}\) such that
where \(r=k-{\mathrm{rk}}(A_1)\), \(s={\mathrm{rk}}(A)+{\mathrm{rk}}(A_1)-k\),
in which \(S_1\) and \({{\widehat{S}}_1 } \) are both positive diagonal matrices.
If \(S_1= { \mathrm diag}(\alpha _1,\alpha _2,\ldots ,\alpha _s)\) and \({{\widehat{S}}_1 }=\mathrm{diag}(\beta _1,\beta _2,\ldots ,\beta _s)\) satisfy \(1>\alpha _1 \ge \ldots \ge \alpha _s>0\), \(1>\beta _s\ge \ldots \ge \beta _1>0\), \(\alpha _i^2+\beta _i^2=1\), \(i=1,\ldots ,s\), and there exists a positive diagonal matrix \(\Sigma _2={ \mathrm diag}\left( \sigma _1(D), \ldots ,\sigma _k(D)\right) \), in which \(\sigma _1(D), \ldots ,\sigma _k(D) \) are the positive singular values of D, and two unitary matrices \(P\in {\mathbb {C}}^{k\times k}\) and \(Q\in {\mathbb {C}}^{n\times n}\) satisfy
then (2.3) is the well-known Q-SVD of A and \(A_1\).
Denoting \(A^{-}=W^{-1}\Sigma ^{\dag } U^{H}\) and \(A_1^{-}=W^{-1}\Sigma _1^{\dag } V^{H}\), we know that \(A^{-}\in A\{1,3\}\), so it suffices to check that \(AA^{-}=A A^{\dag }\).
3 Solutions to Problem 1.1
In this section, we solve Problem 1.1 proposed in Sect. 1, and get two general forms of the least squares solutions.
Theorem 3.1
Suppose that \(A \in {{\mathbb {C}}}^{m\times n}\), \(A_1 \in {{\mathbb {C}}}^{w \times n}\), \(C\in {{\mathbb {C}}}^{m\times p}\), \(C_1 \in {{\mathbb {C}}}^{w \times p}\), k, r, s, and the decompositions of A and \(A_1\) are as in Lemma 2.5. Partition
Let t denote the rank of \(C_{11}\), and let the SVD of \(C_{11} \in {{\mathbb {C}}}^{\left( {w - k + r} \right) \times p}\) be
where \(\mathrm{T}\in {{\mathbb {C}}}^{ t \times t}\) is a nonsingular matrix, \(U_1 \in {\mathbb {U}}_{w - k + r} \) and \(V_1 \in {\mathbb {U}}_{p }\). Partition
Also suppose that the SVD of \({\mathcal {C}}\) is given in (2.1), and denotes \(\mathrm{rk}({\mathcal {C}})=c\). Then there exists a matrix \(X\in {\mathbb {C}}^{n\times p}\) satisfying (1.3) if and only if
If \(c+t \le b \le \min \left\{ \mathrm{rk}\left( A_1,\ \ C_1\right) , c+t+k-r-s, p\right\} \), then
and a general form for X which satisfies (1.3) is
where \({\mathcal {Z}} \in {{\mathbb {C}}}^{\left( {n - k}\right) \times p}\), \({\mathcal {Y}} \in { {\mathbb {C}}}^{(k-r- s )\times t}\) and \({\mathcal {X}}_{21} \in { {\mathbb {C}}}^{(k-r- s )\times c}\) are arbitrary matrices, and \({\mathcal {X}}_{22} \in { {\mathbb {C}}}^{(k-r- s )\times \left( {p-t- c}\right) }\) satisfies \( \text{ rk }\left( {{\mathcal {X}}_{22} } \right) = b-t - c\).
If \(t \le b < c+t\), then
and a general form for X which satisfies (1.3) is
where \({\mathcal {Z}} \in {{\mathbb {C}}}^{\left( {n - k}\right) \times p}\), \({\mathcal {Y}} \in { {\mathbb {C}}}^{(k-r- s )\times t}\) and \({\mathcal {X}}_{21} \in { {\mathbb {C}}}^{(k-r- s )\times c}\) are arbitrary matrices, and when \(\lambda _{b-t} > \lambda _{ b-t + 1} \),
when \(q_2< b-t < q_1 \le r\) and \(\lambda _{q_2 }>\lambda _{q_2 + 1} = \ldots = \lambda _{q_1 } > \lambda _{q_1 + 1} \),
in which \({\mathcal {Q}}\) is an arbitrary matrix satisfying \({\mathcal {Q}} \in {\mathbb {C}}^{\left( {q_1 -q_2 } \right) \times \left( {b-t - q_2 }\right) }\) and \({\mathcal {Q}}^H{\mathcal {Q}} = I_{b-t - q_2 } \).
Proof
Partition
Then from (3.2) and (3.3), we have
According to (3.10),
Hence,
Denoting \(k_1=b-t\), we obtain
where \(Y \in {{\mathbb {C}}}^{\left( {k - r} \right) \times \left( {p -t} \right) }\) satisfies \(\text{ rk }\left( Y\right) = k_1\). Furthermore, a general form for X which satisfies \( \text{ rk }\left( {C_1 - A_1 X} \right) =b\) is
where \(X_1 \in {{\mathbb {C}}}^{r\times p}\), \({\mathcal {Z}} \in {{\mathbb {C}}}^{\left( {n - k}\right) \times p}\) and \(X_{21} \in {{\mathbb {C}}}^{\left( {k - r} \right) \times t}\) are arbitrary matrices, and \(Y \in {{\mathbb {C}}}^{\left( {k - r} \right) \times \left( {p -t} \right) }\) satisfies \(\text{ rk }\left( Y \right) = k_1\).
Applying the decomposition (2.3) of A and (3.13), we gain
Since the Frobenius norm of a matrix is invariant under unitary transformation, by applying (3.14), we obtain
It is easily to find that
and the matrix \(X_{1}\) satisfying (3.16) can be written uniquely as
Denote \( m_1 =k-r\) and \(n_1 =p-t\), and partition
Then by applying (3.4), we obtain the following identity,
Since \(S_1 {\widehat{S}}_1^{- 1} \) is nonsingular,
and the matrix \(X_{211}\) satisfying (3.19) can be written uniquely as
Furthermore, we denote \({{\mathcal {X}}_1 } = S_1{\widehat{S}}_1^{- 1}Y_1 \) and \( {{\mathcal {X}}_2 } = Y_2 \). Since \(S_1 {\widehat{S}}_1^{- 1} \) is nonsingular and \(\mathrm{rk}(Y)=k_1\), then \( \text{ rk }\left( {{{\begin{matrix} {{\mathcal {X}}_1 } \\ {{\mathcal {X}}_2 } \\ \end{matrix}} }} \right) = k_1\). By applying (3.18) and (3.19), we obtain the following identity,
Then applying Lemma 2.1 to the above, it produces \(0 \le k_1 \le \min \left\{ c{ + }(m_1- s ), n_1\right\} \), that is, \(t\le b\le \min \left\{ c+t+k-r-s, p\right\} \). Combining it with (3.11) leads to (3.5). Combining (3.13–3.21), we gain a general form for X which satisfies (1.3) is
where \({\mathcal {Z}} \in {{\mathbb {C}}}^{\left( {n - k}\right) \times p}\) and \({\mathcal {Y}} \in { {\mathbb {C}}}^{(k-r- s )\times t}\) are arbitrary matrices, and \(Y_1 \in { {\mathbb {C}}}^{ s \times n_1}\) and \(Y_2 \in { {\mathbb {C}}}^{(m_1- s )\times n_1}\) satisfy
Applying Lemma 2.4 and (3.15–3.20) to (3.22) get (3.6–3.9). \(\square \)
By applying generalized inverses, rank formulas and the above lemmas to simplify Theorem 3.1, we obtain the following theorem.
Theorem 3.2
Suppose that \(A \in {{\mathbb {C}}}^{m\times n}\), \(A_1 \in {{\mathbb {C}}}^{w \times n}\), \(C\in {{\mathbb {C}}}^{m\times p}\), \(C_1 \in {{\mathbb {C}}}^{w \times p}\), k, r, s, and the decompositions of A and \(A_1\) are given in Lemma 2.5. Denote
and the SVD of \({\widehat{{\mathcal {C}}}}\) as
where \(\Lambda = \text{ diag }\left\{ {\lambda _1 ,\ldots ,\lambda _c} \right\} \), \(\lambda _1 \ge \cdots \ge \lambda _c > 0\), \({\mathcal {U}}_1\) and \({\mathcal {V}}_1\) are unitary matrices of appropriate sizes. Then there exists a matrix \(X\in {\mathbb {C}}^{n\times p}\) satisfying (1.3) if and only if
If \(d+r-k \le b \le \min \left\{ \mathrm{rk}\left( A_1,\ \ C_1\right) , d-s, p\right\} \), then
and a general form for X which satisfies (1.3) is
where \({\widehat{{\mathcal {Z}}}}\in {\mathbb {C}}^{n\times p}\) and \({\widehat{{\mathcal {Y}}}}\in {\mathbb {C}}^{w\times p}\) are arbitrary matrix, and \({\widehat{{\mathcal {X}}}_2}\in {\mathbb {C}}^{w\times p}\) satisfies
If \(t \le b < d+r-k \), then
and a general form for X which satisfies (1.3) is
where \({\widehat{{\mathcal {Z}}}}\in {\mathbb {C}}^{n\times p}\) and \({\widehat{{\mathcal {Y}}}}\in {\mathbb {C}}^{w\times p}\) are arbitrary matrices, and \({\widehat{{\mathcal {X}}}_1}\in {\mathbb {C}}^{w\times p}\) and \({\widehat{{\mathcal {X}}}_2}\in {\mathbb {C}}^{w\times p}\) satisfy
and
when \(\lambda _{b-t} > \lambda _{b-t+1} \),
when \(q_2< b-t < q_1 \le r\) and \(\lambda _{q_2 }>\lambda _{q_2 + 1} = \cdots = \lambda _{q_1 } > \lambda _{q_1 + 1} \),
in which \({\mathcal {Q}}\) is an arbitrary matrix satisfying \({\mathcal {Q}} \in {\mathbb {C}}^{\left( {q_1 -q_2 } \right) \times \left( {b-t - q_2 }\right) }\) and \({\mathcal {Q}}^H{\mathcal {Q}} = I_{b-t - q_2 } \).
Proof
From (2.3) and \(A_1A_1^{-}=A_1 A_1^{\dag }\), it is easy to find that
and
It follows that \( \mathrm{rk}\left( C_{11}\right) = \mathrm{rk}\left( \left( I_w-A_1A_1^\dag \right) C_1\right) = \mathrm{rk}\left( {{\begin{matrix} A_1, &{} C_1 \\ \end{matrix} }} \right) - \mathrm{rk}\left( A_1\right) =t\).
From (2.3), \(A^{-}=W^{-1}\Sigma ^{\dag } U^{H}\) and \(A_1^{-}=W^{-1}\Sigma _1^{\dag } V^{H}\), we obtain
This gives
Applying (3.31), (1.1a) and (1.1b) to (3.24), we obtain
Furthermore, applying (2.3) and (3.1–3.4) to (3.23), we obtain
Thus, \(\mathrm{rk}({\widehat{{\mathcal { C}}}})=\mathrm{rk}( {\mathcal { C}})=c=d -t -k +r\). Hence (3.5\('\)) follows from (3.5).
From (2.3) and (3.1), we obtain
Hence (3.6\('\)) follows from (3.6) and (3.33).
Since (3.32), \({\mathcal {C}}\) and \({\widehat{{\mathcal {C}}}}\) have the same singular values. Hence (3.8\('\)) follows from (3.8) and (3.33).
Using (2.3), (3.1) (3.2) and (3.29), we obtian
From (2.3),(3.30) and (3.34),it is easy to find that
and
where \({\widehat{{\mathcal {Z}}}}\in {\mathbb {C}}^{n\times p}\), \({\widehat{{\mathcal {Y}}}}\in {\mathbb {C}}^{w\times p}\), \( {\mathcal {Y}} \in {\mathbb {C}}^{(k-r- s )\times t}\) and \({\mathcal {Z}} \in {\mathbb {C}}^{(n-k)\times p}\) are arbitrary matrices. Furthermore, using (3.30), (3.32) and (3.34), we obtain
Since \({\widehat{{\mathcal {X}}}_2}\in {\mathbb {C}}^{w\times p}\) satisfies (3.26), we obtain
where \( {{\mathcal {X}}_2 }\in {\mathbb {C}}^{(k-r-s)\times (p-t)}\) satisfies
Hence (3.7\('\)) follows from (3.22) and (3.35–3.37).
Since \({\mathcal {C}}\) and \({\widehat{{\mathcal {C}}}}\) have the same singular values, by applying Lemma 2.2, (2.1), (2.1\('\)), (3.28), (3.30) and (3.34), we obtain
where \( {\mathcal {X}}_1\in {\mathbb {C}}^{s\times (p-t)}\) satisfies \(\left\| {{\mathcal {C}} - {\mathcal {X}}_1 }\right\| =\min \) subject to \( { \text{ rk }\left( {\begin{matrix} {{\mathcal {X}}_1 } \\ {{\mathcal {X}}_2 } \\ \end{matrix}} \right) = k_1}\). Hence (3.9\('\)) follows from (3.22), (3.35) and (3.38). \(\square \)
We provide an example to illustrate that Theorem 3.2 is feasible.
Example 3.1
Take
Then \(r=1\), \(s=2\), \(k=4 \), \(m=n=w=p=5\), \(t=2\), \(\mathrm{rk}\left( A_1, \ C _1\right) =5\),
and
Compute the SVD of \({\mathcal {C}}\) by Matlab7 on a personal computer
Thus, by Theorem 3.1, there exists a rank-constrained least squares solution X to Problem 1.1 if and only if \(2 \le b \le 5\). When \(b = 4\),
and a general form for X satisfying (3.39) is given as follows
where \(x_i\), \(y_j\) and \(z_l\) are arbitrary, \(i=1,2\), \(j=1,2\) and \(l=1,\ldots ,5\).
When \(b = 2\),
and a general form for X satisfying (3.40) is given as follows
where \(y_j\) and \(z_l\) are arbitrary, \(j=1,2\) and \(l=1,\ldots ,5\).
Remark 3.1
By applying SVD and Q-SVD, we get two general forms of the least squares solutions of \(AX=C\) subject to \( \text{ rk }\left( {C_1 - A_1 X} \right) = b\). One thing worthy of note is that it seems hard to obtain one general form of the least squares solutions of \(AXB=C\) subject to \( \text{ rk }\left( {C_1 - A_1 XB_1} \right) = b \).
Investigate its reason, it is the matrix decomposition that is key tool to prove processing of Theorems 3.1 and 3.2. Thus we will focus on introducing a corresponding matrix decomposition in further study.
References
Ben-Israel, A., Greville, T.N.E.: Generalized Inverses: Theory and Applications, 2nd edn. Springer, Berlin (2003)
Chu, D., Hung, Y.S., Woerdeman, H.J.: Inertia and rank characterizations of some matrix expressions. SIAM J. Matrix Anal. Appl. 31, 1187–1226 (2009)
Demmel, J.W.: The smallest perturbation of a submatrix which lowers the rank and constrained total least squares problems. SIAM J. Numer. Anal. 24, 199–206 (1987)
Duan, G.-R.: Analysis and Design of Descriptor Linear Systems. Springer, New York (2010)
Eckart, C., Young, G.: The approximation of one matrix by another of lower rank. Psychometrika 1, 211–218 (1936)
Friedland, S., Torokhti, A.: Generalized rank-constrained matrix approximations. SIAM J. Matrix Anal. Appl. 29, 656–659 (2007)
Golub, G.H., Van Loan, C.F.: Matrix Computations, 2nd edn. The Johns Hopkins University Press, Baltimore (1989)
Golub, G.H., Hoffman, A., Stewart, G.W.: A generalization of the Eckart–Young–Mirsky matrix approximation theorem. Linear Algebra Appl. 88–89, 317–327 (1987)
Hnětynková, I., Plešinger, M., Sima, D.M.: Solvability of the core problem with multiple right-hand sides in the TLS sense. SIAM J. Matrix Anal. Appl. 37, 861–876 (2016)
Liu, X., Li, W., Wang, H.: Rank constrained matrix best approximation problem with respect to (skew) Hermitian matrices. J. Comput. Appl. Math. 319, 77–86 (2017)
Marsaglia, G., Styan, G.P.H.: Equalities and inequalities for ranks of matrices. Linear Multilinear Algebra 2, 269–292 (1974)
Mirsky, L.: Symmetric gauge functions and unitarily invariant norms. Q. J. Math. 11, 50–59 (1960)
Paige, C.C., Saunders, M.A.: Towards a generalized singular value decomposition. SIAM J. Numer. Anal. 18, 398–405 (1981)
Soto-Quiros, P., Torokhti, A.: Improvement in accuracy for dimensionality reduction and reconstruction of noisy signals. Part II: the case of signal samples. Signal Process. 154, 272–279 (2019)
Sou, K.C., Rantzer, A.: On the generalized matrix approximation problems in the spectral norm. Linear Algebra Appl. 436, 2331–2341 (2012)
Shen, D., Wei, M., Liu, Y.: Minimum rank (skew) Hermitian solutions to the matrix approximation problem in the spectral norm. J. Comput. Appl. Math. 288, 351–365 (2015)
Torokhti, A., Soto-Quiros, P.: Improvement in accuracy for dimensionality reduction and reconstruction of noisy signals. Part I: the case of random signals. Signal Process. 154, 338–349 (2019)
Tian, Y., Wang, H.: Relations between least-squares and least-rank solutions of the matrix equation \(AXB = C\). Appl. Math. Comput. 219, 10293–10301 (2013)
Wang, H.: On least squares solutions subject to a rank restriction. Linear Multilinear Algebra 63, 264–273 (2015)
Wang, Q.W., He, Z.H.: Solvability conditions and general solution for mixed Sylvester equations. Automatica 49, 2713–2719 (2013)
Wei, M., Wang, Q.: On rank-constrained Hermitian nonnegative-definite least squares solutions to the matrix equation \(AXA^H =B\). Int. J. Comput. Math. 84, 945–952 (2007)
Wei, M., Shen, D.: Minimum rank solutions to the matrix approximation problems in the spectral norm. SIAM J. Matrix Anal. Appl. 33, 940–957 (2012)
Acknowledgements
The authors wish to extend their sincere gratitude to Professor Eugene E. Tyrtyshnikov and the referees for their precious comments and suggestions.
Funding
This work is supported by Guangxi Natural Science Foundation (Grant Number 2018GXNSFAA138181), the Special Fund for Scientific and Technological Bases and Talents of Guangxi (Grant Number 2016AD05050) and the Special Fund for Bagui Scholars of Guangxi. Funding was provided by National Natural Science Foundation of China (Grant No. 11401243).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
No potential conflict of interest was reported by the author.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wang, H. Least squares solutions to the rank-constrained matrix approximation problem in the Frobenius norm. Calcolo 56, 47 (2019). https://doi.org/10.1007/s10092-019-0339-y
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10092-019-0339-y