Abstract
The low-rank quaternion matrix approximation has been successfully applied in many applications involving signal processing and color image processing. However, the cost of quaternion models for generating low-rank quaternion matrix approximation is sometimes considerable due to the computation of the quaternion singular value decomposition (QSVD), which limits their application to real large-scale data. To address this deficiency, an efficient quaternion matrix CUR (QMCUR) method for low-rank approximation is suggested, which provides significant acceleration in color image processing. We first explore the QMCUR approximation method, which uses actual columns and rows of the given quaternion matrix, instead of the costly QSVD. Additionally, two different sampling strategies are used to sample the above-selected columns and rows. Then, the perturbation analysis is performed on the QMCUR approximation of noisy versions of low-rank quaternion matrices. And we also employ the proposed QMCUR method to color image recovery problem. Extensive experiments on both synthetic and real data further reveal the superiority of the proposed algorithm compared with other algorithms for getting low-rank approximation, in terms of both efficiency and accuracy.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Quaternion [1] as a mathematical concept was originally introduced by Hamilton in 1843. As an extension of complex numbers, a quaternion number consists of one real part and three imaginary parts. A quaternion matrix is a generalization of a complex matrix in quaternion algebra. By now, quaternions and quaternion matrices have widely used in signal processing [2, 3], machine learning [4, 5], color image processing [6, 7], and other fields [8, 9]. By encoding the red, green, and blue channel pixel values of a color image on the three imaginary parts of quaternion matrix, this method perfectly fits the color image structure and effectively preserves the inter-relationship between the color channels [10].
As an emerging mathematical tool, low-rank quaternion matrix approximation (LRQA) has attracted much attention in the field of color image processing, such as color face recognition [11], color image inpainting [12,13,14,15], and color image denoising [16, 17]. For instance, Chen et al. [18] proposed the quaternion nuclear norm-based LRQA for color image denoising and inpainting. This method fully utilizes the high correlation among RGB channels by extending low-rank matrix approximation into the quaternion domain. Yu et al. [19] further extended the weighted nuclear norm minimization (WNNM) into the quaternion domain and proposed the quaternion-based WNNM method for color image restoration. Notably, the work [20] proposed the robust quaternion matrix completion to minimize the nuclear norm and \(l_1\)-norm. This approach provides an exact recovery guarantee under certain conditions and shows superior performance for color image recovery. Moreover, Jia et al. [21] introduced the patch group-based nonlocal self-similarity (NSS) prior scheme to learn explicit NSS models, and subsequently offered the NSS-based quaternion matrix completion (QMC) algorithm to reconstruct color images, leading to a promising result. Furthermore, they extended this approach to tensor NSS-based QMC for color videos inpainting. Additionally, in [22], the authors employed quaternion logarithmic norm to achieve a more accurate low-rank approximation. However, a key limitation of these methods for generating low-rank quaternion matrix approximation is that they need to compute the quaternion singular value decomposition (QSVD) in each iteration and suffer from computational deficiency, especially for large-scale data.
In addition to rank minimization, recent studies have utilized quaternion matrix decomposition and randomized techniques to improve the performance of LRQA [13, 23]. For instance, Miao et al. [24] suggested the matrix factorization for the target quaternion matrix, followed by three quaternion-based bilinear factor matrix norm factorization methods for low-rank quaternion matrix completion, which can avoid expensive calculations. Liu et al. [25] presented a randomized QSVD algorithm for low-rank matrix approximation. The randomized QSVD algorithm reduces the computational cost compared to traditional QSVD for large-scale data. Is there any method with lower computational cost for low-rank approximation of a quaternion matrix?
In recent years, the matrix CUR (MCUR) method [26,27,28] for fast low-rank approximation of real matrices has been actively investigated because of its ability to efficiently handle large-scale problems. The MCUR method approximates a low-rank matrix by directly sampling a subset of columns and rows from the original matrix and representing it as a product of three small-scale matrices. Owing to the random column/row selection strategy, the MCUR decomposition shows great potential for reducing the computational costs and preserving the properties of the original data matrix compared with other methods [29, 30].
To further enhance the approximation performance and boost the computational efficiency of LRQA, we consider the efficient quaternion matrix CUR (QMCUR) method for low-rank approximation. Precisely, the QMCUR approximation of a quaternion matrix is obtained by utilizing the actual rows and columns of the original quaternion matrix. Additionally, we employ two different sampling strategies to sample the above-selected columns and rows. In summary, the introduction of the QMCUR approximation and the utilization of sampling strategies offer promising avenues for achieving low-rank approximation. The main contributions of this paper are twofold:
-
We consider the QMCUR method for low-rank approximation of a quaternion matrix. This approach helps to reduce computational costs and improve the precision of low-rank approximation.
-
The perturbation error bound of the proposed approximation method is studied and demonstrated under the spectral norm. We also employ the QMCUR method to color image recovery problem and develop an algorithm to solve it.
-
Experimental results for synthetic data and color images demonstrate that the proposed QMCUR approximation method achieves a substantial speed-up while maintaining accuracy compared with other methods.
The rest of the paper is organized as follows. The main notation and preliminary for quaternions are presented in Section 2. Section 3 gives the details of the proposed quaternion CUR approximation method and perturbation error bound, and presents the QMCUR method for color image recovery. We provide numerical experiments in Section 4. Finally, some concluding remarks will be given in Section 5.
2 Preliminaries
2.1 Notation
Throughout this paper, \(\mathbb {R}\), \(\mathbb {C}\), and \(\mathbb {Q}\) respectively denote the set of real numbers, the set of complex numbers, and the set of quaternions. The set of all positive integers is denoted by \(\mathbb {N}\), and the symbol [n] represents the set of integers \(\{1, \dots , n\}\) for any \(n\in \mathbb {N}\). A scalar, a vector, and a matrix are written as x, \(\textbf{x}\), and \(\textbf{X}\), respectively. A dot (above the variable) is used to denote a quaternion variable \(\mathbb {Q}\) (e.g., \(\dot{x}\), \(\dot{\textbf{x}}\), and \(\dot{\textbf{X}}\) respectively represent a quaternion scalar, a quaternion vector, and a quaternion matrix). We use \(\dot{\textbf{X}}(I,:)\) and \(\dot{\textbf{X}}(:,J)\) to denote the row and column submatrices with indices I and J, respectively. Here, \((\cdot )^*, (\cdot )^\top , (\cdot )^H,\) and \((\cdot )^{\dagger }\) represent the conjugate, transpose, conjugate transpose, and Moore-Penrose inverse, respectively.
2.2 Quaternion and quaternion matrices
The set of quaternions \(\mathbb {Q}\) is a linear space over \(\mathbb {R}\), with an ordered basis of 1, \(\varvec{i}\), \(\varvec{j}\), and \(\varvec{k}\). Here, \(\varvec{i}\), \(\varvec{j}\), and \(\varvec{k}\) are three imaginary units with the multiplication laws: \(\varvec{i}^2 = \varvec{j}^2 = \varvec{k}^2 = \varvec{ijk} = -1\), \(\varvec{ij} = -\varvec{ji} = \varvec{k}\), \(\varvec{jk} = -\varvec{kj} = \varvec{i}\), \(\varvec{ki} = -\varvec{ik} = \varvec{j}\). An quaternion number \(\dot{x}\in \mathbb {Q}\) is of the form \(\dot{x} = {x}_0 + x_1\varvec{i} + x_2\varvec{j} + x_3\varvec{k}\) \(\in \) \(\mathbb {Q}\) with \(x_0, x_1, x_2, x_3 \in \mathbb {R}\). In particular, \(\dot{x}\) is called a pure quaternion if the real component \(x_0\) equals 0. The conjugate and modulus of \(x\) are defined as \(\dot{x}^* = x_0 - x_1\varvec{i} - x_2\varvec{j} -x_3\varvec{k}\) and \(|\dot{x}| = \sqrt{x_0^2 + x_1^2 + x_2^2 + x_3^2}\), respectively.
Analogously, a quaternion matrix \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\) can be written as \(\dot{\textbf{X}}= \textbf{X}_0 + \textbf{X}_1\varvec{i} + \textbf{X}_2\varvec{j} + \textbf{X}_3\varvec{k}\) with \(\textbf{X}_0, \textbf{X}_1, \textbf{X}_2, \textbf{X}_3\in \mathbb {R}^{m\times n}\). For given \(\dot{\textbf{X}}=(\dot{x}_{st})\in \mathbb {Q}^{m\times n}\), the conjugate of \(\dot{\textbf{X}}\) is denoted by \(\dot{\textbf{X}}^*= (\dot{x}^*_{st})\in \mathbb {Q}^{m\times n}\), the transpose of \(\dot{\textbf{X}}\) is denoted by \(\dot{\textbf{X}}^\top = (\dot{x}_{ts})\in \mathbb {Q}^{n\times m}\), and the conjugate transpose of \(\dot{\textbf{X}}\) is denoted by \(\dot{\textbf{X}}^H= (\dot{x}^*_{ts})\in \mathbb {Q}^{n\times m}\). Moreover, the Frobenius norm of \(\dot{\textbf{X}}\) is defined as \(\Vert \dot{\textbf{X}}\Vert _F = \sqrt{\text {Tr}(\dot{\textbf{X}}^H\dot{\textbf{X}})}= \sqrt{\sum _{s=1}^m\sum _{t=1}^n|\dot{x}_{st}|^2}\), where \(\text {Tr}(\cdot )\) is the trace operator, and the spectral norm \(\Vert \dot{\textbf{X}}\Vert _2\) is defined as \(\Vert \dot{\textbf{X}}\Vert _2=\max _{\varvec{\dot{a}}\in \mathbb {Q}^n,\Vert \varvec{\dot{a}}\Vert _2 = 1}\Vert \dot{\textbf{X}}\varvec{\dot{a}}\Vert _2\). The following gives some definitions and properties of quaternion matrices.
Definition 1
(The rank of quaternion matrix[31]) The maximum number of right (left) linearly independent columns (rows) of a quaternion matrix \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\) is called the rank of \(\dot{\textbf{X}}.\)
Definition 2
(Quaternion singular value decomposition (QSVD)[31]) Let \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\) be of rank k. There exist unitary quaternion matrices \(\dot{\textbf{W}}\in \mathbb {Q}^{m\times m}\) and \(\dot{\textbf{V}}\in \mathbb {Q}^{n\times n}\) such that \( \dot{\textbf{X}}=\dot{\textbf{W}}\mathbf {\Sigma }\dot{\textbf{V}}^{H},\) where \(\mathbf {\Sigma }=\left( \begin{aligned}&\textbf{D}_k&\textbf{0}\\&\textbf{0}&\textbf{0} \end{aligned}\right) \in \mathbb {R}^{m\times n}\), and \(\textbf{D}_k\) = diag(\(\sigma _1,\dots ,\sigma _k\)) \(\in \mathbb {R}^{k\times k}\) is a real diagonal matrix and has k positive entries \(\sigma _k\) (\(i = 1,\dots , k\)) on its diagonal (i.e., positive singular values of \(\dot{\textbf{X}}). \) The truncated QSVD of \(\textbf{A}\) with rank k is denoted by \(\textbf{A}_k=\dot{\textbf{W}}_k\mathbf {\Sigma }_k\dot{\textbf{V}}_k^{H}\), where \(\dot{\textbf{W}}_k= \dot{\textbf{W}} (:, 1:k)\in \mathbb {Q}^{m\times k}\), \(\dot{\textbf{V}}_k= \dot{\textbf{V}}(:, 1:k)\in \mathbb {Q}^{n\times k}\), and \(\mathbf {\Sigma }_k\) is a \(k \times k\) matrix containing the largest k singular values.
Definition 3
(Cayley-Dickson form[32]and Complex adjoint form [31]) Let \(\dot{\textbf{X}}= \textbf{X}_0 + \textbf{X}_1\varvec{i} + \textbf{X}_2\varvec{j} + \textbf{X}_3\varvec{k},\) \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\), \(\textbf{X}_0,\textbf{X}_1,\textbf{X}_2,\textbf{X}_3\in \mathbb {R}^{m\times n}.\) Then the Cayley-Dickson form of \(\dot{\textbf{X}}\) is expressed by \(\dot{\textbf{X}} = \textbf{X}_a+\textbf{X}_b\varvec{j}\in \mathbb {Q}^{m\times n},\) where \(\textbf{X}_a=\textbf{X}_0+\textbf{X}_1\varvec{i},\textbf{X}_b =\textbf{X}_2+\textbf{X}_3\varvec{j},\) and \(\textbf{X}_a,\textbf{X}_b\in \mathbb {C}^{m\times n}.\) The complex adjoint form of \(\textbf{X}\) is formulated as
The relation between the QSVD of quaternion matrix \(\dot{\textbf{X}}\) and the SVD of its equivalent complex matrix \(\textbf{P}_{\dot{\textbf{X}}}\in \mathbb {C}^{2m\times 2n}\) (\(\textbf{P}_{\dot{\textbf{X}}} = \textbf{W}\mathbf {\Sigma '}\textbf{V}^H\)) is defined as follows [3, 33, 34]
such that \(\dot{\textbf{X}} = \dot{\textbf{W}}\mathbf {\Sigma }\dot{\textbf{V}}^H,\) where
and \(\text {row}_{odd}(\textbf{M})\), \(\text {col}_{odd}(\textbf{M})\) extracts the odd rows and odd columns of matrix \(\textbf{M}\), respectively. Consequently, we obtain the QSVD of quaternion matrix \(\dot{\textbf{X}}\) as \(\dot{\textbf{X}} = \dot{\textbf{W}}\mathbf {\Sigma }\dot{\textbf{V}}^H.\)
Based on QSVD \(\dot{\textbf{X}} = \dot{\textbf{W}}\mathbf {\Sigma }\dot{\textbf{V}}^H\) we can further obtain the Moore-Penrose inverses [35, 36] of quaternion matrix as
More information about quaternion matrices can be found in [3, 31, 35].
3 Proposed approach
In this section, we first present the QMCUR method, which is designed to efficiently compute a low-rank approximation of a large-scale quaternion matrix with low computational costs and comparable accuracy. Then we provide the perturbation estimates for the QMCUR approximation of the noisy version of low-rank matrices. Furthermore, We employ the QMCUR method to color image recovery problem and develop an algorithm to solve it.
3.1 QMCUR-based low-rank approximation method
We now present the QMCUR approximation method. Let \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\) be a low-rank quaternion matrix with a predefined rank k. Consider row indices \(I\subseteq [m]\) and column indices \(J\subseteq [n]\) satisfying \(|I|, |J|\ge k\). Denote a column submatrix \(\dot{\textbf{C}} = \dot{\textbf{X}}(:, J)\) whose columns span the column space of \(\dot{\textbf{X}}\) and a row submatrix \(\dot{\textbf{R}} = \dot{\textbf{X}}(I,:)\) whose rows span the row space of \(\dot{\textbf{X}}\). Then the QMCUR approximation of \(\dot{\textbf{X}}\) is a product of the form
where \(\dot{\textbf{C}}\in \mathbb {Q}^{m\times |J|}\) with |J| columns from the quaternion matrix \(\dot{\textbf{X}}\) and \(\dot{\textbf{R}}\in \mathbb {Q}^{|I|\times n}\) with |I| rows from the quaternion matrix \(\dot{\textbf{X}}\). It’s obvious that the core quaternion matrix \(\dot{\textbf{U}}\in \mathbb {Q}^{|J|\times |I|}\) should be computed to yield the smallest error. The optimal choice for the core quaternion matrix \(\dot{\textbf{U}}\) in the least-squares sense is \(\dot{\textbf{U}}=\dot{\textbf{C}}^{\dagger }\dot{\textbf{X}}\dot{\textbf{R}}^{\dagger }\) [37] because
Note that one may replace \(\dot{\textbf{C}}^{\dagger }\) in (5) by \(\dot{\textbf{C}}^{\dagger }\dot{\textbf{X}}\dot{\textbf{R}}^{\dagger }\) to improve the approximation, resulting in the QMCUR approximation \(\dot{\textbf{C}}\dot{\textbf{C}}^{\dagger }\dot{\textbf{X}}\dot{\textbf{R}}^{\dagger }\dot{\textbf{R}}\). Algorithm 1 summarizes the overall process of the QMCUR approximation.
In Algorithm 1, the matrix \(\dot{\textbf{U}}\in \mathbb {Q}^{|J|\times |I|}\) is computed by \(\dot{\textbf{U}}=\dot{\textbf{C}}^{\dagger }\dot{\textbf{X}}\dot{\textbf{R}}^{\dagger }.\) Here, the computation of \(\dot{\textbf{C}}^{\dagger }\) relies on its QSVD (see (4) for more details) [35, 36]. If the QSVD of \(\dot{\textbf{C}}\in \mathbb {Q}^{m\times |J|}\) is \( \dot{\textbf{C}}=\dot{\textbf{W}}\mathbf {\Sigma }\dot{\textbf{V}}^{H},\) then the Moore-Penrose inverse of \(\dot{\textbf{C}}\) is given by
And the calculation of \(\dot{\textbf{R}}^{\dagger }\in \mathbb {Q}^{n\times |I|}\) is similar to that of \(\dot{\textbf{C}}^{\dagger }\). Additionally, we set the size of \(\dot{\textbf{C}}\) to be \(m \times k\log k\) and \(\dot{\textbf{R}}\) to be \(k\log k \times n\) for the QMCUR approximation method. The indices used to determine \(\dot{\textbf{C}}\) and \(\dot{\textbf{R}}\) are sampled using two different strategies [38, 39]. In the first strategy, the sampling probabilities \(p_j^{\text {col}}\) and \(q_i^{\text {row}}\) for each column j and row i of the quaternion matrix \(\dot{\textbf{X}}\) are based on the Euclidean norm of the columns and rows, which are respectively defined as
This strategy is referred to as QMCUR_length. In the second strategy, the sampling probabilities \( p_j^{\text {unif}} \) and \( q_i^{\text {unif}} \) are respectively defined as
This approach is referred to as QMCUR_uniform.
3.2 Computational complexity
Now, we discuss the computational complexity of the proposed algorithm. Let \(\dot{\textbf{X}}\in \mathbb {H}^{m\times n}\) be a given quaternion matrix. According to (2), the computation of the QSVD for a \(m\times n\) quaternion matrix is equivalent to the computation of the SVD of a \(2m\times 2n\) complex matrix (its complex adjoint). Thus, the computational complexity of QSVD for a \(m\times n\) quaternion matrix is \(\mathcal {O}(\text {min}(8mn^2, 8nm^2))\). Then, computing \(\dot{\textbf{C}}^{\dagger }\) requires performing QSVD on matrices with size of \(m\times |J|\) and multiplication, which cost \(\mathcal {O}((9|J|^2m+|J|m^2).\) Similarly, computing \(\dot{\textbf{R}}^{\dagger }\) requires performing QSVD on matrices with size of \(|I|\times n\) and multiplication, which cost \(\mathcal {O}(9|I|^2n+|I|n^2).\) Computing \(\dot{\textbf{U}}\) and \(\dot{\textbf{X}}\) only involves multiplication costing \(\mathcal {O}((|I|+|J|)mn+|I||J|(m+n))\). In summary, the computational cost of the developed algorithm is \(\mathcal {O}((8|J|^2m+8|I|^2n+|J|m^2+|I|n^2+(|J|+|I|)mn).\)
3.3 Perturbation estimates for CUR approximation
In practical applications, quaternion matrix data is often perturbed, such as noise in pictures, which may cause huge errors. In this section, we present the perturbation analysis suggested by the QMCUR approximation described above. Our main task will be to consider matrices in the form of
where \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\) with rank k, and \(\dot{\textbf{E}}\in \mathbb {Q}^{m\times n}\) is an arbitrary noise quaternion matrix drawn from a certain distribution.
Before analyzing the perturbation, we will first introduce some notations. If \({\tilde{\dot{\textbf{X}}}} = \dot{\textbf{X}} + \dot{\textbf{E}}\), and we consider \({\tilde{\dot{\textbf{C}}}} = {\tilde{\dot{\textbf{X}}}}(:,J)\in \mathbb {Q}^{m\times |J|}\), \({\tilde{\dot{\textbf{R}}}} = {\tilde{\dot{\textbf{X}}}}(I,:)\in \mathbb {Q}^{|I|\times n}\), and \({\tilde{\dot{\textbf{U}}}} = {\tilde{\dot{\textbf{C}}}} ^{\dagger }{\tilde{\dot{\textbf{X}}}}{\tilde{\dot{\textbf{R}}}} ^{\dagger }\in \mathbb {Q}^{|J|\times |I|}\) for selected index sets I and J. Then we write
where \({\dot{\textbf{C}}} ={\dot{\textbf{X}}}(:,J)\in \mathbb {Q}^{m\times |J|}\), \({\dot{\textbf{R}}}= {\dot{\textbf{X}}}(I,:)\in \mathbb {Q}^{|I|\times n}\), and \({\dot{\textbf{U}}} ={\dot{\textbf{C}}}^{\dagger }{\dot{\textbf{X}}}{\dot{\textbf{R}}}^{\dagger }\in \mathbb {Q}^{|J|\times |I|}.\) For ease of notation, we will use the conventions that \({\dot{\textbf{E}}_J}=\dot{\textbf{E}}(:, J)\in \mathbb {Q}^{m\times |J|}\) and \({\dot{\textbf{E}}_I}= {\dot{\textbf{E}}}(I,:)\in \mathbb {Q}^{|I|\times n}\). The following theorem provides a perturbation estimate for the QMCUR approximation.
Theorem 1
Let \({\tilde{\dot{\textbf{X}}}} = \dot{\textbf{X}} + \dot{\textbf{E}}\) for a fixed but arbitrary \(\dot{\textbf{E}}\in \mathbb {Q}^{m\times n}\). Using the notation in (11), then the following holds:
Furthermore, suppose that \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\) has rank k and its truncated QSVD is \(\dot{\textbf{X}}=\dot{\textbf{W}}_k \mathbf {\Sigma }_k\dot{\textbf{V}}_k^{H}\), where \(\dot{\textbf{W}}_k\in \mathbb {Q}^{m\times k}\), \(\mathbf {\Sigma }_k\in \mathbb {R}^{k\times k}\), and \(\dot{\textbf{V}}_k\in \mathbb {Q}^{n\times k}.\) Then we have
where \(\dot{\textbf{W}}_{k,I} = \dot{\textbf{W}}_{k}(I,:)\in \mathbb {Q}^{|I|\times k}\) and \(\dot{\textbf{V}}_{k,J} = \dot{\textbf{V}}_{k}(J,:)\in \mathbb {Q}^{|J|\times k}\) for index sets I, J.
In a word, Theorem 1 shows that the error estimation in the QMCUR method is controlled by the pseudoinverses of the submatrices of the orthogonal singular vectors and is linear in the norm of the noise \(\dot{\textbf{E}}\). To establish Theorem 1, we will introduce the following lemmas.
Lemma 1
Suppose that \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\) has rank k with the QMCUR approximation: \(\dot{\textbf{X}}\approx \textbf{CUR}\), where \(\dot{\textbf{C}} = \dot{\textbf{X}}(:,J)\) with selected column indices J, \(\dot{\textbf{R}} = \dot{\textbf{X}}(I,:)\) with selected row indices I, and \(\dot{\textbf{U}} = \dot{\textbf{C}}^\dagger \dot{\textbf{X}}\dot{\textbf{R}}^\dagger \). Then we have
Proof
Recall that \(\dot{\textbf{C}} = \dot{\textbf{X}}(:,J)\). Thus, span(\(\dot{\textbf{C}}\)) \(\subset \) span(\(\dot{\textbf{X}}\)). On account of the fact that \(\dot{\textbf{X}}=\dot{\textbf{C}}\dot{\textbf{C}}^{\dagger }\dot{\textbf{X}}\dot{\textbf{R}}^{\dagger }\dot{\textbf{R}}\), we have span(\(\dot{\textbf{X}}\)) \(\subset \) span(\(\dot{\textbf{C}}\)). A similar argument shows that span(\(\dot{\textbf{R}}^H\)) \(\subset \) span(\(\dot{\textbf{X}}^H\)). The conclusion follows. \(\square \)
Lemma 2
The following hold:
Proof
Note that
The final inequality holds because the first norm term is 0 by identity of the Moore-Penrose pseudoinverse and \(\Vert \textbf{I}-{\tilde{\dot{\textbf{C}}}}{\tilde{\dot{\textbf{C}}}} ^\dagger \Vert _2\le 1\) as this is an orthogonal projection operator. Since \(\text {rank}(\dot{\textbf{C}}) = \text {rank}(\dot{\textbf{X}})\), we have \(\dot{\textbf{X}} = \dot{\textbf{C}}\dot{\textbf{C}}^\dagger \dot{\textbf{X}}\). Then
The second inequality is derived by mimicking the above argument. The conclusion follows. \(\square \)
Lemma 3
Suppose \(\dot{\textbf{X}}, \dot{\textbf{C}}, \dot{\textbf{U}}, \) and \(\dot{\textbf{R}}\) are as in (5) such that \( \dot{\textbf{X}}= \dot{\textbf{C}}\dot{\textbf{U}}\dot{\textbf{R}}\), and suppose that rank\((\dot{\textbf{X}})\) = k. Let \(\dot{\textbf{X}}={\dot{\textbf{W}}}_k \mathbf {\Sigma }_k{\dot{\textbf{V}}}_k^{H}\) be the truncated QSVD of \(\dot{\textbf{X}}\). Then the following hold:
where \(\dot{\textbf{W}}_{k,I} = \dot{\textbf{W}}_{k}(I,:)\in \mathbb {Q}^{|I|\times k}\) and \(\dot{\textbf{V}}_{k,J} = \dot{\textbf{V}}_{k}(J,:)\in \mathbb {Q}^{|J|\times k}\) for selected index sets I, J.
Proof
Begin with the fact that \(\dot{\textbf{R}}={\dot{\textbf{W}}}_k(I,:) \mathbf {\Sigma }_k{\dot{\textbf{V}}}_k^{H}={\dot{\textbf{W}}}_{k,I} \mathbf {\Sigma }_k{\dot{\textbf{V}}}_k^{H}. \) Then we have
Notice that \(\dot{\textbf{W}}_{k,I}\) has full column rank, \(\mathbf {\Sigma }_k\) is a \(k\times k\) matrix with full rank, and \(\dot{\textbf{V}}_k^{H}\) has orthogonal rows (i.e., \((\dot{\textbf{V}}_k^{H})^\dagger = \dot{\textbf{V}}_k\)). In this case,
Consequently, the following holds:
The second equality above follows from the unitary invariance of the norm [40]. Similarly, we have
whereupon the conclusion follows from the fact that \( \Vert (\dot{\textbf{V}}_{k,J}^H)^{\dagger } = (\dot{\textbf{V}}_{k,J}^\dagger )^H\) which has the same norm as \(\dot{\textbf{V}}_{k,J}^\dagger \). The conclusion follows. \(\square \)
Proof of Theorem 1
First note that
The second inequality follows from the fact that \(\Vert \dot{\textbf{C}}\dot{\textbf{C}}^{\dagger }\Vert _F = 1\). Next, using the formula \( {\tilde{\dot{\textbf{X}}}} = \dot{\textbf{X}} + \dot{\textbf{E}}\), we have \(\Vert {\tilde{\dot{\textbf{X}}}}(\textbf{I}-{\tilde{\dot{\textbf{R}}}} ^{\dagger }{\tilde{\dot{\textbf{R}}}} )\Vert \le \Vert \dot{\textbf{X}}(\textbf{I}-{\tilde{\dot{\textbf{R}}}} ^{\dagger }{\tilde{\dot{\textbf{R}}}} )\Vert +\Vert \dot{\textbf{E}}\Vert \) since \(\Vert \textbf{I}-{\tilde{\dot{\textbf{R}}}} ^\dagger {\tilde{\dot{\textbf{R}}}} \Vert _2\le 1\). Then applying Lemma 2, we obtain
Moreover, from Lemma 3, it follows that
3.4 QMCUR approximation for color image recovery
In this section, we apply the QMCUR approximation method to recover color images. The problem of recovering color images can be formulated as a quaternion matrix completion problem, aiming to find its missing entries through the following optimization. The mathematical model is
where rank(\(\cdot \)) denotes the rank function, \({\dot{\textbf{X}}}\in \mathbb {Q}^{m\times n}\) and \(\dot{\textbf{Y}}\in \mathbb {Q}^{m\times n}\) represent the recovered and observed quaternion matrices, respectively. \(\Omega \) represents the set of observed elements, and \(P_{\Omega }(\dot{\textbf{X}})\) is the projection operator that keeps entries in \(\Omega \) and zeros out others. However, directly solving the problem (26) is difficult as the rank minimization problem is known as NP-hard [41].
The quaternion nuclear norm minimization (QNNM) [18] is a popular convex surrogate of non-convex rank function, and its minimization method is as following
However, the computational of QNNM can be quite expensive for large-scale data, due to calculating the QSVD of quaternion matrices. Therefore, if the rank of the matrix \(\dot{\textbf{X}}\) is known [42,43,44], the QNNM method can be more effectively replaced by the concept of quaternion decomposition. In this case, the problem (26) is presented as follows
Intuitively, this problem seeks the matrix \(\dot{\textbf{X}}\) of rank k that best fits the given data \(\dot{\textbf{Y}}\).
Similar to the Algorithm framework in [43], the optimization problem (28) can be expressed as follows
where \(\dot{\textbf{M}}\in \mathbb {Q}^{m\times n}\) is the auxiliary tensor variable. Therefore, we can solve the optimization problem (28) alternatively over the variables \(\dot{\textbf{X}}\) and \(\dot{\textbf{M}}\). The solution to the minimization problem (28) can be approximated using the following iterative procedures:
where t represents the iteration number, \(\mathcal {L}\) is the an operator used to calculate a low-rank quaternion matrix approximation of the quaternion matrix \(\dot{\textbf{X}}\), which can be achieved by Algorithm 1, and \(\textbf{1}\) is a matrix in which all components are equal to one. The algorithm mainly consists of two steps: low-rank quaternion matrix approximation (30) and masking computation (31). We summarize the proposed method in Algorithm 2. It starts with the initial incomplete data \(\dot{\textbf{X}}^0\) and iteratively enhances the approximate solution until a stopping criterion is satisfied or the maximum number of iterations is reached.
4 Numerical experiments
In this section, we conduct numerical experiments to assess the accuracy and computation time of the proposed QMCUR method for low-rank quaternion matrix approximation.
Example 4.1
In this experiment, we demonstrate the error estimation in Theorem 1 using simulation data. We randomly generate a quaternion matrix \(\dot{\textbf{X}}\) as a product \(\dot{\textbf{X}} = {\dot{\textbf{W}}}_k\mathbf {\Sigma }_k{\dot{\textbf{V}}}_k^{H}\in \mathbb {Q}^{m\times m}\), where \(\dot{\textbf{W}}_k\in \mathbb {Q}^{m\times k}, \dot{\textbf{V}}_k\in \mathbb {Q}^{m\times k}\) are unitary matrices, and \(\mathbf {\Sigma }_k\) is a diagonal \(k \times k\) matrix with positive diagonal elements. In addition, the random noise quaternion matrix is given by \(\dot{\textbf{E}}=\textbf{E}_0+\textbf{E}_1\varvec{i}+\textbf{E}_2\varvec{j}+\textbf{E}_3\varvec{k}\in \mathbb {Q}^{m\times n}\), where the entries of \(\textbf{E}_t (t = 0, 1, 2, 3)\) are i.i.d. Gaussian variables with mean zero and variance \(\sigma \). Then we obtain the perturbed quaternion matrix
In this experiment, we assume that \( m = 500 \) and \( k = 50 \). By applying Algorithm 1 to compute the rank-k approximation of \({\tilde{\dot{\textbf{X}}}}\) with \(\sigma = 10^{-1}, 10^{-2}, \dots , 10^{-6}\), Fig. 1 shows the results. As observed, the error \(\Vert \dot{\textbf{X}}-{\tilde{\dot{\textbf{C}}}}{\tilde{\dot{\textbf{U}}}}{\tilde{\dot{\textbf{R}}}} \Vert _2\) increases linearly as \(\Vert \dot{\textbf{E}}\Vert _2\) increases, which confirms the conclusion of Theorem 1.
Example 4.2
In this experiment, we test the proposed QMCUR method on simulated quaternion data and compare their performance with that of QSVD [33], qSVD4 [45], lansvdQ [46], and CSP-QM [7]. We still consider the perturbed quaternion matrix \({\tilde{\dot{\textbf{X}}}}\) in Example 4.1. Figure 2 displays the relative error and running time computed by different methods with the varying size m when k =10. This error is measured by
where \(\dot{\textbf{X}}\) denotes the quaternion matrix that represents the clean data and \(\dot{\textbf{M}}\) is obtained by all methods.
As observed, QMCUR\(\_\)length and QMCUR\(\_\)uniform methods in Algorithm 1 are significantly faster than QSVD when the data size (i.e., m) is sufficiently large. In particular, the errors for QMCUR_uniform are smaller than those obtained by the QSVD algorithm in the noiseless case (i.e., \(\sigma \) = 0). This is mainly because only the QSVD of the smaller sizes \(\dot{\textbf{C}}\) and \(\dot{\textbf{R}}\) need to be computed, not the QSVD of the quaternion matrix \(\dot{\textbf{X}}\). The proposed methods require less computation time and give more accurate approximations than the qSVD4, lansvdQ, and CSP-QM methods. This observation verifies that the proposed QMCUR method enhance computational performance by calculating the QSVD of selected rows and columns. In summary, the proposed methods achieve a substantial speed-up while maintaining accuracy compared to other methods, especially when m is sufficiently large.
Example 4.3
In this example, we evaluate the performance of the proposed QMCUR method for color image compression. The dataset used in this example is the Set27Footnote 1 and two images are shown in Fig. 4. Each image, with a size of 683 \(\times \) 1024 \(\times \) 3, is represented by a pure quaternion matrix \(\dot{\textbf{X}}= \mathbf {{R}}\varvec{i} + \textbf{G}\varvec{j} + \textbf{B}\varvec{k}\in \mathbb {Q}^{683\times 1024}\), where \(\mathbf {{R}}\), \(\textbf{G}\), and \(\textbf{B}\) represent the RGB (red, green, blue) channels of color image. The elements of \(\dot{\textbf{X}}\) are \(\dot{x}_{st} = r_{st}\varvec{i} + g_{st}\varvec{j} + b_{st}\varvec{k}\), where \(r_{st}\), \(g_{st}\), and \(b_{st}\) denote the red, green, and blue pixel values at the position (s, t) of the color image, respectively. Notably, the number of rows and columns is sethui to klogk, where k represents the predefined rank of the quaternion matrix \(\dot{\textbf{X}}\in \mathbb {Q}^{m\times n}\).
Figure 3 illustrates the relative error and running time of all methods for various selected ranks. As observed, the performance of the proposed methods is superior to that of QSVD in terms of relative error when k is set to a suitable size. In terms of running time, the proposed methods are much faster than QSVD. Fig. 4 presents the reconstructed image quality for various target ranks, showing little difference in the visual effect between each original image and the approximate image obtained by the proposed methods and QSVD.
Example 4.4
In this part, we test the Algorithm 2 using four-color images of size 512\(\times \)768\(\times \)3 from the Kodak PhotoCD DatasetFootnote 2 under the random missing scenario. Several low-rank matrix completion methods are compared in this experiment, including LRQA-1 [18], QLNF [22], and Q-DFN [24]. The missing ratio (MR) is defined as the ratio of the number of missing elements to the total number of elements. To provide a numerical evaluation of the recovered results, the peak signal-to-noise rate (PSNR), structural similarity (SSIM), and running time per iteration (in seconds) are employed. The higher PSNR and SSIM value imply better recovered results.
In Table 1, we list the PSNR, SSIM, and time values of different methods on color images, and the best values are highlighted in bold. Compared with competing methods, the proposed methods (QMCUR_length and QMCUR_uniform) achieve the best results in terms of all quantitative metrics. Furthermore, the proposed QMCUR_uniform is the fastest, demonstrating its feasibility on large-scale color images. To clearly observe the recovery performance, Fig. 5 presents the recovered images of four color images by all methods with MR = 80\(\%\). As observed, the proposed methods generate better results than the compared methods. These observations demonstrate that our methods can be applicable for large-scale data processing.
5 Conclusion
This work presented an efficient quaternion matrix CUR method for computing low-rank approximation of a quaternion matrix. The QMCUR approximation offers a balance between accuracy and computational costs by selecting specific column and row submatrices of a given quaternion matrix. Besides, we conducted a perturbation analysis of the proposed approximation method and concluded that the error in the quaternion spectral norm is correlated with the noise quaternion matrix in the first order. Experimental results illustrated that the QMCUR approximation methods are significantly faster than comparative low-rank quaternion matrix approximation methods, without sacrificing the quality of reconstruction on both synthetic and color image datasets. In the future, we will extend the proposed methods to other high-dimensional data processing, e.g., quaternion tensor completion [47], tensor completion [48, 49] and multi-dimensional image recovery [50].
Availability of supporting data
The data sets generated during and/or analyzed during the current study are available from the corresponding author on reasonable request.
References
Hamilton, W.R.: Elements of quaternions. Green, & Company, Longmans (1866)
Ell, T.A., Le Bihan, N., Sangwine, S.J.: Quaternion fourier transforms for signal and image processing. John Wiley & Sons (2014)
Le Bihan, N., Mars, J.: Singular value decomposition of quaternion matrices: a new tool for vector-sensor signal processing. Signal Processing. 84(7), 1177–1199 (2004)
Hirose, A., Yoshida, S.: Generalization characteristics of complex-valued feedforward neural networks in relation to signal coherence. IEEE Transactions on Neural Networks and Learning Systems. 23(4), 541–551 (2012)
Zhou, H., Zhang, X., Zhang, C., Ma, Q.: Quaternion convolutional neural networks for hyperspectral image classification. Engineering Applications of Artificial Intelligence. 123, 106234 (2023)
Wang, G., Zhang, D., Vasiliev, V.I., Jiang, T.: A complex structure-preserving algorithm for the full rank decomposition of quaternion matrices and its applications. Numerical Algorithms. 91(4), 1461–1481 (2022)
Zhang, D., Jiang, T., Jiang, C., Wang, G.: A complex structure-preserving algorithm for computing the singular value decomposition of a quaternion matrix and its applications. Numerical Algorithms. 95(4), 267–283 (2024)
Guo, Z., Jiang, T., Vasil’ev, V., Wang, G.: Complex structure-preserving method for schrödinger equations in quaternionic quantum mechanics. Numerical Algorithms. 1–17 (2023)
Jiang, T., Guo, Z., Zhang, D., Vasil’ev, V.: A fast algorithm for the schrödinger equation in quaternionic quantum mechanics. Applied Mathematics Letters. 150, 108975 (2024)
Qi, L., Luo, Z., Wang, Q., Zhang, X.: Quaternion matrix optimization: Motivation and analysis. Journal of Optimization Theory and Applications. 193(1–3), 621–648 (2022)
Zou, C., Kou, K.I., Wang, Y.: Quaternion collaborative and sparse representation with application to color face recognition. IEEE Transactions on Image Processing. 25(7), 3287–3302 (2016)
Chen, J., Ng, M.K.: Color image inpainting via robust pure quaternion matrix completion: Error bound and weighted loss. SIAM Journal on Imaging Sciences. 15(3), 1469– 1498 (2022)
Li, C., Liu, Y., Wu, F., Che, M.: Randomized block krylov subspace algorithms for low-rank quaternion matrix approximations. Numerical Algorithms. 1–31 (2023)
Xu, T., Kong, X., Shen, Q., Chen, Y., Zhou, Y.: Deep and low-rank quaternion priors for color image processing. IEEE Transactions on Circuits and Systems for Video Technology. 33(7), 3119–3132 (2023)
Jia, Z., Ng, M.K., Wang, W.: Color image restoration by saturation-value total variation. SIAM Journal on Imaging Sciences. 12(2), 972–1000 (2019)
Gai, S., Yang, G., Wan, M., Wang, L.: Denoising color images by reduced quaternion matrix singular value decomposition. Multidimensional Systems and Signal Processing. 26, 307–320 (2015)
Huang, C., Li, Z., Liu, Y., Wu, T., Zeng, T.: Quaternion-based weighted nuclear norm minimization for color image restoration. Pattern Recognition. 128, 108665 (2022)
Chen, Y., Xiao, X., Zhou, Y.: Low-rank quaternion approximation for color image processing. IEEE Transactions on Image Processing. 29, 1426–1439 (2020)
Yu, Y., Zhang, Y., Yuan, S.: Quaternion-based weighted nuclear norm minimization for color image denoising. Neurocomputing. 332, 283–297 (2019)
Jia, Z., Ng, M.K., Song, G.: Robust quaternion matrix completion with applications to image inpainting. Numerical Linear Algebra with Applications. 26(4), 2245 (2019)
Jia, Z., Jin, Q., Ng, M.K., Zhao, X.: Non-local robust quaternion matrix completion for large-scale color image and video inpainting. IEEE Transactions on Image Processing. 31, 3868–3883 (2022)
Yang, L., Miao, J., Kou, K.I.: Quaternion-based color image completion via logarithmic approximation. Information Sciences. 588, 82–105 (2022)
Ren, H., Ma, R., Liu, Q., Bai, Z.: Randomized quaternion qlp decomposition for low-rank approximation. Journal of Scientific Computing. 92(3), 80 (2022)
Miao, J., Kou, K.I.: Quaternion-based bilinear factor matrix norm minimization for color image inpainting. IEEE Transactions on Signal Processing. 68, 5617–5631 (2020)
Liu, Q., Ling, S., Jia, Z.: Randomized quaternion singular value decomposition for low-rank matrix approximation. SIAM Journal on Scientific Computing. 44(2), 870–900 (2022)
Goreinov, S.A., Tyrtyshnikov, E.E., Zamarashkin, N.L.: A theory of pseudoskeleton approximations. Linear Algebra and its Applications. 261(1), 1–21 (1997)
Aldroubi, A., Hamm, K., Koku, A.B., Sekmen, A.: CUR decompositions, similarity matrices, and subspace clustering. Frontiers in Applied Mathematics and Statistics. 4 (2019)
Lin, P., Peng, S., Xiang, Y., Li, C., Cui, X., Zhang, W.: Structure-oriented CUR low-rank approximation for random noise attenuation of seismic data. IEEE Transactions on Geoscience and Remote Sensing. 61, 1–13 (2023)
Drineas, P., Mahoney, M.W., Muthukrishnan, S.: Relative-error CUR matrix decompositions. SIAM Journal on Matrix Analysis and Applications. 30(2), 844–881 (2008)
Wang, S., Zhang, Z.: Improving CUR matrix decomposition and the nyström approximation via adaptive sampling. The Journal of Machine Learning Research. 14(1), 2729–2769 (2013)
Zhang, F.: Quaternions and matrices of quaternions. Linear Algebra and its Applications. 251, 21–57 (1997)
Schafer, R.D.: On the algebras formed by the Cayley-Dickson process. American Journal of Mathematics. 76(2), 435–446 (1954)
Xu, Y., Yu, L., Xu, H., Zhang, H., Nguyen, T.: Vector sparse representation of color image using quaternion matrix analysis. IEEE Transactions on Image Processing. 24(4), 1315–1329 (2015)
Xiao, X., Chen, Y., Gong, Y., Zhou, Y.: 2D quaternion sparse discriminant analysis. IEEE Transactions on Image Processing. 29, 2271–2286 (2020)
Kyrchei, I.: Weighted singular value decomposition and determinantal representations of the quaternion weighted Moore-Penrose inverse. Applied Mathematics and Computation. 309, 1–16 (2017)
Kyrchei, I.: Determinantal representations of the Moore-Penrose inverse over the quaternion skew field and corresponding cramer’s rules. Linear and Multilinear Algebra. 59(4), 413–431 (2011)
Wang, Q.: The general solution to a system of real quaternion matrix equations. Computers & Mathematics with Applications. 49(5), 665–675 (2005)
Drineas, P., Kannan, R., Mahoney, M.W.: Fast monte carlo algorithms for matrices iii: Computing a compressed approximate matrix decomposition. SIAM Journal on Computing. 36(1), 184–206 (2006)
Chiu, J., Demanet, L.: Sublinear randomized algorithms for skeleton decompositions. SIAM Journal on Matrix Analysis and Applications. 34(3), 1361–1383 (2013)
Ling, C., He, H., Qi, L.: Singular values of dual quaternion matrices and their low-rank approximations. Numerical Functional Analysis and Optimization. 43(12), 1423–1458 (2022)
Gillis, N., Glineur, F.: Low-rank matrix approximation with weights or missing data is np-hard. SIAM Journal on Matrix Analysis and Applications. 32(4), 1149–1165 (2011)
Liu, J., Musialski, P., Wonka, P., Ye, J.: Tensor completion for estimating missing values in visual data. IEEE Transactions on Pattern Analysis and Machine Intelligence. 35(1), 208–220 (2013)
Ahmadi-Asl, S., Asante-Mensah, M.G., Cichocki, A., Phan, A.H., Oseledets, I., Wang, J.: Fast cross tensor approximation for image and video completion. Signal Processing. 213, 109121 (2023)
Vandereycken, B.: Low-rank matrix completion by riemannian optimization. SIAM Journal on Optimization. 23(2), 1214–1236 (2013)
Li, Y., Wei, M., Zhang, F., Zhao, J.: Real structure-preserving algorithms of householder based transformations for quaternion matrices. Journal of Computational and Applied Mathematics. 305, 82–91 (2016)
Jia, Z., Ng, M.K., Song, G.: Lanczos method for large-scale quaternion singular value decomposition. Numerical Algorithms. 82, 699–717 (2019)
Miao, J., Kou, K.I., Liu, W.: Low-rank quaternion tensor completion for recovering color videos and images. Pattern Recognition. 107, 107505 (2020)
Li, B., Zhao, X., Zhang, X., Ji, T., Chen, X., Ng, M.K.: A learnable group-tube transform induced tensor nuclear norm and its application for tensor completion. SIAM Journal on Imaging Sciences. 16(3), 1370–1397 (2023)
Li, B., Zhao, X., Ji, T., Zhang, X., Huang, T.: Nonlinear transform induced tensor nuclear norm for tensor completion. Journal of Scientific Computing. 92(3), 83 (2022)
Lin, J., Huang, T., Zhao, X., Ji, T., Zhao, Q.: Tensor robust kernel PCA for multidimensional data. IEEE Transactions on Neural Networks and Learning Systems. 1–13 (2024)
Acknowledgements
This research is supported by University of Macau (MYRG2022-00108-FST, MYRG-CRG2022-00010-ICMS), the Science and Technology Development Fund, Macau S.A.R (0036/2021/AGJ), the National Key Research Project (2022YFB3904104), and the Key Project from National Natural Science Foundation of China (42230406).
Funding
This research is supported by University of Macau (MYRG2022-00108-FST, MYRG-CRG2022-00010-ICMS), the Science and Technology Development Fund, Macau S.A.R (0036/2021/AGJ), the National Key Research Project (2022YFB3904104), and the Key Project from National Natural Science Foundation of China (42230406).
Author information
Authors and Affiliations
Contributions
Pengling Wu: Methodology, Software, Writing - Original Draft. Kit Ian Kou: Conceptualization, Validation, Resources. Hongmin Cai: Methodology, Validation, Writing - Review and Editing. Zhaoyuan Yu: Formal analysis, Writing - Review and Editing.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Ethical Approval
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wu, P., Kou, K.I., Cai, H. et al. Efficient quaternion CUR method for low-rank approximation to quaternion matrix. Numer Algor (2024). https://doi.org/10.1007/s11075-024-01923-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11075-024-01923-8