1 Introduction

Recently, solving the inverse problem of images has attracted extensive attention of scholars [4, 9]. In the context of imaging, several methods have been proposed to exploit the sparsity of image patches in a sparsifying transform domain or learned dictionary to reconstruct images, especially in the field of Compressed Sensing (CS) [11]. CS conducts sampling and compression at the same time by utilizing the redundancy of the signal. This technology mainly consists of three important stages, i.e., signal sparse representation, sensing matrix design and signal reconstruction. And finding the best sparsifying basis is the prerequisite for compressed sensing theory.

The conventional CS sparse representation methods often utilize fixed sparsifying transforms such as Fourier transform, discrete cosine transform, wavelet transform, etc. However, for lacking of translation and rotation invariance, the orthogonal basis is not enough to capture the various features of images. In recent years, with the in-depth study of dictionary learning algorithms [8, 20, 21], researchers have extended CS to the redundant dictionary, which is adaptively learned from the processed signal itself. Hence, the sparse decomposition of signals in redundant dictionaries has become a research hotspot in sparse coding field. Various studies have shown that the learned dictionaries can commendably match the structure of the signal or image itself, and exhibit great performance in image classification [3], image denoising [13, 17], image super-resolution [15], etc.

The K-singular value decomposition (K-SVD) [1] algorithm, an iterative method that alternates between sparse coding stage and dictionary atoms update stage, is one of the most well-known dictionary learning algorithms. It has been widely used in face recognition [14, 22, 25], image processing [2, 18] and so on. This method updates the dictionary columns jointly with an update of the sparse representation coefficients related to it. Thus it helps to avoid matrix inversion calculation and accelerate convergence. However, the K-SVD algorithm needs to calculate SVD excessively for the whole dictionary atoms during each iteration procedure. Since SVD operation has large computational cost as the dimensions of the input vector increases, the speed of the K-SVD algorithm may slow down. Furthermore, only the largest singular value and its corresponding singular vector are used while the others are abandoned directly, which results in the waste of computing resources.

Ref. [26] proposed the K-RBP dictionary learning algorithm, a novel dictionary update method, to sparsely represent video frame. This algorithm makes changes of K-SVD in the dictionary update stage and improves the manner of calculating the largest singular value and its corresponding singular vector. Based on Ref. [26], we exploit a novel dictionary learning algorithm, called adaptive K-bilateral random projections (AK-BRP), to relieve the above limitations of the K-SVD algorithm. In the sparse coding stage, the sparsity constraint is associated with the current dictionary to obtain an adaptive sparsity constraint; while in the dictionary update stage, the bilateral random projections (BRP) [6, 7], a method of low rank approximation, is used to directly compute the largest singular value and its corresponding singular vector. Besides, the adaptive sparsity is employed by associating the maximal number of dictionary atoms in the sparse coding stage with the iterated updated dictionaries. Using the dictionaries learned by our new method as sparse representation for video frame compressed sensing, experimental results on a wide range of video frames for CS recovery have shown that our proposed algorithm is better than K-SVD algorithm in terms of computation expense, running time and sparse representation.

The remainder of this paper is organized as follows. Section 2 briefly reviews CS theory and the K-SVD dictionary learning algorithm. Section 3 provides the proposed dictionary learning algorithm in detail. The compressed sensing algorithm based on our adaptive redundant dictionary learning for video frame recovery is detailed in Section 4. Experimental results are reported in Section 5. Finally, in Section 6, we conclude this paper.

2 Background

2.1 Compressed sensing

Compressed sensing theory indicates that if a signal x ∈ R n × 1 is sparse or compressible in an orthogonal basis or tight framework Ψ = (ψ 1, ⋯, ψ n )T, (here Ψ satisfies Ψ ⋅ Ψ T = I), we can use a random sensing matrix Φ ∈ R m × n(m < < n), which is irrelevant to the sparsifying transform base, to obtain the linear observational vector y ∈ R m × 1 of the sparse transform coefficient vector Θ = Ψ T x. Then the original signal x can be exactly reconstructed from the linear measurements y, whose number is much smaller than that of the original signal. The CS model can be formulated as

$$ y=\varPhi x=\varPhi \varPsi \varTheta $$
(1)

It has been well known that the sparsity degree of the signal, which is a key to achieve accurate reconstruction, plays an important role in the field of compressed sensing. The higher degree of a signal, the higher reconstruction precision it will have. Thus, finding a domain in which the signal has a high degree of sparsity is one of the main challenges CS recovery should face [24]. The conventional CS recovery algorithms mainly use a set of fixed sparsifying domains such as DCT, wavelet and FFT domain. These algorithms are not able to capture the various geometrical features contained in the high dimensional data and result in poor CS recovery performance for high dimensional signals. Recently, with the rapid development of signal sparse representation methods based on over-complete dictionaries, the study of CS sparse coding is mostly extended to redundant dictionaries. Nowadays, designing a more efficient redundant dictionary has become the emphasis.

Another fundamental principle that CS relies on is incoherent projection [5]. CS theory requires that Φ, the sensing matrix, and Ψ, the sparse basis, must satisfy Restricted Isometry Property (RIP) [4, 5]. Only in this way can we reconstruct the original signal accurately.

The purpose of CS reconstruction is to recover \( \overset{\frown }{\varTheta } \) from the linear measurements y, and then obtain the approximation of the original signal by \( \overset{\frown }{x}=\varPsi \overset{\frown }{\varTheta } \). It can be described as the following optimization problem

$$ \underset{\varTheta }{ \min }{\left\Vert y-\varPhi \varPsi \varTheta \right\Vert}_2^2+\lambda {\left\Vert \varTheta \right\Vert}_0 $$
(2)

where λ is a non-negative parameter, ‖v0 is the l 0 norm, counting the number of nonzero elements of the vector v, and ‖v2 is the l 2 norm which is defined as \( {\left\Vert v\right\Vert}_2=\sqrt{{\displaystyle \sum {v}_i^2}} \).

Though Eq. (2) is a NP-hard problem, greedy algorithms, such as orthogonal matching pursuit (OMP), can obtain approximate solution with relatively small running time. When l 0 norm is replaced with l 1 norm, which adds all the absolute values of the entries in the vector, the minimization problem can be solved by the convex relaxation methods such as BP, least absolute shrinkage method and LASSO.

2.2 Dictionary learning

In this sub-section, we will describe the K-SVD algorithm, which is one of the most commonly used methods of dictionary learning, in detail. Given a training set X ∈ R n × N whose columns are the form of {x i } N i = 1 , and A = {a i } N i = 1  ∈ R K × N is the corresponding coefficient. Then, the dictionary learning process is to find a possible optimal dictionary D ∈ R n × K(K > n), which contains K prototype signal atoms for columns {d j } K j = 1 , for sparse representation of the given training samples X It can be expressed as

$$ \underset{D,A}{ \min }{\left\Vert X-DA\right\Vert}_F^2\kern0.5em s.t.\kern0.5em \forall i,{\left\Vert {a}_i\right\Vert}_0\le {T}_0 $$
(3)

where ‖‖ F denotes the Frobenius norm of a matrix, ‖‖0 is the l 0 norm, and T 0 is the maximum number of non-zero atoms used in the sparse representation vector.

The specific implementation steps of the K-SVD algorithm is outlined in Algorithm 1 [1]:

figure e

3 The proposed algorithm

From the process of the K-SVD algorithm, it is not difficult to find that the singular value decomposition (SVD), which is used for low rank approximation, plays a significant role in K-SVD. However, the use of SVD restrains K-SVD in terms of execution time. Especially when the dimension of the input vectors increases, the computation time will rapidly increase, which worsen the K-SVD algorithm. Consequently, we would like to propose a novel fast dictionary learning algorithm named adaptive K-BRP algorithm. Since dictionary learning algorithm is usually treated as iterations of sparse coding stage and dictionary update stage, we will describe our new dictionary learning algorithm in these two stages.

3.1 Sparse coding stage

Unlike the sparse coding stage of K-SVD algorithm, in our method, we associate the sparsity upper-bound with the iterated updated dictionary by using the coherence of the current dictionary, to obtain an adaptive sparsity constraint, sequentially reducing the reconstruction errors iteratively.

Suppose that T j is the sparsity upper-bound for each iteration process, T j denotes

$$ {T}_j=\frac{1}{2}\left(1+\frac{1}{\mu \left({D}_j\right)}\right)\kern0.6em j=1,2,\cdots, P $$
(5)

where μ(D) is the mutual coherence of the dictionary D, which describes the maximum similarity between the over-complete dictionary atoms, expressed as [10]

$$ \mu (D)= \max \left|\left\langle {d}_i\cdot {d}_j\right\rangle \right|,\kern0.7em i\ne j $$
(6)

Here, the μ(D) function is valued between 0 and 1. The minimum is reached for an orthogonal dictionary and the maximum for a dictionary containing at least two collinear atoms [16].

In order to expound the rationality of our adaptive sparsity constraint T j , the following theorem is introduced [10, 12]:

Theorem 3.1

Let D ∈ R n × K(K > n) be the dictionary and μ(D) denote the mutual coherence of the dictionary D. Suppose that x = Da, where a is sparse, if the sparsity S, the number of nonzero entries of the correct coefficients, satisfies Eq. (7)

$$ S<\frac{1}{2}\left(1+\frac{1}{\mu (D)}\right) $$
(7)

then the following conclusions can be deduced:

  1. 1)

    The solution a of the minimization problem (4) must be the sparsest and unique;

  2. 2)

    The l 0 norm minimization problem in Eq. (3) can be equivalent to the l 1 norm;

  3. 3)

    Any pursuit algorithm (such as OMP) can find out the best linear combination of S atoms from the dictionary D

Hence, the defined T j can ensure precise reconstruction of the sparse signal, and is reasonably practicable.

Replacing T 0 in Eq. (4) with T j , then the sparse coding stage is to solve the following optimization problem

$$ \underset{a_i}{ \min }{\left\Vert {x}_i-{D}_j{a}_i\right\Vert}_2^2\kern0.8em s.t.\kern0.6em {\left\Vert {a}_i\right\Vert}_0\le {T}_j\kern0.6em \forall \kern0.4em i=1,2,\cdots, N;\kern0.7em j=1,2,\cdots, P $$
(8)

This problem can be solved by greedy algorithms, such as OMP, ROMP, etc.

3.2 Dictionary update stage

As is known to all, K-SVD algorithm employs singular value decomposition (SVD), a low rank approximation method, to obtain the update of dictionary atoms, since SVD can achieve the minimum reconstruction error. However, the use of SVD restricts K-SVD algorithm in terms of computation time, especially for high dimensional data. To this end, we would like to adopt another method of low rank approximation, named bilateral random projections (BRP) [6, 7], to reduce the computation time.

Given \( {E}_k^R\in {R}^{n\times \left|{\omega}_k\right|} \), the restricted representation error matrix that corresponds to examples that use the atom d k in the update stage of the K-SVD algorithm, then the r rank bilateral random projections (BRP) of E R k is described as

$$ \left\{\begin{array}{l}{Y}_1={E}_k^R{A}_1\\ {}{Y}_2={\left({E}_k^R\right)}^T{A}_2\end{array}\right. $$
(9)

wherein \( {A}_1\in {R}^{\left|{\omega}_k\right|\times r} \) and A 2 ∈ R n × r are random matrices.

Then the fast rank-r approximation of E R k can be expressed as

$$ {\widehat{E}}_k^R={Y}_1{\left({A}_2^T{Y}_1\right)}^{-1}{Y}_2^T $$
(10)

However, it has been pointed out in reference [6] that when singular values of E R k decay slowly, Eq. (10) may perform poorly. At this point, we need to apply power scheme [19] to Ê R k for improving the approximation precision. In the power scheme modification, we instead calculate the BRP of a matrix R k  = (E R k (E R k )T)q E R k , whose singular values decay faster than E R k . Particularly, λ i ( R k ) = λ i (E R k )2q + 1, here λ i (⋅) denotes the ith singular value of the matrix. Both E R k and R k share the same singular vectors. Thereafter, we will be able to obtain the BRP of R k according to that of E R k as

$$ \left\{\begin{array}{l}{Y}_1={\tilde{E}}_k^R{A}_1\\ {}{Y}_2={\left({\tilde{E}}_k^R\right)}^T{A}_2\end{array}\right. $$
(11)

According to Eq. (10), the BRP based r rank approximation of R k is

$$ {\widehat{\tilde{E}}}_k^R={Y}_1{\left({A}_2^T{Y}_1\right)}^{-1}{Y}_2^T $$
(12)

In order to obtain the fast r rank approximation of E R k , we need to calculate the QR decomposition of Y 1 and Y 2 respectively

$$ \left\{\begin{array}{l}{Y}_1={Q}_1{R}_1\\ {}{Y}_2={Q}_2{R}_2\end{array}\right. $$
(13)

The fast low-rank approximation of E R k with rank r is then given by

$$ {\widehat{E}}_r={\left({\tilde{E}}_k^R\right)}^{\frac{1}{2q+1}}={Q}_1{\left[{R}_1{\left({A}_2^T{Y}_1\right)}^{-1}{R}_2^T\right]}^{\frac{1}{2q+1}}{Q}_2^T $$
(14)

Because in each dictionary atom update process, only the largest singular value and its corresponding singular vector are used, we just need to calculate the fast 1-rank approximation of E R k . Then, the random matrices A 1 and A 2 can be transformed into random vectors, which will greatly accelerate the speed of the algorithm.

4 Compressed sensing algorithm based on the AK-BRP dictionary learning

In this section, we will elaborate the scheme of video frame compressed sensing based on the proposed dictionary learning algorithm in detail.

Assuming that a video sequence is composed of I frames, of size W × L, then the video sequence can be expressed as

$$ squ={f}_r\left(x,y\right) $$
(15)

where 1 ≤ x ≤ L, 1 ≤ y ≤ W, 1 ≤ r ≤ I,

Dividing each frame of the video sequence into small non-overlapping blocks of equal size (i.e., size \( \sqrt{n}\times \sqrt{n} \)), here we suppose that W and L are both the integral multiple of \( \sqrt{n} \), then each frame can be divided into WL/n blocks. Among them, the corresponding vector x t r of the tth image patch f t r (x, y) in the rth frame can then be formed by the way of row stacking or column stacking, expressed as

$$ {x}_r^t={\left({f}_r^t\left(1,1\right),{f}_r^t\left(1,2\right)\cdots {f}_r^t\left(1,\sqrt{n}\right),{f}_r^t\left(2,1\right),\cdots {f}_r^t\left(2,\sqrt{n}\right)\cdots {f}_r^t\left(\sqrt{n},1\right)\cdots {f}_r^t\left(\sqrt{n},\sqrt{n}\right)\right)}^T $$
(16)

where the vector x t r  ∈ R n × 1, 1 ≤ t ≤ WL/n, 1 ≤ r ≤ I,

For each image patch, the corresponding measurements vector of CS is y t r  = Φx t r  + ε t r , where ε t r is the random noise in measurement process (here in this paper, we assume the same sampling matrix is applied to each block). In the compressed sensing process, we exploit the proposed AK-BRP dictionary learning algorithm to train the redundant dictionary D = D P , and then represent each patch x t r by a linear combination of Da t r (i.e. x t r  = Da t r ), wherein D ∈ R n × K (each atom of D has unit norm), a t r  ∈ R K × 1 is the sparse representation coefficient. Because the sparsest representation of x t r corresponds to the vector of weight coefficient a t r with the smallest number of non-zeros, we use greedy iterative algorithms, such as OMP, to find the coefficient vector a t r , and then reconstruct the tth block by x t r  = Da t r . And lastly, the reconstructed rth frame of the video sequence can be obtained by reshaped every tth block in the rth frame. The whole procedure is as follows

figure f

5 Experiment simulations and analysis

In this section, we carry out several experimental simulations on two standard CIF (frame size: 352 × 288, luminance only) video sequences: Foreman and Akiyo, to evaluate the performance of our proposed dictionary learning algorithm for CS sparse representation.

In our experiments, each test frame is divided into 8 × 8 non-overlapping blocks and the CS measurements are obtained by applying a Gaussian random projection matrix to each block (here, each block is sampled by the same sensing matrix). The other parameter setting of proposed algorithm is as follows: the over-complete dictionary size is chosen to be 64 × 256 and the number of iterations for training is set to 30. All implementations involved in the experiments are coded in Matlab R2011a. Computations are performed on a ThinkPad computer with Intel (R) Core (TM) i5-5250U CPU @ 1.6 GHz, 8GB memory, and Windows 8 operating system.

5.1 Visual effect of dictionary learning

In order to validate that the improved dictionary learning algorithm has better performance in sparse representation, the K-SVD dictionary learning algorithm and the proposed AK-BRP dictionary learning algorithm are respectively used to represent the 23th frame of the Foreman video sequence by learning dictionaries from the processed frame itself. In the experiment, the sparsity of K-SVD is set to be 5, where sparsity is defined as the maximum number of non-zero coefficients used to represent each block in the sparse coding stage. The redundant dictionaries trained by these two algorithms are shown in Fig. 1.

Fig. 1
figure 1

Visual comparison of two types of learned dictionaries. From left to right: the redundant dictionary trained by K-SVD algorithm; the redundant dictionary trained by AK-BRP algorithm

5.2 Improvement on accuracy in video frame compressed sensing

In this sub-section, we will show the influence of the proposed dictionary learning algorithm on the performance of CS recovery. In our simulation, the K-SVD redundant dictionary and the AK-BRP redundant dictionary are respectively selected as the sparsifying basis of CS, and the OMP algorithm is used as CS reconstruction algorithm to restore every test frame. For conveniently describing, the two methods are credited as KSVD-CS and AK-BRP-CS, respectively.

To evaluate the reconstruction quality of each algorithm, PSNR (Peak Signal to Noise Ratio, unit: dB) is used to evaluate the objective image quality. Besides, a recently proposed powerful perceptual quality metric FSIM (Feature SIMilarity) [23] is also calculated to evaluate the visual quality. It has been recognized that the higher FSIM value means the better visual quality. Comparisons of execution time (unit: s), PSNR and FSIM for five gray test frames of Foreman in the case of 0.4 to 0.6 sampling rate are provided in Tables 1, 2 and 3, respectively. Here, in order to eliminate the randomness, all the values of each test frame are averaged over 10 executions.

Table 1 Time Comparison with Various CS Methods Based on Different Sparsifying Basis
Table 2 PSNR Comparison with Various CS Methods Based on Different Sparsifying Basis
Table 3 FSIM Comparison with Various CS Methods Based on Different Sparsifying Basis

Tables 1, 2 and 3 show that compared with KSVD-CS algorithm, our proposed algorithm is not only time saving, but also achieves the highest PSNR and FSIM, which can effectively reduce the reconstruction error. In order to compare the merits and demerits of each algorithm more intuitively, we carry out experiments at different sampling rates ranging from 0.3 to 0.9, and the comparison results are presented in Fig. 2. Wherein, the values of running time, PSNR and FSIM are averaged over 5 or 10 random gray test frames. And two types of videos (Foreman and Akiyo) are used.

Fig. 2
figure 2

Comparison of reconfiguration performance of KSVD-CS and AK-BRP-CS algorithms. a Comparison of CS reconstruction running time; b comparison of CS reconstruction PSNR; c comparison of CS reconstruction FSIM

From Fig. 2a, we can discover that the running time of our AK-BRP-CS algorithm is much lower than that of the KSVD-CS algorithm for both Foreman and Akiyo videos. The reason is that our method can obtain adaptive sparsity in the sparse coding stage, and only calculates the largest singular value and the corresponding singular vector during each dictionary atom update stage, which accelerates the speed of the algorithm. Moreover, Fig. 2b shows that the reconstruction performances of the two algorithms are not very good for low sampling rate. However, with the increase of sampling rate, the two algorithms’ reconstruction results both have a significant facelift. It is clear that our method improves the CS reconstruction PSNR remarkably compared to KSVD-CS. Figure 2c once again demonstrates that the proposed algorithm has better visual quality than the KSVD-CS algorithm.

Figure 2a-c also show that there is negligible difference between performances of 5 and 10 test frames. As a result, it is reasonable to consider 5 random test frames and the performance is acceptable. In addition, since for both Foreman and Akiyo video sequences, AK-BRP-CS algorithm is superior to KSVD-CS, only results on “Foreman” will be shown in the following experiments.

Specifically, some visual results of the recovered 6th frame of “Foreman” by the two algorithms are presented in Fig. 3. Here, the sampling rate is equal to 0.4.

Fig. 3
figure 3

Visual quality comparison of CS recovery on the 6th frame with subrate = 0.4. From left to right: the KSVD-CS recovered frame; the AK-BRP-CS recovered frame

Obviously, the proposed algorithm shows much clearer and better visual results than the KSVD-CS algorithm in the same experimental conditions. And the results fully reflect the superiority of our algorithm. However, because of the randomicity and adaptivity of our algorithm, there is some poor local performance of AK-BRP-CS, e.g. the top-left corner of the AK-BRP-CS recovered frame.

5.3 Effect of sparse coding stage

The superior performance of the proposed algorithm has been adequately demonstrated in the previous section. In this sub-section, we will verify the influences of our adaptive sparsity constraint on the algorithm performance. To make a better comparison, we change the sparse coding stage of the proposed AK-BRP dictionary learning algorithm by using the sparse coding method of the conventional KSVD algorithm, forming the K-BRP dictionary learning algorithm. In our comparison experiments, the K-BRP dictionary and the AK-BRP dictionary are respectively selected as the sparsifying basis of CS, and the OMP reconstruction algorithm is used to restore each test frame. For ease of the elaboration, the CS based on these two dictionaries are credited as K-BRP-CS and AK-BRP-CS, respectively. Figure 4 shows the comparison of reconfiguration performance of K-BRP-CS, whose sparsity is respectively set to 3, 5 and 8, and AK-BRP-CS algorithms. Wherein, the values of running time, PSNR and FSIM are averaged over the five gray test frames of “Foreman”, whose values are all averaged over 10 executions.

Fig. 4
figure 4

Comparison of reconfiguration performance of K-BRP-CS and AK-BRP-CS algorithms in the case of different sparsity. a Comparison of CS reconstruction running time; b comparison of CS reconstruction PSNR; c comparison of CS reconstruction FSIM

As can be seen from Fig. 4a, with the increase of the sparsity, the running time of the K-BRP-CS algorithm increases progressively. Especially, when the sparsity is 3, the running time of K-BRP-CS algorithm is close to that of AK-BRP-CS algorithm, which means that the adaptive selection of sparsity in each iteration of our proposed algorithm is less than 3. In addition, Fig. 4b, c both illustrate that the change of sparsity has little effect on the reconstruction performance of K-BRP-CS algorithm, since the CS reconstruction PSNR and FSIM are almost unanimous in the four cases. To sum up, our AK-BRP-CS algorithm simultaneously ensures the reconstruction accuracy and selects the smaller sparsity in every iteration step to decrease the running time, which fully reflects the feasibility of the designed adaptive sparse coding mode.

5.4 Effect of dictionary update stage

In this sub-section, the superiority of the proposed dictionary learning algorithm in the dictionary update stage will be further certificated by the simulation results. In order to make a better comparison, we improve the dictionary update mode of the traditional K-SVD algorithm by using the proposed adaptive sparse coding method, forming an adaptive K-SVD (AK-SVD) dictionary learning algorithm. Similar to Sub-section 5.3, the CS based on AK-SVD dictionary also be credited as AK-SVD-CS for conveniently describing. The comparisons of reconstruction property of AK-SVD-CS and AK-BRP-CS algorithms are presented in Fig. 5. Wherein, the values of running time, PSNR and FSIM are averaged over the five gray test frames of “Foreman”, whose values are all averaged over 10 executions.

Fig. 5
figure 5

Comparison of reconfiguration performance of AK-SVD-CS and AK-BRP-CS algorithms. a Comparison of CS reconstruction running time; b comparison of CS reconstruction PSNR; c comparison of CS reconstruction FSIM

Obviously, compared with the AK-SVD-CS algorithm, our AK-BRP-CS algorithm saves as much as half the running time, and improves the CS reconstruction PSNR and FSIM about 3–5 dB, 0.07-0.15 respectively, when the sampling rate among 0.3 to 0.9. The reason is that the AK-BRP-CS algorithm only calculates the largest singular value and the corresponding singular vector in the dictionary update stage by exploiting the bilateral random projections (BRP) method, and results in a great reduction in the computation cost. In addition, due to the stronger robustness of the BRP method in comparison with SVD decomposition, our algorithm can improve the reconstruction accuracy to some extent. Above all, the way of updating the dictionary in this paper is more practical.

6 Conclusion

We present a new method, named adaptive K-BRP (AK-BRP), for learning patch-sparse adaptive redundant dictionaries to improve the accuracy of CS reconstruction. The proposed dictionary learning algorithm associates the sparsity upper-bound with the iterated updated dictionary in the sparse coding stage and adopts bilateral random projections to update each dictionary atom in the following dictionary update stage. A great deal of comparative experiments show that our proposed algorithm achieves significant sparse representation performance improvements over the traditional K-SVD dictionary learning algorithm.