1 Introduction

With the development of imaging technology in space satellite optical cameras based on the multispectral time-delay-integration-charge-coupled-device (TDICCD), the requirements of the performance specifications, such as field of view (FOV), spatial resolution, time resolution, spectral resolution, radiometric resolution and wide swath, have been constantly improved in recent years [1,2,3,4]. At the same time, the number of TDICCDs which are used for splicing into long linear arrays is growing. The read speed rate of TDICCD and the quantization bits of an analog-to-digital convertor in a TDICCD video processor have also been improving. These factors have led to a dramatic increase in the total amount of digital image data of multispectral TDICCD cameras [5,6,7]. However, the storage capacities of satellite solid-state recording devices are limited and satellite downlink channel bandwidth is also restricted. The present storage devices and downlink technology have difficultly adapting the huge amount of data relating to multispectral TDICCD images. Therefore, the multispectral image data must be compressed when TDICCD cameras work in orbit.

Presently, the prevailing compression algorithms generally use a three dimensional (3D) transform technology to compress multispectral images [8]. The 3D transform-based compression approach of multispectral images usually has an image transform stage, such as discrete cosine transform (DCT) [9], discrete wavelet transform (DWT) [10] and Karhunen–Loeve transform (KLT) [11]. The transform stage can remove spatial and spectral redundancies. Then, the transformed coefficients are encoded by entropy encoding algorithms, such as Embedded Zero-tree Wavelet [12], Embedded Block Coding with Optimization Truncation (EBCOT) [13], Set Partitioning in Hierarchical Trees (SPIHT) [14], Set Partitioned Embedded Block Coder (SPECK) [15], and Bit Plane Encoder (BPE) [16, 17]. The typical transform-based algorithms are JPEG2000, which are suitable for still images [18]. In the JPEG2000 algorithm, EBCOT technology is a very efficient approach to remove the redundancy between wavelet coefficients. The EBCOT technology enables JPEG2000 to have an excellent compression performance. However, it is too complex to be implemented in a space satellite TDICCD camera. Another typical transform-based algorithm is the Consultative Committee for Space Data Systems (CCSDS)—Image Data Compression (CCSDS-IDC) algorithm [19,20,21,22,23]. The CCSDS-IDC algorithm is mainly used in space application. The CCSDS-IDC algorithm contains two stages: (1) DWT; (2) BPE. The BPE is a zero tree encoder. It makes full use of the structures of spatiotemporal orientation trees in a bit plane. In other words, grandchildren coefficients also become not important when children coefficients are not important. The zero-tree characteristic causes the bit plane to have a large number of zero areas. Taking full advantage of these zero areas can improve coding efficiency. The CCSDS-IDC algorithm has the characteristics of progressive coding and fault-tolerant capability. In addition, BPE is low-complexity and occupies less storage capacity, which is suitable for the application of on-board cameras. However, the CCSDS-IDC method decreases the average peak signal to noise ratio (PSNR) by 2 dB compared with JPEG2000. In addition, CCSDS-IDC is only suitable for two dimensional (2D) images, which cannot exploit the spectral redundancy between 3D images. Realizing the needs of remote sensing imaging sensors capturing multiple spectral bands, the CCSDS standard committee [21] published the CCSDS 123.0-B-1: Lossless multispectral and hyperspectral image compression recommended standard [22] and CCSDS 122.1-B-1: spectral pre-processing transform for multispectral and hyperspectral image compression [23]. These two multispectral and hyper-spectral image compression standards could provide the references of onboard lossless, near-lossless, and lossy compression. The compression standard of multispectral and hyperspectral images may be extended by the CCSDS committee in future more effective compression.

Many compression methods of multispectral images have been proposed [24] for space optical imaging systems. In 2004, Enrico et al. proposed a new context-based adaptive lossless image coding method to implement the lossless and near-lossless compression for hyperspectral images [25]. In 2007, Enrico et al. proposed a distributed source coding technique for the lossless compression of hyperspectral images [26]. Didier et al. used Graphics Processing Units to implement a lossless hyperspectral data compression system for space applications [27]. These lossless compression algorithms of multispectral and hyperspectral images can provide the references for the lossy compression of multispectral and hyperspectral images from optical remote sensors. In 2013, Enrico et al. analysed the predictor of the predictive lossy compression paradigm [28]. This work outlines the directions for improvement of the algorithms, especially in the treatment of noisy channels, and investigates the use of appropriate statistical models for the entropy-coding stage. Kiely et al. proposed a 3D progressive wavelet-based compressor for hyperspectral images [29]. They used a 3D wavelet decomposition structure to exploit correlations in all three dimensions of hyperspectral data sets, while facilitating elimination of spectral ringing artefacts. Klimesh et al. analysed the spectral ringing artefacts in hyperspectral image data compression to improve the reconstructed band quality [30]. These multispectral and hyperspectral image compression methods can efficiently complete the compression mission for space optical imaging sensors with multiple spectral bands. They also can provide the reference for the future space compression standard of multispectral and hyperspectral image compression.

The compression methods outlined above are mainly used in multispectral and hyperspectral images that have multiple bands. However, for now, compression algorithms of multispectral images with relatively few spectral bands usually ignore the spectral redundancy. Each band is encoded independently, which causes the problem of low compression performance for remote sensing multispectral images with relatively few spectral bands. Achieving such a compressed multispectral image with a few bands remains a crucial but unsolved issue. In order to address this problem, two methods are proposed in our previous work. In [31], a one-dimensional (1D) DCT was used to exploit the spectral information, and a Tucker decomposition (TD) in spatial transform domain was used to remove residual spectral redundancies between bands and residual spatial redundancies intra bands. A deep coupling approach was also used to manage the TD processing. This method focused on removing spatial redundancies. In [32], a low-complexity removing spectral redundancy algorithm was investigated, while a traditional method using DWT and BPE was used to remove the spatial redundancies.

To address the low compression performance of remote sensing multispectral images which have relatively a few bands, this paper proposes an effective compression method for these images based on a fast KLT in conjunction with a distributed post-transform in the wavelet domain. This is able to achieve good performance with moderate complexity, so as to adapt to the application requirements of new generation of high-resolution multispectral TDICCD cameras with wide FOV. The proposed method can also be used in space smart cameras [33] integrated with navigation devices [34,35,36], in space compressed imaging cameras for micro-satellites [37], and for space visual-tracking systems [38, 39].

The rest of this paper is organized as follows. Section 2 presents the architecture of the method. In Sect. 3, the experimental results are demonstrated. Section 4 concludes the paper.

2 Proposed compression method

2.1 Overall architecture

Multispectral images are divided into two categories: multispectral image having a few bands (MIHFB) and multispectral image having normal bands (MIHNB). The MIHFB usually has a high spatial resolution and a relatively low spectral resolution. In other words, the MIHFB has a high spatial correlation (i.e. spatial redundancy) between two adjacent pixels in a spectral band, while the spectral correlation between two adjacent bands is relatively low. On the contrary, the MIHNB usually has a high spectral resolution and a relatively low spatial resolution. Thus, the MIHNB has the high spectral correlation between two adjacent bands, while the spatial correlation between two adjacent pixels in a spectral band is relatively low. Based on the two different features of MIHFB and MIHNB, MIHFB usually requires a compression algorithm with the high-performance in the spatial decorrelation, while the spectral decorrelation is relatively low. MIHNB usually requires a compression algorithm with the high-performance in the spatial decorrelation, while the spectral decorrelation is relatively low. In this paper, we define images having a band number of below 10 as MIHFB, and those having a band number of above 10 as MIHNB. MIHFBs are obtained by a multispectral TDICCD, which is composed of several TDICCD arrays in parallel and produces several bands simultaneously (Fig. 1). The optics reflected and radiated by a ground object converges on the thin optical film of the TDICCD surface through the optical system. Each band TDICCD array captures optical energy to obtain the corresponding spectral band image. Each band image contains 1D spatial information of ground objects. At this point, a 1D spectral and 1D spatial image is obtained by this multispectral TDICCD camera. When the camera moves along push-broom direction, additional 1D spatial information of ground objects is obtained. Through iterations of this process, the multispectral TDICCD camera produces 3D images. The number of bands of multispectral image can be customized by the user. The imaging system in Fig. 1 uses four bands, a number widely used in multispectral imaging having a few bands.

Fig. 1
figure 1

Imaging principle of multispectral TDICCD with four bands, where Lf is the focal length of the camera, LH is the orbital height

In our previous work, compression approaches of panchromatic and multispectral images are thoroughly researched [16, 17, 19]. In this paper, a new compression method of multispectral images for remote sensing camera having comparatively few spectral bands is proposed. Four bands are used as an example to illustrate the proposed method. To balance computational complexity and compression performance of the MIHFB, an efficient scheme based on low-complexity fast KLT (FKLT) combined with improved post transform (PT) and Distributed Source Coding (DSC) integrated with CCSDS-IDC (DSC-CCSDS) is proposed for multispectral images having a few bands. This method integrates FKLT, improved post-transform in the wavelet domain, optimized CCSDS-IDC and distributed source coding to remove the spatial redundancy, spectral redundancy and bit information redundancy. The proposed compression encoder has than lower complexity than traditional ones, making it suitable for on-board camera applications.

Figure 2 shows the proposed compressor architecture of the MIHFB. The number of bands is four. First, all bands of multispectral images are transformed using a FKLT to remove the spectral redundancy. Then, a 3-level 9/7 lifting 2D DWT is attached to each principal component band to remove the intra-band spatial redundancy. Here, the lifting 2D DWT is a float form. In addition, an improved post-transform having only one fast principal component analysis (PCA) basis is attached to wavelet sub-bands of each band to exploit the remaining redundancy between adjacent wavelet coefficients intra-sub-band. Finally, an optimized DSC-CCSDS codec is used as an entropy coder to remove information redundancy and residual redundancy between the adjacent bands.

Fig. 2
figure 2

The proposed coding architecture for MIHFBs

2.2 Fast KLT

KLT can be considered as an excellent method to remove the spectral redundancies between multispectral images [40,41,42,43]. However, the computation complexity of KLT is too high to be implemented in remote sensing applications. To lower the complexity, the following calculating strategy is adopted:

  1. 1.

    Spatial tiling

    The computation complexity of KLT is proportional to the size of each band [44, 45]. However, space camera has large wide-swath and high time resolution. The size of each band is very large. To accelerate computation speed, a tiling strategy is used in which a 3D multispectral image is spatially tiled. Each spatially tiled sub-block is considered as a new 3D multispectral image. The new 3D multispectral image is attached by KLT

  1. 2.

    Spectral band tiling

    In [45], the computational complexity of KLT is proportional to the number of bands in the multispectral images. A spectral band tiling strategy is used based on the idea of the clustering successive approximation method in [46,47,48], and [66]. Each spatially tiled sub-block, which has L spectral bands, is spectrally tiled to form several groups. Each two spectral bands are considered as a spectral tiled group. A KLT is attached to each spectral tiled group. Each spatially tiled sub-block is transformed by n level KLT. In first level transform, the L spectral bands can be spectrally tiled into L/2 groups. Each spectral tiled group includes two spectral bands and is transformed by KLT to form the first principal component spectral band and the second principal component spectral band. For each spatially tiled sub-block, the L/2 first principal component spectral bands and the L/2 s principal component spectral bands can be obtain in the first level transform. Since most of image energy is focused on the first principal component spectral bands, the redundancy between spectral tiled bands has been removed. Then, the L/2 first principal component spectral bands in the first level can be transformed by KLT at the second level. The transform approach in the second level is the same as the first. The transform approach at each level is the same until the n level transform is completed. Finally, the L principal component spectral bands can be obtained. When the number of spectral bands in one level is an odd number, the unpaired component exists. In this case, the unpaired component is directly forwarded to the next level. In other words, the unpaired components are seen as a principal component at successive levels.

  1. 3.

    Covariance matrix computation

    To reduce the computational complexity of KLT, a sub-sampling strategy is used to calculate the covariance matrix of the spectral tiled group, as in [42, 49]. In the sub-sampling strategy, 1% of the spectrum vectors are selected randomly to evaluate the covariance matrix, while the PSNR is not decreased. The complexity of the evaluation of the covariance matrix of full KLT is \(\left( {B_{{{\text{KLT}}}}^{2}{M_{{\text{KLT}}}}{N_{{\text{KLT}}}}} \right),\) where BKLT is the band number, and MKLT and NKLT are the size of each band in the horizontal and vertical direction. After the use of the sub-sampling strategy, the computational complexity of the covariance matrix is decreased to \(\left( {B_{{{\text{KLT}}}}^{2}{M_{{\text{KLT}}}}{N_{{\text{KLT}}}}{\text{/100}}} \right).\)

  1. 4.

    Eigenvalues and eigenvectors computation

    The Jacobi algorithm is used to implement the eigenvalue and eigenvector computations in the KLT process. In order to increase the calculation speed and to be easily realized by hardware, an optimized Jacobi algorithm is used based on the pipeline and parallel architecture detailed in [50]-[51].

  1. 5.

    Spectral transform computation

    Several lifting strategies of KLT are used to implement spectral transform computation, as given in [52,53,54,55,56]. According to the aforementioned tiling spectral bands, the associated lifting structure of spectral transform computation can be designed. The spectral transform is expressed as

$$G={V^T}H=\left[ {\begin{array}{*{20}{c}} {{G_1}} \\ {{G_2}} \end{array}} \right]={V^T}\left[ {\begin{array}{*{20}{c}} {{H_1}} \\ {{H_2}} \end{array}} \right],$$
(1)

where V is eigenvector of the covariance matrix of H=[H1, H2]T. The covariance matrix of H is denoted as Cov(H), which can be expressed as:

$${\text{Cov}}(H)=\frac{1}{4}{H^T}H=\left[ {\begin{array}{*{20}{c}} {{{\operatorname{cov} }_{11}}}&{{{\operatorname{cov} }_{12}}} \\ {{{\operatorname{cov} }_{21}}}&{{{\operatorname{cov} }_{22}}} \end{array}} \right].$$
(2)

V is expressed as

$${V^{\text{T}}}{\text{=}}\left[ {\begin{array}{*{20}{c}} {{v_{11}}}&{{v_{12}}} \\ {{v_{21}}}&{{v_{22}}} \end{array}} \right].$$
(3)

The eigenvector is computed as:

$${v_{11}}={v_{22}}=\sqrt {\frac{1}{2}+\frac{{({{\operatorname{cov} }_{11}} - {{\operatorname{cov} }_{22}})}}{{2\eta }}} =\sqrt {1 - {v_{21}}^{2}} ,$$
(4)
$${v_{21}}= - {v_{12}}=\frac{{{{\operatorname{cov} }_{12}}}}{{\left| {{{\operatorname{cov} }_{12}}} \right|}}\sqrt {\frac{1}{2} - \frac{{{{\operatorname{cov} }_{11}} - {{\operatorname{cov} }_{22}}}}{{2\eta }}} ,$$
(5)
$$\eta =\sqrt {{{({{\operatorname{cov} }_{11}} - {{\operatorname{cov} }_{22}})}^2}+4{{\operatorname{cov} }_{12}}{{\operatorname{cov} }_{21}}} .$$
(6)

Let g = v21, V is expressed as

$$V{\text{=}}\left[ {\begin{array}{*{20}{c}} {\sqrt {1 - {g^2}} }&{ - g} \\ g&{\sqrt {1 - {g^2}} } \end{array}} \right].$$
(7)

Since Det (A) = 1 [57], V can be decomposed as:

$$V{\text{=}}\left[ {\begin{array}{*{20}{c}} t&{ - g} \\ g&t \end{array}} \right]{\text{=}}\left[ {\begin{array}{*{20}{c}} 1&0 \\ f&1 \end{array}} \right]\left[ {\begin{array}{*{20}{c}} 1&{ - g} \\ 0&1 \end{array}} \right]\left[ {\begin{array}{*{20}{c}} 1&0 \\ f&1 \end{array}} \right],$$
(8)

where t = (1 − g2)0.5 and f = (t − 1)/− g. Based on Eq. 8 the lifting structure of spectral transform computation is designed as shown in Fig. 3. The computation structure requires only nine operations and a conditional permutation.

Fig. 3
figure 3

Lifting structure of FKLT for MIHFBs

2.3 Improved PT having only fast PCA basis

To implement post-transform [58, 59], each sub-band of all the high frequency sub-bands, such as LHi, HLi, HHi, i = 1,2,…,L, is organized into several blocks. When the distance between wavelet coefficients is larger than 4 pixels, the correlation between coefficients is relatively weak [60]. Therefore, each block is composed of 4 × 4 wavelet coefficients. The post-transform compression idea is that wavelet coefficient blocks are further transformed on a one particular direction basis (such as Bandelet, DWT, DCT and PCA) in a dictionary. The post-transform key idea is that wavelet coefficient blocks perform transform on a one certain direction basis in a dictionary. A wavelet coefficient block is denoted by \(f\), which is composed of 16 wavelet coefficients. f can also be considered as a vector with A (A = 16) elements in \({\mathbb{R}}^{A}\). Let D denote the post-transform dictionary, which is composed of multiple orthonormal bases of \({\mathbb{R}}^{A}\). Let S be the number of basis in D. Let Bb be an orthonormal basis function in \(D=\left\{ {{B_1},{B_2}, \ldots {B_b}} \right\},\) where \(b \in [1,S]\) is the basis index. The A vectors of the orthonormal basis Bb are expressed by \(\phi _{a}^{b}\) with \(a \in [1,A],B_{b} = \left\{ {\phi _{a}^{b} } \right\}_{{a(a = 1)}}^{{bA}} .\) Let fb be a post-transformed coefficients block. The post-transform can be expressed as:

$${f^b}=\sum\limits_{{a=1}}^{{16}} {\left\langle {f,\phi _{a}^{b}} \right\rangle } \cdot \phi _{a}^{b},$$
(9)

where f is a wavelet coefficients block.

Each wavelet coefficient block is projected onto \(S\) bases in the post-transform dictionary. The \(S\) post-transformed block can be obtained. Therefore, a best post-transformed coefficient block needs to be selected from the \(S\) post-transformed blocked. In general, a minimum Rate-Distortion (RD) Lagrangian cost method is used to select a best post-transformed block, which can be expressed as:

$$L(f_{q}^{b})=D(f_{q}^{b})+\lambda \cdot R(f_{q}^{b}),$$
(10)

where \(f_{q}^{b}\) represents the quantized post-transformed coefficients, the quantization step is \(q,\;\left( {f_{q}^{b}} \right)\) is the square error between fb and \(f_{q}^{b}\), \(\lambda\) is a Lagrangian multiplier, and \(R\;\left( {f_{q}^{b}} \right)\) is the allocated bit-rate to encode \(f_{q}^{b}\).

In our previous work [16, 17, 19] an improved post transform encoder was proposed. Some optimized strategies were used, such as a very simple dictionary which can be composed of the Hadamard transform (HT) and DCT basis, l0 and l1 norm for the best post-transform basis selection. Our previous work points out that PCA is the best post transform basis. However, PCA has very heavy computation complexity. It is not suitable for post-transform application. Because the fast strategy is utilized in computing KLT in the above section, the fast PCA is also implemented. So, the fast PCA as a post transform basis is used in this paper. The proposed encoder is shown in Fig. 4. The improved post-transform in the DWT domain is used to remove spatial redundancies. A fast PCA basis is adopted to further exploit the spatial redundancy within a wavelet-transform sub-band due to better sparse performance.

Fig. 4
figure 4

Improved post-transform encoder for MIHFBs

The post-transformed coefficients are encoded by the BPE encoder which is a zero tree encoder. However, the post-transform destroys the zero structure of wavelet coefficients. Each wavelet sub-band is performed by 4 × 4 post-transform. In one BPE block, the structure relationship between grandchildren and children coefficients is 2 × 2 children coefficients and 4 × 4 grandchildren. The post-transform block is 4 × 4. The process is expressed in Fig. 5.

Fig. 5
figure 5

Post-transform block structure

If BPE builds the block structure as traditional wavelet spatiotemporal orientation trees, the 4 × 4 post-transform block of HL2, LH2 and HH2 consists of children coefficients in four orientation trees, and the 4 × 4 post-transform block of HL3, LH3 and HH3 consists of grandchildren coefficients in four orientation trees. The four orientation trees can be expressed as:

$${\text{SO}}{{\text{T}}^a}=\left\{ {C_{0}^{a},C_{1}^{a},C_{2}^{a},G_{0}^{a},G_{1}^{a},G_{2}^{a}} \right\},$$
(11)
$${\text{SO}}{{\text{T}}^b}=\left\{ {C_{0}^{b},C_{1}^{b},C_{2}^{b},G_{0}^{b},G_{1}^{b},G_{2}^{b}} \right\},$$
(12)
$${\text{SO}}{{\text{T}}^c}=\left\{ {C_{0}^{c},C_{1}^{c},C_{2}^{c},G_{0}^{c},G_{1}^{c},G_{2}^{c}} \right\},$$
(13)
$${\text{SO}}{{\text{T}}^d}=\left\{ {C_{0}^{d},C_{1}^{d},C_{2}^{d},G_{0}^{d},G_{1}^{d},G_{2}^{d}} \right\}.$$
(14)

In BPE encoding, the four orientation trees are located in four separate encoding blocks of BPE. Grandchildren coefficients in each orientation tree are exactly one post-transform block. However, children coefficients are part of one post-transform block. This changes the original zero tree relationship between grandchildren and children coefficients. In this paper, a new BPE method is proposed to encode post-transform coefficients. To ensure the structures of spatiotemporal orientation trees, the structure of the traditional block is modified to four orientation trees representing one block, which is shown in Fig. 6.

Fig. 6
figure 6

Modifying the structure of the post-transform in the proposed method, the points represent the grandchildren coefficients in different directions

The structure of the improved block includes three families. In each family, father coefficients are 4 × 4 coefficients, children coefficients are 4 × 4 coefficients, and grandchildren coefficients are 8 × 8 coefficients. Since grandchildren coefficients consist of a 4 × 4 post-transform block and a new family block is 8 × 8, the four post-transform coefficients of HL3, LH3 and HH3 in Fig. 5 are reordered. The energy of the reconstructed post-transform block can be concentrated in the top left corner, and low coefficients are concentrated in other parts. This can ensure the spatial orientation trees, which are beneficial, to be encoded by tranB, tranD, tranG, and \({\text{tra}}{{\text{n}}_{{H_{i,j}}}}\). Finally, the four orientation trees are separated into four blocks, which are encoded by traditional BPE.

2.4 DSC-CCSDS strategy

The DSC method can obtain better compression performance of multispectral or hyperspectral images in the remote sensing application [61,62,63,64,65,66,67]. In our previous work in [19], a DSC strategy was combined with CCSDS-IDC to encode transformed coefficients. In this paper, the encoding process is optimized. The optimized coding process of the Slepian-Wolf bit plane encoding (SW-BPE) is summarized in Fig. 7.

Fig. 7
figure 7

DSC-CCSDS encoding flow in the proposed method

3 Experimental results

3.1 Experiment setup

The proposed approach is tested and analyzed on a self-developed multispectral compression testing platform to verify the performance and feasibility of the proposed approach. The testing platform is shown in Fig. 8. The testing platform includes a multispectral image simulation source, a compression system, a compression and storage sever (CSS), a decoding unit, and a display system. The CSS can produce simulated multispectral images with three and four bands, which are transmitted to the image simulation source unit. The image simulation source unit adjusts the output line frequency, image size, and output time to simulate TDICCD output. The compression system compresses the received multispectral image or simulated multispectral image. The proposed method is implemented and tested on the compression system. The compression system uses Xilinx FPGA with a 32-bit MicroBlaze as the processor. The compressed streams are received and decoded by a decoding unit. In the decompression method, a deblocking filter is used to remove the block effects because the multispectral images are spatially titled before spectral transform in the proposed method. The reconstructed image is transmitted to the CSS.

Fig. 8
figure 8

Compression testing platform of the proposed method

3.2 Compression experiments using image simulation sources

In order to verify the proposed compression scheme, the image simulation source unit produces several groups of multispectral remote sensing images that are transmitted to the compression system. They are rich in texture to verify validity and feasibility of the proposed approach. Each group has three bands, measuring radiation from the Earth filtered at wavelengths red, green, and blue. Each band of the remote sensing image is 8 bpp (bits per pixel). The FKLT is performed on each group of the multispectral remote sensing images. A three-level 9/7 Daubechies DWT and a fast PCA-based PT are performed on each band, and then the post-transformed coefficients are quantized. Figure 9 demonstrates the three reconstructed bands and the magnified sub-images. Each band is 491 × 491 pixels. The code rate of the key bands is 2.0 bpp. The PSNR of three bands can reach 51.56 dB, 50.48 dB, and 50.29 dB, respectively. After multispectral remote sensing images are transformed by the proposed FKLT-PT scheme, the bit-depth of transformed high frequency coefficients is 0–6 bit. The bit-depth of most of the coefficients is 0–4 bit. There are no large high-frequency coefficients. All reconstructed bands have no blocking effect, and high recovery-quality bands can be obtained.

Fig. 9
figure 9

Reconstructed multi-band images using the propose method at 2.0bpp, (a) reconstructed band1, (b) reconstructed band2, and (c) reconstructed band3

In addition, the image simulation source unit produces several groups of Quickbird multispectral images to test the proposed algorithm. Each group has four bands, measuring radiation from the Earth filtered at wavelengths red, green, blue, and near-infrared. Each band in the remote sensing, multispectral images is 8 bpp. Figure 10 demonstrates the first reconstructed band and the magnified sub-image. From the magnified sub-image, the proposed algorithm can effectively preserve the edge information of the original multispectral images. Other groups can also obtain the same results.

Fig. 10
figure 10

Magnification of one segment of the first reconstructed band using the proposed method

Finally, a multi-spectral image is used to compare the compression experiments with two different methods. The captured images are compressed by the proposed algorithm and CCSDS-IDC approach at 0.5bpp. Figure 11 demonstrates the reconstructed image. Figure 11a, d show the original image and magnification of one segment S. Figures 11b, e show the reconstructed image of CCSDS-IDC. Figure 11c, f show the reconstructed image of the proposed method. As the figures show, the CCSDS-IDC method has the fuzzy phenomenon of a reconstructed image at low bit rates. The proposed method has no ringing effect and emerging distortion. Other bands depict the same results. Therefore, the proposed method works well for image compression while edge and outline information is substantially retained.

Fig. 11
figure 11

Reconstructed multi-band images with different methods at 0.5 bpp, (a) The original first band, (b) Reconstructed band of S using CCSDS-IDC method, (c) Reconstructed band of S using the proposed method, (d) Magnification of one segment of the first band, (e) and (f) Magnification of two reconstructed images

3.3 Compression performance testing experiments

In order to objectively evaluate the performance of the proposed compression algorithm, the CSS injects multiple groups of AVIRIS uncalibrated multispectral images, including cities and buildings, into the image simulation sources to perform the compression experiment at different bit rates (BR). Each group is 512 pixel × 512pixel × 4, while the bit depth of every pixel is 8bit.

The compression bit rate is set to 0.25–2.0 bpp. The PSNR of the original and reconstructed multispectral images is used as the evaluation equation, which can be expressed as:

$${\text{PSNR}}=20{\log _{10}}\left( {\frac{{{2^Q} - 1}}{{\sqrt {{\text{MSE}}} }}} \right)({\text{dB}}),$$
(15)
$${\text{MSE}}=\frac{1}{{W \times H \times B}}\sum\limits_{{k=0}}^{{B - 1}} {\sum\limits_{{i=0}}^{{W - 1}} {\sum\limits_{{j=0}}^{{H - 1}} {{{\left( {{x_{i,j,k}} - x_{{i,j,k}}^{\prime }} \right)}^2}} } } ,$$
(16)

where Q is the quantization bits of the original image sample, B is the band number of the multispectral image, W and H are the size of each band, \({x_{i,j,k}}\) and \(x_{{i,j,k}}^{\prime }\) denote the pixel value at (i, j,k) in original and reconstructed multispectral images, respectively.

First, the three groups (Hawaii, Jasper Ridge, and Lunar Lake,) of multispectral images with different textures are used as the testing images. Figure 12 shows the PSNR results of three group AVIRIS multispectral images using the proposed algorithm. In Fig. 12, the PSNR reaches above 40dB at 0.25–2 bpp. Second, the proposed algorithm on multiple groups for multispectral images is compared with PCA, SPIHT + 2D-DWT with KLT [68], SPECK + 2D-DWT [69] with KLT, and POT [70]. These four methods use the quantization method and entropy encoder of the JPEG2000 [71]. From Fig. 12, a compression algorithm is applied to two different multispectral images and two different PSNRs are obtained. We use 50 images from the testing image database are used to measure the corresponding PSNR. The average PSNR is considered as the PSNR of the corresponding method. The PSNR comparisons are demonstrated in Fig. 13a, b. Due to the full usage of FKLT and PT in the DWT domain, the proposed compression achieves a good compression performance. The proposed method can improve PSNR by 0.3–1.7 dB against SPIHT + 2D-DWT with KLT at 3.5–0.25 bpp.

Fig. 12
figure 12

Tested PSNR results with the proposed method on three group multispectral images

Fig. 13
figure 13

PSNR comparison with traditional compression method at different bit rate

Third, the proposed algorithm is compared with the CCSDS-IDC space compression standard method. As the same considerations with Fig. 13, the same compression algorithm has a different PSNR when a different multispectral image is used as the input. The same groups (50 images from the testing image database) of multispectral images are used as test input for compression algorithms. The average PSNR is considered as the PSNR of the corresponding method. Figure 14 shows the comparison of results for different compression algorithms. Since the compression method based on CCSDS-IDC without KLT does not remove the spectral redundancies, the compression performance is 2dB lower than CCSDS-IDC with KLT. In the proposed algorithm, FKLT and improved PT in the DWT domain are utilized to remove spatial and spectral redundancies, and DSC-CCSDS encodes the transform coefficients. Therefore, the PSNR of the proposed algorithm is 0.15–1.57 dB higher than CCSDS-IDC with KLT.

Fig. 14
figure 14

Test result comparison (PSNR) with space compression standard at different bit rate

Finally, other evaluation parameters are used to analyze the compression performance of the proposed method. The CSS is injected by the same multispectral images. Other image quality assessment methods are performed by using the mean measure of structural similarity (MSSIM) [72] and visual information fidelity (VIF) [73]. The MSSIM and VIF are based on the hypothesis that the human visual system (HVS) is highly adopted for extracting structural information. The reference methods use the CCSDS-IDC method with KLT and the CCSDS-IDC method without KLT. The MSSIM and VIF values at different compressed ratios are shown in Table 1. Because several key technologies are used, such as the FKLT to remove spectral redundancy, and the improved post-transform method, the proposed method outperforms the traditional saliency-based methods.

Table 1 MSSIM and VIF values at different compression ratio

3.4 Compression complexity analysis

In the proposed method, FKLT is performed on the spectral dimension to exploit spectral information. Let N be the band number of the multispectral image. The spectral band tiling is used based on multi-level decomposition to perform KLT. Let K be the number of sub-band unit (SBU) decomposed by spectral band tiling. For each SBU, N = 2. Only a small sample of the image spatial locations is also used to compute the covariance matrix. Let ρ be Sub-sampling factor of the covariance matrix. The KLT calculation complexity of each SBU is shown in Table 2.

Table 2 Calculation complexity of each SBU, M and L are the number of spatial locations

After applying FKLT, three-level 2D DWT is applied to the spatial bands. Let L be the decomposition level of 2D DWT. For the \({I_1} \times {I_2} \times {I_3}\) MIHFBs performing 2D-DWT, the calculation complexity is \(\left( {8N{I_1}{I_2}{I_3}(1 - {2^{ - 2L}}} \right){\text{/}}6.\) In the proposed method, a 9/7 2D-DWT is used to transform each band of MIHFBs. Three-level decomposition is performed. Therefore, the calculation complexity of the proposed method is \(O(9{I_1}{I_2}{I_3}{\text{/}}7).\) After MIHFBs performing 2D DWT, The fast PCA is used based on pairwise orthogonal transform to implement post-transform. Each post-transform block has 4 × 4 coefficients. Each block is taken as m vectors, and each vector has n components. The computational cost is 12mn+21n-8m-23. After 2D-DWT coefficients performing post-transform, The BPE-DSC encoder adopts low-density parity-check codes (LDPC), of which the encoding complexity is \(O\left( {{N^2}} \right).\) In this paper, quasi-cyclic LDPC are used for fast encoding. The encoding complexity is \({0.17^2}{N^2}+O(N),\) where N is the code length. The basic operations are very simple, as they are sums in modulo-2 arithmetic.

In the following section, the calculation times of the proposed compression method are tested and analysed. TDICCD imaging mode is used to test the compression time of the proposed method. The line frequency of TDICCD is 0.86 KHz. The following compression time of the proposed algorithm is only used to perform the evaluations of compression time. The compression algorithm is not optimized for FPGA implementation. These evaluations are based on the lossy compression of remote sensing multispectral images with four bands. The size of each band is 3072 × 128. The compression time of the proposed algorithm is compared with that of other algorithms. Table 3 compares the results of the compression time for different compression algorithms. The processing time of the proposed algorithm is 0.049 us/sample, which indicates less time than other methods. The compression time of 128 × 3072 needs only 19.26 ms. According to the different principles of TDICCD imaging, the proposed compression algorithm can be optimized on an FPGA. An optimized implementation on a FPGA can take minimal time.

Table 3 The results of complexity comparison

In addition, a compression method with a low memory requirement can be used in memory constrained environments. In order to evaluate the occupancy of the resources, the XC2V6000-6FF1152 FPGA is utilized to implement the proposed algorithm. These resources can implement the operation of the compression algorithm. The design language is VerilogHDL, the development platform is ISE8.2, and the synthesis tool is XST. Table 4 demonstrates the occupancy of resources of the proposed approach. From Table 3, the LUT occupies 81%, Slices occupy 76%, and BRAM occupies 86%. Various indicators are lower than 95%, which meet the requirement of our project.

Table 4 The occupancy of resources

Overall, the proposed algorithm in this paper achieves the excellent lossy compression performance and is very suitable for remote sensing multispectral images with a few bands.

4 Conclusion

In order to improve the compression performance of remote sensing MIHNBs, a low-complexity and efficient compression scheme has been proposed based on a distributed post-transform in the wavelet domain in conjunction with a fast spectral decorrelator. The new method uses a FKLT as the spectral decorrelator, followed by a post-transform having only fast PCA basis in the wavelet domain for the spatial decorrelation. In this algorithm, a new distributed BPE is proposed for encoding the post-transform coefficients. Multispectral images having different textures are used to verify validity and test compression performance of the proposed algorithm. Experimental results show that the proposed compression method offers the excellent compression performance based on different performance evaluation indices, such as PSNR, MSSIM and VIF. The experimental results are compared with various state-of-the-art compression techniques and exhibit a performance improvement of 0.3–1.7 dB PSNR over current techniques. The proposed algorithm is very suitable for MIHFBs and has the potential to compress MIHNBs and hyperspectral images. The next stage will be to optimize the proposed algorithm on an onboard hardware platform. We will also simplify the proposed algorithm, slightly compromising performance, to develop an on-board compressor.