Keywords

1 Introduction

HYPERSPECTRAL imaging (HSI) is one of the most powerful technologies to remotely detect and recognize the material properties of the object in the interest scene for the broadband wavelength and high spectral resolution [1,2,3] However, a large number of spectral bands pose a great challenge for information extraction. For example, in hyperspectral image classification, the classification accuracy decreases with the increase of the number of bands relative to a small number of tag samples. Therefore, it is very necessary to reduce the amount of data and save resources. Two methods are selected for dimension reduction: feature extraction and band selection. Feature extraction uses mapping methods to convert raw data to fewer new features including principal component analysis (PCA) and independent component analysis (ICA) [4]. However, feature extraction will change the original data to a certain degree. Different from feature extraction, bands selection selects proper band subsets from the original data set [5]. It not only preserves the original features and physical meaning of remote sensing data but also reduces the size of the data to achieve the goal of dimensionality reduction.

Recently, some hyperspectral BS methods have been proposed for hyperspectral classification and target detection. Weiwei sunpresented a novel symmetric sparse representation (SSR) method to solve the band selection problem in hyperspectral imagery classification [6]. H Su et al. proposed a particle swarm optimization (PSO)-based system to select bands and determine the optimal number of bands to be selected simultaneously [7]. Low-rank representation (LRR) is a new tool for robustly and efficiently processing high-dimensional data. Xiaotao Wang proposed a predimension-reduction algorithm that couples weighted low-rank representation (WLRR) with a skinny intrinsic mode functions (IMFs) dictionary for hyperspectral image (HSI) classification. Alex Sumarsono et al. introduced the LRR model into the hyperspectral image and used it to estimate the number of subspaces [8,9,10,11]. However, these low-rank-based methods do not make full use of the resulting low-rank information.

In this paper, we developed an unsupervised BS approach, to be called low-rank band selection (LRBS) for hyperspectral image processing where no a priori knowledge available to be used. The organization of this paper is organized as follows. In Sect. 2, the proposed method is specifically described. Section 3 presents the extensive experiments and conclusions are drawn in Sect. 4.

2 LRBS Method

2.1 The LRR Model

A hyperspectral image is denoted as \( X = (x_{1} ,x_{2} , \cdots ,x_{L} )^{T} \in R^{L \times n} \), where L is the number of bands, n is the total number of pixels, and \( \varvec{x}_{\varvec{i}} = (x_{i,1} ,x_{i,2} , \cdots ,x_{i,n} )^{T} \) is the \( i^{th} \) spectral band. The HSI data includes the low-rank target parts and the noise part. Therefore, the following model is defined as low-rank representation of HSI:

$$ \min\Gamma ({\mathbf{Z}}) + \lambda \left\| {\mathbf{E}} \right\|_{1,2} ,\quad st.\text{ }\quad X = DZ + E, $$
(1)

where \( \Gamma ( \bullet ) \) is the coefficient of the low-rank matrix, D is the dictionary of HSI, E denotes the noise, \( \lambda \) is the balance parameter of the two parts of the equation. You can get \( \lambda \) by

$$ \lambda = \frac{1}{{\sqrt {\log L} }} $$
(2)

The hyperspectral image contains redundancy and noise information. The singular value decomposition is used to filter the image noise to get the data dictionary in this paper. The specific approach is as follows: First, singular value decomposition is performed on \( X = U\Sigma V \), and V is right singular matrix with the size of n*m, U is left singular matrix with the size of L*m, \( \Sigma \) is semi-positive definite m*m order diagonal matrix whose diagonal elements are singular value. The data dictionary is calculated by the following equation:

$$ D = U*\Sigma *V^{{\prime }} $$
(3)

2.2 Solution of the LRR Model

The rank minimization problem is known to be NP-hard, and the most popular choice is to replace rank with the nuclear norm to make the minimization to be convex relaxation problem. Therefore, the Eq. (1) can be written as follow:

$$ \hbox{min} \left\| Z \right\|_{*} + \lambda \left\| E \right\|_{1,2} ,\quad s.t.\quad X = DZ + E, $$
(4)

where \( \left\| Z \right\|_{*} \) is the nuclear norm of Z.

The augmented Lagrange multiplier (ALM) method is quite prevalent in the LRR-related works [12]. In this section, inexact-ALM is extended to handle the proposed LRR problem in (4). By introducing the intermediate variable J, it can be rewritten in the following equivalent form

$$ \hbox{min} \left\| J \right\|_{*} + \lambda \left\| E \right\|_{1,2} \quad s.t.\text{ }\quad X = DZ + E,Z = J $$
(5)

With the aid of the Lagrange multipliers Y1 and Y2, (5) can be redefined in the final optimization problem (6)

$$ \begin{aligned} & \mathop {\hbox{min} }\limits_{{Z,E,J,Y_{1} ,Y_{2} }} ||Z||_{*} + \lambda ||E||_{1,2} + < Y_{1} ,X - DZ - E > \\ & \quad + < Y_{2} ,Z - J > + \frac{\mu }{2}(||X - DZ - E||_{F}^{2} + ||Z - J||_{F}^{2} ) \\ \end{aligned} $$
(6)

In (6), each unknown variable of Z, E, J, Y1, and Y2 is optimized in iterative fashion that is to successively optimize one variable with the other fixed variables. The solving procedure is summarized in Algorithm 1. In each iteration, the variables J and E are updated by the singular value thresholding operator and the \( l_{2,1} \) norm thresholding operator. As for the variable Z, it has a closed solution for each column shown in step 2.2. The steps of algorithm is described as bellow.

figure a

2.3 The LRBS Algorithm

This section describes the proposed band selection algorithm based on low-rank representation (LRBS). We use the technique of spectral clustering algorithm to compute the eigenvectors of graph Laplacian which weight is composed of the low-rank matrix. The algorithm implementation needs to use K-means to cluster a data set into some subsets. The number of categories k in k-means is obtained by virtual dimensionality (VD) algorithm. Different from the classic K-means, in this paper, we utilized a fixed initial cluster centers instead of the random initial ones.

Assume that \( \left\{ {{\mathbf{b}}_{l} } \right\}_{l = 1}^{L} \) is a set of band number of a hyperspectral image cube where bl is the l th spectral band represented by a column vector, \( {\mathbf{b}}_{l} = \left( {b_{l1} ,b_{l2} , \cdots ,b_{lN} } \right)^{T} \) and \( \left\{ {b_{li} } \right\}_{i = 1}^{N} \) is the set of all N pixels in the l th band image. The implementation steps of the BS method proposed in the paper is listed as following.

figure b

3 Hyperspectral Image Experiments

In this section, we conducted a series of experiments to evaluate the classification performance of the selected bands by different BS methods. The compared algorithms include uniform BS (UBS), minimum estimated abundance covariance (MEAC), multigraph determinantal point process (MDPP), dominant set extraction BS (DSEBS), and the proposed BS method. The experiment platform is listed as follows: the machine operating system used is Windows10, the machine configuration is: CPU frequency 2.50 GHz, running memory 8 GB software environment: MATLAB R2014a. The three hyperspectral datasets we used in the experiment include Purdue University’s Indiana Indian Pines image, Salinas Valley image, and the image of University of Pavia.

Table 1 shows the specific band subsets selected using different methods on three datasets. Tables 2, 3, and 4 show the classification accuracy of the bands selected using different methods for hyperspectral classification. It can be seen from Table 2 that the band subsets selected by the LRSB method have higher classification accuracy than other classifications in subsequent classifications such as 1st, 2nd, 6th, 7th, 8th, 9th, and 14th classes in Purdue’s data. It can also be seen from Tables 3 and 4 that the band classification selected by the LRBS method generally achieves higher classification accuracy in image classification. Therefore, the experiment shows that the LRBS algorithm provided in this paper can indeed select a subset of frequency bands that perform better.

Table 1. Bands selected by UBS, MEAC, MDPP, DSEBS, LRBS
Table 2. PD, POA, PR calculated from the classification results for Purdue’s data
Table 3. PD, POA, PR calculated from the classification results for Salinas
Table 4. PD, POA, PR calculated from the classification results for University of Pavia

4 Conclusion

This paper proposed an unsupervised low-rank representation of the band selection method. First, the low-rank representation of the hyperspectral image is solved using an inexact -ALM algorithm to obtain a low-rank coefficient matrix. Each column of the low-rank coefficient is used as a vertex of the graph to perform spectral clustering to determine the initial k-means. The clustering center is clustered to finally obtain a band that satisfies the conditions. Experiments on three HSI datasets show that the frequency band selected by the LRBS algorithm can select a better band subset for image classification.