1 Introduction

In recent years, subspace segmentation has played an important role in image representation, computer vision, and other related fields [1,2,3,4,5]. The ultimate aim of subspace segmentation is to classify data samples to their corresponding subspaces. According to the surveys in related research works [6,7,8], major treatment methods for subspace segmentation include algebraic methods [9, 10], iterative methods [11, 12], statistical methods [13, 14], and spectral clustering-based methods [6,7,8, 15,16,17,18,19,20,21]. Among them, the methods based on spectral clustering have achieved great successes in practical applications.

Generally speaking, a spectral clustering based subspace segmentation algorithm consists of two steps: (1) constructing an affinity matrix for a given data set; (2) obtaining the final segmentation result via a special spectral clustering algorithm (e.g. Ncut [22]) with the constructed affinity matrix. It can be observed that the obtained affinity matrix usually determines the performance of a spectral clustering based subspace segmentation algorithm.

The existing methods for learning affinity matrices could all be regarded as self-representation models. The principal ideology of a self-representation model is that each sample data can be linearly represented by the other data samples from the same data set. For that, Elhamifar and Vidal [15, 20] first proposed a sparse subspace clustering (SSC) algorithm, which aims to calculate a reconstruction coefficient vector by l1-minimization [23] for each data sample and then construct a sparse affinity matrix by concentrating all the reconstruction coefficient vectors. SSC is designed to obtain the sparsest representation of each sample, but the global structures of data sets are usually ignored. To capture the global structures of grossly corrupted data sets, Liu et al. [7, 18] developed a low-rank representation algorithm (LRR) to discover the reconstruction coefficients matrix for all data points jointly. LRR expects the coefficient matrix to have a minimal rank, and this goal can be satisfied by minimizing its nuclear norm. It’s proved that the minimal nuclear norm constraint would make the coefficient matrix is capable of discovering the global structure of corrupted data points. Therefore, LRR performs better than SSC in corrupted subspace segmentation problems.

In addition, a lot of approaches have been proposed to show the capabilities of low-rank methods in different applications [24,25,26,27,28,29]. Wang et al. [24] developed a novel DCT (discrete cosine transform) regularized low-rank method to handle image recovery problems for data with large percentage of corruption. In [25], for solving matrix completion problems, Hu et al. used a kind of truncated nuclear norm, which is gotten by the nuclear norm subtracted by the sum of the largest few singular values, to replace the classical nuclear norm. To accelerate the convergence of the TNNR, Liu et al. presented a TNNR-WRE algorithm, in which different weights were given to the corresponding rows of the residual error matrix [26]. Wang et al. formulated a new robust PCA algorithm to discover the low-rank and noisy components simultaneously [27]. Li et al. [28] proposed a discriminative multi-view interactive image re-ranking algorithm which could describe the images sufficiently by integrating user relevance feedback capturing users’ intentions and multiple features. Li et al. [29] proposed an efficient image retrieval algorithm, called SPA (Spatially Pooled Attributes for image retrieval), which encoded weak spatial information into attribute embedding.

The wide usages in image applications of LRR and its successes in constructing affinity graphs have encouraged lots of related subsequent researches. For example, Chen et al. [30] considered a symmetric low-rank representation to preserve the high-dimensional data sets’ subspace structure. Zhuang et al. [31] presented the locality-preserving low-rank representation to obtain an undirected graph from a mixture of nonlinear manifolds. Zhang et al. [32] claimed that the structured low-rank representation of a data matrix could be learned by using a supervised method to create a discriminant and reconstructive dictionary. Zhuang et al. introduced a non-negative low rank and sparse representation (NNLRSR) method [33] by adding non-negative and sparse constraints of the reconstruction coefficient matrix into LRR. Experiments showed that these algorithms can get promising results in subspace segmentation for different data sets.

However, the original LRR algorithm can only work well under the condition that the subspaces of a data set are totally independent. Therefore, Tang et al. [34] raised a structure-constrained low-rank representation (SCLRR) to handle disjoint subspace segmentation problems. SCLRR incorporates the structure information of data sets by appending a regularization term in the objective function of LRR, which is a feasible approach for solving multiple disjoint subspace segmentation problems. On the basis of SCLRR, Li et al. [35] advanced a structure-constrained low-rank dictionary learning (SCLRDL) algorithm to combine low-rank constraint with structure information into coefficient matrix for image classification. Wu et al. [36] overcome the shortcomings of existing techniques in handling disjoint subspaces and proposed a CS-LRR algorithm to perform optimal spectral clustering of the subspaces. To find the optimal low-rank affinity matrices of disjoint data sets for Ncut, Wei et al. proposed a SCSLRR algorithm [8] to combine the objection functions of K-means, Ncut, and LRR together.

For SCLRR, its appended regularization term brings an additional parameter into LRR and different values of the parameter have great effect on the efficiency of the algorithm. Furthermore, there is already a parameter for the residual term in the original LRR algorithm, hence SCLRR has been a more complex double parameters problem. In Fig. 1, we give a simple experiment on Fashion-MINIST [37] data set to observe the sensitivity of SCLRR to the additional regularization parameter β, it can be seen that the performance of SCLRR varies drastically with β.

Fig. 1
figure 1

The sensitivity of SCLRR to parameter β when λ is fixed, λ is the parameter to the residual term

Therefore, we attempt to find a better way to introduce the structure information of data sets into LRR and solve the disjoint subspace segmentation problems. In this paper, a new improved structured low-rank representation algorithm, abbreviated as ISLRR, is presented. In ISLRR, we introduce the structure information of data sets into the equality constraint term of LRR. It can be found that ISLRR evades the difficulty of tuning two parameters. Experiments on some benchmark data sets demonstrate that ISLRR is easy to implement and efficient for disjoint subspace segmentation tasks.

The main framework of this article is as follows: In Sect. 2, we briefly introduce LRR and SCLRR. In Sect. 3, we give the motivation of ISLRR and present an optimization algorithm to solve ISLRR. In Sect. 4, we give some further discussions about the proposed algorithm. Experiments for verifying the effectiveness of our algorithm are shown in Sect. 5. Finally, Conclusions are presented in Sect. 6.

2 Related Algorithms

In this section, we briefly describe LRR and its extended algorithm SCLRR.

2.1 Low-Rank Representation

Suppose a data matrix \( {\text{X}} = \left[ {{\text{x}}_{1} ,{\text{x}}_{2} , \ldots ,{\text{x}}_{n} } \right] \in {\mathbb{R}}^{d \times n} \), where d is the dimensions of each data sample and n is the number of all samples. Because each data sample can be lineally represented by the data samples from the same subspace, the model can be written as:

$$ {\text{X}} = {\text{AZ}} , $$
(1)

where \( {\text{A}} = \left[ {a_{1} ,a_{2} , \ldots ,{\text{a}}_{m} } \right] \in {\mathbb{R}}^{d \times m} \) is a dictionary and Z is the reconstruction coefficient matrix. In LRR, X is used as the dictionary and Z is hoped to have minimal rank. Then the purpose of LRR can be expressed as:

$$ \mathop { \hbox{min} }\limits_{Z} rank\left( Z \right)\;{\text{s}} . {\text{t}} .\;{\text{X}} = {\text{XZ}}. $$
(2)

However, in real-world applications, data points are usually noisy and even grossly corrupted. So the objective function of LRR can be rewritten as:

$$ \mathop {\hbox{min} }\limits_{Z,E} \parallel Z\parallel_{* } + \lambda \parallel E\parallel_{2,1 } \;{\text{s}} . {\text{t}} .\;{\text{X}} = {\text{XZ}} + {\text{E}} , $$
(3)

where \( \parallel E\parallel_{2,1 } = \sum\nolimits_{i = 1}^{n} {\sqrt {\sum\nolimits_{j = 1}^{n} {\left( {\left[ E \right]_{ij} } \right)^{2} } } } \) ([E]ij is the (i, j)-th element of E) is called l2,1-norm of noise E and \( \lambda > 0 \) is a parameter to balance the contributions of two parts.

Augmented Lagrange multipliers (ALM) [18] method can be used to solve the problem (3). Once Z is obtained, we can define \( {\text{G}} = \left( {Z + Z^{T} } \right)/2 \) (ZT is the transpose of Z) as the affinity graph and use Ncut [22] to get the final clustering results.

2.2 Structure-Constrained Low-Rank Representation

As we know, LRR algorithm can only work well on data sets with independent subspaces. Independent subspaces segmentation is a special case of the disjoint subspace segmentation. To improve the capability of LRR for handling disjoint subspace segmentation problems, SCLRR adds a weighted sparse matrix \( \left| {{\text{M}} \odot {\text{Z}}} \right|_{1} = \sum\nolimits_{i,j} {M_{ij} \left| Z \right|_{i,j} } \) (M is a matrix which characterizes the locality structure information of a data set) into the objective function to enhance the structure information of Z. For that, the penalty of rank is weaker than that in the objective function which only considers the nuclear norm, therefore, the structure constraint can also improve the rank of the solution. As the above descriptions, considering the noise and outlines, SCLRR can be defined as follows:

$$ \mathop {\hbox{min} }\limits_{Z,E} \parallel Z\parallel_{* } + \beta \left| {{\text{M}} \odot {\text{Z}}} \right| + \lambda \parallel E\parallel_{2,1 } \;{\text{s}} . {\text{t}} .\;{\text{X}} = {\text{XZ}} + {\text{E}}, $$
(4)

where \( \odot \) is called Hadamard product [38], β and λ are two parameters to balance the devotions of three parts. If two matrices A, B both have same size m × n, the Hadamard product means that:

$$ {\text{A}} \odot {\text{B}} = \left[ {\left[ {\begin{array}{*{20}l} {\left[ A \right]_{11} \left[ B \right]_{11} } \hfill & \cdots \hfill & {\left[ A \right]_{1n} \left[ B \right]_{1n} } \hfill \\ \vdots \hfill & \ddots \hfill & \vdots \hfill \\ {\left[ A \right]_{m1} \left[ B \right]_{m1} } \hfill & \cdots \hfill & {\left[ A \right]_{mn} \left[ B \right]_{mn} } \hfill \\ \end{array} } \right]} \right] . $$
(5)

SCLRR was proved to be a better way to draw out the subspace structures of different data sets than LRR.

3 Improved Structured Low-Rank Representation

In this section, we will present an improved structured low-rank representation algorithm (ISLRR). Moreover, an optimizing algorithm is also proposed for solving ISLRR problem.

3.1 Motivation

Based on the analysis of [34], it can be found that the structure of the solution to LRR affects the subspace segmentation results directly. We could hope that the coefficient matrices obtained by LRR would be block-diagonal, namely, the relationships between intra-subspace data samples are strong and the relationships between inter-subspace data samples are weaker.

However, LRR is insufficient to handle disjoint subspace segmentation problems. For the sake of improving the ability of LRR, it is a clear way to ameliorate the structure of its solution. The purpose of introducing structural information into the algorithm is that we can achieve the block-diagonal solution even if subspaces are disjoint. From the descriptions in Sect. 2.2, SCLRR can reveal the subspace structures of data sets more precisely than LRR, however, the difficulty of SCLRR is greatly increased because of the adjustment of the additional parameter β, and an improper value of β will decrease the performance of SCLRR sharply.

Stated thus, we expect a better algorithm without extra constraint to discover the structure information of data sets. As described in SCLRR [34], M is a weighted matrix that reflects the structural relationships between data samples and a new weighted constraint \( \left| {{\text{M}} \odot {\text{Z}}} \right|_{1} \) is considered to help Z to reveal the structure information of data sets better. Different from SCLRR, we apply the weighed constraint \( {\text{M}} \odot {\text{Z}} \) into the equality constraint item of LRR to form a new dictionary, so that a better low rank matrix for subspace segmentation could be obtained. Considering the noise and outliers, we define the formula of ISLRR as follows:

$$ \mathop {\hbox{min} }\limits_{Z,E} \parallel Z\parallel_{* } + \lambda \parallel E\parallel_{2,1 } \;{\text{s}} . {\text{t}} .\;{\text{X}} = {\text{X}}\left( {{\text{M}} \odot {\text{Z}}} \right) + {\text{E,}} $$
(6)

It is obvious that if we set all the elements of M as 1, we can get LRR from ISLRR. Hence, LRR is actually a special case of LRR. After Z is obtained, we define the affinity matrix as \( {\text{G}} = \left( {Z + Z^{T} } \right)/2 \) and then compute the segmentation results by Ncut [22].

In general, the property of preferable weight matrix M is that the dissimilarities between samples from intra-class are smaller and dissimilarities between interclass samples are larger. An ideal M [34] can be defined as:

$$ \left[ M \right]_{ij} = 1 - { \exp }\left( { - \frac{{1 - \left| {x_{i}^{*T} x_{j}^{*} } \right|}}{\sigma }} \right) , $$
(7)

where \( x_{i}^{*} \) and \( x_{j}^{*} \) are the normalized data points of xi and xj respectively, and σ is set as the mean of \( 1 - \left| {x_{i}^{*T} x_{j}^{*} } \right| \) for all pairwise \( x_{i}^{*} \) and \( x_{j}^{*} \) empirically in this paper.

3.2 Optimization

In order to solve the optimization of problem (6), we introduce two auxiliary variable J and T into the objective function. Then (6) can be represented as the following equivalent form:

$$ \mathop {\hbox{min} }\limits_{Z,E,J,T} \parallel J\parallel_{* } + \lambda \parallel E\parallel_{2,1 } \;{\text{s}} . {\text{t}} .\;{\text{X}} = {\text{XT}} + {\text{E}},\quad {\text{Z}} = {\text{J}},{\text{M}} \odot {\text{Z}} = {\text{T}} . $$
(8)

The augmented Lagrangian function of (8) is:

$$ \begin{aligned} & {\mathcal{L}}\left( {Z,J,T,E,Y_{a} ,Y_{b} ,Y_{c} } \right) = \parallel J\parallel_{ * } + \lambda \parallel E\parallel_{2,1 } + < Y_{a} ,X - XT - E > + < Y_{b} ,Z - J > \\ & \quad + < Y_{c} ,{\text{M}} \odot {\text{Z}} - {\text{T}} > + \frac{\mu }{2}\left( {\left\| {X - XT - E} \right\|_{F}^{2} + \left\| {Z - J} \right\|_{F}^{2} + \left\| {{\text{M}} \odot {\text{Z}} - {\text{T}}} \right\|_{F}^{2} } \right), \\ \end{aligned} $$
(9)

where \( Y_{a} ,Y_{b} \) and \( Y_{c} \) are Lagrange multipliers and \( \mu > 0 \) is a parameter. We can use an iterative approach to optimize the above unknown variables.

3.2.1 Update J with Other Variables Fixed

Suppose \( J^{t} \) and other variables have been obtained in the t-th iteration, ignore the irreverent terms of \( J \) in Eq. (9), we have:

$$ \begin{aligned} J^{t + 1} & = argmin_{{J^{t} }} \left\| {J^{t} } \right\|_{*} + < Y_{b}^{t} ,Z^{t} - J^{t} > + \frac{{\mu^{t} }}{2}\left\| {Z^{t} - J^{t} } \right\|_{F}^{2} \\ & = argmin_{{J^{t} }} \frac{1}{{\mu^{t} }}\left\| {J^{t} } \right\|_{*} + \frac{1}{2}\left\| {J^{t} - \left( {Z^{t} + \frac{{Y_{b}^{t} }}{{\mu^{t} }}} \right)} \right\|_{F}^{2} \\ \end{aligned} $$
(10)

then \( J^{t + 1} = U^{t} \varTheta_{{\frac{1}{{\mu^{t} }}}} \left( {S^{t} } \right)V_{t}^{T} \), where \( U^{t} S^{t} V_{t}^{T} \) is the SVD of matrix \( Z^{t} + \frac{{Y_{b}^{t} }}{{\mu^{t} }} \) and \( \Theta \) is a singular value thresholding operator [39].

3.2.2 Update T with Other Variables Fixed

Similar to the above method, we drop irrelevant items of T, then we have:

$$ \begin{aligned} {\text{T}}^{t + 1} & = argmin_{{{\text{T}}^{t} }} < Y_{a}^{t} ,X - XT^{t} - E^{t} > + < Y_{c}^{t} ,{\text{M}} \odot {\text{Z}}^{t} - T^{t} > \\ & \quad + \frac{{\mu^{t} }}{2}\left( {\left\| {X - XT^{t} - E^{t} } \right\|_{F}^{2} + \left\| {{\text{M}} \odot {\text{Z}}^{t} - T^{t} } \right\|_{F}^{2} } \right) \\ & = argmin_{{{\text{T}}^{t} }} \frac{1}{2}\left( {\left\| {X - XT^{t} - E^{t} + \frac{{Y_{a}^{t} }}{{\mu^{t} }}} \right\|_{F}^{2} + \left\| {{\text{M}} \odot {\text{Z}}^{t} - T^{t} + \frac{{Y_{c}^{t} }}{{\mu^{t} }}} \right\|_{F}^{2} } \right) \\ \end{aligned} $$
(11)

then \( {\text{T}}^{t + 1} = \left( {X^{T} X + I_{n} } \right)^{ - 1} \left( {X^{T} \left( {X - E^{t} + \frac{{Y_{a}^{t} }}{{\mu^{t} }}} \right) + \left( {{\text{M}} \odot {\text{Z}}^{t} + \frac{{Y_{c}^{t} }}{{\mu^{t} }}} \right)} \right) \). In is an n × n identity matrix.

3.2.3 Update Z with Other Fixed Variables

Based on the above computed \( J^{t + 1} \) and \( {\text{T}}^{t + 1} \), in the t-th iteration, we have

$$ \begin{aligned} {\text{Z}}^{t + 1} & = argmin_{{{\text{Z}}^{t} }} < Y_{b}^{t} ,Z^{t} - J^{t + 1} > + < Y_{c}^{t} ,{\text{M}} \odot {\text{Z}}^{t} - T^{t + 1} > \\ & \quad + \frac{{\mu^{t} }}{2}\left( {\left\| {{\text{Z}}^{t} - J^{t + 1} } \right\|_{F}^{2} + \left\| {{\text{M}} \odot {\text{Z}}^{t} - T^{t + 1} } \right\|_{F}^{2} } \right) \\ & = argmin_{{{\text{Z}}^{t} }} \frac{1}{2}\left( {\left\| {{\text{Z}}^{t} - J^{t + 1} + \frac{{Y_{b}^{t} }}{{\mu^{t} }}} \right\|_{F}^{2} + \left\| {{\text{M}} \odot {\text{Z}}^{t} - T^{t + 1} + \frac{{Y_{c}^{t} }}{{\mu^{t} }}} \right\|_{F}^{2} } \right), \\ \end{aligned} $$
(12)

then \( {\text{Z}}^{t + 1} = \left( {J^{t + 1} - \frac{{Y_{b}^{t} }}{{\mu^{t} }} + T^{t + 1} - \frac{{Y_{c}^{t} }}{{\mu^{t} }}} \right)./\left( {EE + M} \right) \), where EE is an n × n matrix with each element equals 1. Suppose \( C = A./B \), then \( \left[ C \right]_{ij} = \left[ A \right]_{ij} /\left[ B \right]_{ij} \).

3.2.4 Update E with Other Fixed Variables

Abandon the irrelevant terms in (9) w.r.t E, then we could have:

$$ \begin{aligned} E^{t + 1} & = argmin_{{{\text{E}}^{t} }} \lambda \parallel E^{t} \parallel_{2,1 } + < Y_{a}^{t} ,X - XT^{t + 1} - E^{t} > + \frac{{\mu^{t} }}{2}\left\| {X - XT^{t + 1} - E^{t} } \right\|_{F}^{2} \\ & = argmin_{{{\text{E}}^{t} }} \frac{\lambda }{\mu }\parallel E^{t} \parallel_{2,1 } + \frac{1}{2}\left\| {X - XT^{t} - E^{t} + \frac{{Y_{a}^{t} }}{{\mu^{t} }}} \right\|_{F}^{2} , \\ \end{aligned} $$
(13)

Finally, the update schemes of for the Lagrange multipliers \( Y_{a} ,Y_{b} \) and \( Y_{c} \) and parameter μ are summarized as follows:

$$ \begin{aligned} & Y_{a}^{t + 1} = Y_{a}^{t} + \mu^{t} \left( {X - XT^{t + 1} - E^{t + 1} } \right), \\ & Y_{b}^{t + 1} = Y_{b}^{t} + \mu^{t} \left( {{\text{Z}}^{t + 1} - J^{t + 1} } \right), \\ & Y_{c}^{t + 1} = Y_{c}^{t} + \mu^{t} \left( {{\text{M}} \odot {\text{Z}}^{t + 1} - T^{t + 1} } \right), \\ & \mu^{t + 1} = \hbox{min} \left( {\mu_{max} ,\rho \mu^{t} } \right), \\ \end{aligned} $$
(14)

where \( \mu_{max} \) and ρ are two adjustable parameters.

3.3 Algorithm

According to the descriptions above, ISLRR algorithm can be summarized as follows

figure a

4 Further Discussion

4.1 The Relationship Between SCLRR and ISLRR

Comparing SCLRR with our ISLRR, it can be found that:

  1. 1.

    In order to improve the applications of LRR on data sets with disjoint subspaces, both SCLRR and ISLRR introduce the weighted matrix M to the coefficient matrix Z for enhancing its structure information, as is to say, they both have the same weighted coefficient matrix \( {\text{M}} \odot {\text{Z}} \).

  2. 2.

    In this article, the way to define the weighted matrix M is same as that used in [34], namely \( \left[ M \right]_{ij} = 1 - { \exp }\left( { - \frac{{1 - \left| {x_{i}^{*T} x_{j}^{*} } \right|}}{\sigma }} \right) \). In the future works, we can also take more appropriate ways to get a better M.

  3. 3.

    According to Eq. (4), SCLRR introduces the weighted matrix \( {\text{M}} \odot {\text{Z}} = \sum\nolimits_{i,j} {M_{ij} \left| Z \right|_{i,j} } \) as the sparse constraint item into the objective function of LRR for balancing the nuclear norm to get a better low rank solution. But ISLRR treats the weighed matrix \( {\text{M}} \odot {\text{Z}} = \sum\nolimits_{i,j} {M_{ij} \left| Z \right|_{i,j} } \) as a new dictionary for representing data samples more accurately. Therefore, we can also get a more suitable structural low rank matrix for subspace segmentation.

4.2 Algorithm Convergence

In Sect. 3.2, ISLRR uses ADM to get the optimum solution, whose convergence has been validated when the number of variables is less than or equal to two [40]. Since there are four variables: Z, J, T and E in Algorithm 1, it is difficult to give the theoretical explanation to illustrate the convergence of algorithm. Fortunately, Eckstein et al. [41] provided two sufficient conditions to guarantee the convergence of the algorithms: one is that the dictionary matrix in the equality constraints should be of full column rank; the other is that the optimal solution of each variable is monotonically decreasing in each iteration step. According to [41], the dataset Z is a full column rank matrix, since the row elements of the weight matrix M are not all zero, it is easy to verify that the dictionary matrix \( {\text{M}} \odot {\text{Z}} \) is of full column rank. However, the monotonically decreasing condition cannot be proved directly, while the convexity of Lagrange’s function ensures its effectiveness to a certain extent [41]. In a word, we expect Algorithm 1 to be convergent and use the upper bounded μ to guarantee its theoretical convergence in the alternating direction method, but its strict proof of convergence is worth further discussion.

5 Experiments

In this section, we evaluate the performance of ISLRR for subspace segmentation problems. The close related algorithms including SSC [15], LRR [7, 18], and SCLRR [34] are used for comparisons. The frequently used databases such as Hopkins 155 [42], Fashion_MNIST [37], two face image databases Extended Yale B [43] and AR [44] are used in our experiments.

The computer configurations for these experiments are 3.20 GHz Inter (R) Core (TM) i5-6500 CPU with 4 GB machine memory, 500G hard disk memory, 64 bit windows7 professional operating system, and R2014a version Matlab.

These four databases are described briefly as follows:

Hopkins 155 motion segmentation data set contains 155 sequences, which includes 104 checkerboard sequences, 38 traffic sequences, and 13 others. Each of sequence is an independent clustering task, so there are altogether 156 clustering tasks. The sample images from Hopkins 155 database are given in Fig. 2.

Fig. 2
figure 2

Sample images of Hopkins 155 motion database. a 1R2RC, b arm

Fashion-MNIST data set consists of 60,000 train examples and 10,000 test examples. They are divided into 10 categories: t-shirt, trouser, pullover, dress, coat, sandal, shirt, sneaker, bag, and ankle boo. We select the first 50 pictures of each class’ train data set, and resize them to 20 × 20 pixels in our experiments. The samples are given in Fig. 3a.

Fig. 3
figure 3

Static sample images from three image databases. a Sample images of Fashion-MNIST database, b sample images of the extend Yale B database and c sample images of AR database

Extended Yale B face database consists of 38 objects with a total of 2414 images, each of which has 9 different gestures with 64 Gray scales. Moreover, these photos were taken in different light and facial expressions. In our experiment, each picture is resized to 32 × 32 pixels. The samples are given in Fig. 3b.

AR face database contains more than 4000 face images of 126 objects. Each object has 26 pictures in two modes, therefore each mode has 13 pictures. These images are taken in different views and lights. In our experiment, we select altogether 2000 photos from the first 120 people and adjust each image to 50 × 40 pixels. The image samples are shown in Fig. 3c.

5.1 Comparisons with LRR

In this subsection, we carry the experiments on three static image data sets to observe the relationship between LRR and ISLRR. The first four categories of each data set are selected for this experiment and parameter λ in both algorithms are set to 1. In Fig. 4, \( Z_{LRR} \) is the low rank coefficient matrix of LRR, \( Z_{ISLRR} \) is the low rank constraint matrix of ISLRR. M is the structural features matrix of data samples.

Fig. 4
figure 4

The reconstruction coefficient matrices obtained by LRR and SLRR. a Extend Yale B, b AR and c Fashion-MINST

From Fig. 4, we can see that: (1) \( Z_{ISLRR} \) is a dialog matrix, as is to say, \( Z_{ISLRR} \) can be a better matrix to design an affinity graph for subspace segmentation; (2) the low rank matrix \( Z_{ISLRR} \) is different from \( Z_{LRR} \) of LRR, and it is also not equal to \( Z_{LRR} ./M \). This means that \( Z_{ISLRR} \) could not be directly computed as \( Z_{LRR} ./M \). Therefore, it is reliable to introduce the structure information of data sets into LRR by following the methodology of ISLRR.

5.2 Sensitivity Analysis

In the first section, we took a simple test to observe the sensitivity of SCLRR to the extra parameter β and claimed that the performance of SCLRR varies drastically with β. In this section, a series of experiments are conducted to observe the sensitivity of ISLRR and SCLRR to the corresponding parameters. For Hopkins 155, we choose 1R2RC sequence for this experiment, and the range of parameter β is set as [0.001, 50] while λ is chosen in [0.0001, 20]. For the three static image databases, we select the first three classes of each dataset and set the intervals of λ and β as [0.001, 50]. In the experiments on SCLRR, we fix one parameter at their corresponding optimal point to observe the accuracy along with the change of another parameter. The experiments results are shown in Fig. 5.

Fig. 5
figure 5

The performaces of SCLRR and ISLRR varied with the corresponding parameters. a 1R2RC, b the Extended Yale B, c AR and d Fashion-MINIST

From Fig. 5, we can find that: (1) parameter β will affect the final results of SCLRR directly even if the parameter λ is fixed in the corresponding optimal value; (2) in most cases, the sensitivity of ISLRR to the parameter λ is similar to that of SCLRR when β is fixed in its optimal value. But in general, it is much easier to adjust one parameter in ISLRR than to adjust two in SCLRR.

5.3 Comparisons with Related Algorithms

To further verify the effectiveness of ISLRR, ISLRR and other related algorithms are evaluated on the four databases for comparisons.

5.3.1 Experiments on Motion Sequences

For Hopkins 155 data set, these motion sequences can be regarded as different categories, so we can consider such problems as 120 two-class segmentation problems and 35 three-class segmentation problems. The segmentation results of related algorithms including mean values (Ave.), standard deviations (Std.), and maxima errors (Max.) are reported in Table 1 (The optimal values of different criteria are emphasized in bold).

Table 1 Segmentation error (%) of different algorithms on Hopkins 155 data set

From the Table 1, we can see that ISLRR achieved better results in two motion sequences than those in three motion sequences. Compared with the related algorithms, ISLRR can reach more promising results.

5.3.2 Experiments on Static Image Datasets

In this subsection, we first find the highest accuracies of the four algorithms with the corresponding optimal parameters on three static image databases, and record the running times of each algorithm in each database. The experimental results are shown in Table 2 (The optimal values of different criteria are emphasized in bold).

Table 2 The experiment results of different algorithms on three image data sets

From Table 2, we see that: (1) ISLRR and SCLRR spend more time on three image data sets for better results; (2) ISLRR can obtain the highest segmentation accuracies on all the three databases.

Next, we compare the performances of related algorithms on each data set with different number of classes. In those experiments, we choose q (q is between 3 and the total number classes of data samples) classes data samples to compute their corresponding accuracies. The range of parameter λ is set as [0.001, 20] and parameter β is [0.0001, 50], and the best results are recorded in Fig. 6.

Fig. 6
figure 6

The segmentation accuracies obtained by evaluated algorithms versus the number of classes on different databases. a Experimental results on Fashion-MINIST database, b experimental results on the Extended Yale B database and c experimental results on AR database

From Fig. 6, we can see that: (1) with the increment of segmentation data categories, the segmentation accuracies of these algorithms decline; (2) ISLRR is superior to other algorithms in most cases, especially on Yale B data set. The experimental results show that ISLRR is an effective algorithm in dealing with the segmentation problems of static images.

6 Conclusions

In this paper, a new subspace segmentation algorithm named ISLRR is proposed. In ISLRR, we introduce the structure information of data sets into the equality constraint term of LRR. Comparing with SCLRR, it doesn’t bring any extra parameter, thus it avoids the difficulty for adjusting two parameters. To confirm the algorithm’s effectiveness, three close related algorithms are also evaluated on four image data sets. Experimental results show that ISLRR algorithm is simply and effective for handling subspace segmentation problems.