1 Introduction

The rapid growth in the use of Internet applications and the great concern of security require reliable and automatic personal identification. Traditional knowledge-based and token-based automatic personal identification schemes have forgotten or stolen limitations. Biometrics is an automated method of recognizing an individual based on a physiological or behavioral characteristic, such as face, fingerprint, iris, palmprint etc. Biometric recognition technologies are becoming the foundation of an extensive array of highly secure identification and personal verification solutions. Many large-scale human identity management projects such as US-visit program, EU biometric passport and Hong Kong identity card involve biometrics. And it is also applied to many other applications including access control, computer login and financial transactions authentication. Among different biometric authentications, palmprint recognition is one of the most promising approaches since it has several special advantages such as stable line features, rich texture features, low-resolution imaging, low-cost capturing devices, easy self positioning, and user-friendly interface [27].

Palmprint recognition has been investigated over the past several years. During this period, many different problems related to palmprint recognition have been addressed. Researchers have focused on developing accurate verification algorithms. Various feature extraction and matching algorithms have been proposed to improve the speed of implement and recognition rate [13]. For example, in early times, Lu et al. [18] and Wu et al. [24] proposed two methods based on Principal Components Analysis (PCA) and Linear Discriminant Analysis (LDA), respectively. Ekinci et al, proposed Kernel PCA to improve the palmprint recognition performance and proposed a palmprint recognition approach integrating the Gabor wavelet representation and kernel PCA [4, 5]. However, because of the illumination influence, the recognition performance of appearance based approaches mentioned above is not very satisfying in some aspects. Inspired by the face recognition algorithms, the Gabor based appearance algorithms are proposed on palmprint biometric traits [5, 25]. However, traditional Gabor-based image representation it has got two important drawbacks [20]. First, the computation burden is very heavy. Second, a lot of memory is needed to store Gabor features. And what is important, Gabor-based image representation is not robust to varying illumination. There are usuallythree principal lines in a palmprint: the heart line, the head line, and the life line. The Gabor-based image representation can’t represent the lines very well [10]. Other wavelets also are employed to represent the biometric traits, such as DT-CWT. It cannot represent the palmprint images well [15] though it has better performance than of Gabor based on face images [6].

Orientation based approaches are deemed to have the best performance in palmprint recognition field, because orientation in palmprint contains more discriminative principal line information than other features. A.Kong and D.Zhang were the first authors who investigated the orientation information of the palm lines for palmprint verification and extracted the orientations of the palm lines by using the Winner-take-all Rule, which was defined as Competitive Code (CompC) [12]. Inspired by CompC, W.Jia used a modified FRAT to extract every orientations of palmprint pixels which was defined as Robust Line Orientation Code (RLOC) [11]. And also designed a matching algorithm based on pixel-to-area comparison, which improve the palmprint performance and reduce the Equal Error Rate(EER). Yue use modified FCM based clustering method to find the orientations of Gabor filters in CompC [26]. However, by using only one dominant orientation to represent a local region, it may lose some valuable information because there are cross lines in the palmprint. To better describe the local orientation features and be more robust to image rotation, Z.Guo proposed binary orientation co-occurrence vector (BOCV) to represent multiple orientations for a local region [7]. Compared with RLOC, the EER of BOCV was reduced significantly. By directly using multiple orientation features, H.Li used a the matching score level fusion strategy to fuse of different directional Palm-Codes to obtain the high performance [17]. Those above palmprint recognition algorithms imply that directional representations have certain robustness and can represent the main characters of palmprint.

To improve the implemented efficiently and robustness in the palmprint recognition, we propose novel palmprint recognition approach using directional representations and compressed sensing in this paper. The novel proposed approach can effectively overcome the Gabor-based shortcoming and can significantly improve the performance of appearance based approaches. Based on the principle of structural risk minimization, Support Vector Machines (SVMs) are a set of supervised learning methods for linear classification and regression of data [2]. SVM classifier has good performance in biometric recognition. Compared with SVM and other classification algorithms, perhaps, the simplest classification scheme is a nearest neighbor (NN) classifier to distinguish different biometric traits using Subspace-based features [9]. Under this classifier, an image in the test set is recognized (classified) by assigning to it the label of the closest point in the learning set, where distances are measured in the image space. However, it does not work well under varying lighting conditions. A sparse representation provides deep insights into feature classification. Based on a sparse representation, J. Wright et al propose a general classification algorithm [23]. Usually, a sparse representation computed by l1-minimization [21] and has proven its superior performance on pattern recognition [22, 28].

The rest of this paper is organized in the following. In Section 2, the directional representations of palmprint images using multiple anisotropy filters will be proposed. Feature extraction and dimension reduction using PCA and classification using compressed sensing will be presented in Section 3. Experimental results on PolyU Palmprint Database are given in Section 4. Finally, conclusions are made in Section 5.

2 Image representations using directional representations

Based on physiological constraints and the wavelet theory, a family of 2D Gabor wavelets which model the receptive fields of the simple cells in the brain’s primary visual cortex has been derived in [14]. And Gabor is extensively employed to extract biometric image features and has improved its performance in the biometric trait recognition system for its characteristics to those of human visual system. From some literatures, it can be seen that different representations of palmprint can be adopted for Subspace-based approaches. These representation algorithms include Gabor [4], wavelet transform [5], dual-tree complex wavelet transform [15] et,al. Generally speaking, the commonly used strategy is applying the corresponding filter to convolute with the original palmprint image. However, these mentioned above still have some drawbacks. However, the Computation burden of Gabor is very heavy due to the filtering of the biometric image with a bank of Gabor filtering at many scales and orientations.

An image can be modeled as a piecewise smooth 2D signal with singularities. These properties need anisotropic refinement representations. However, the traditional Gabor isotropic refinement can’t efficiently represent the palmprint structure. The Anisotropic Filter (AF) is initially used in building over-complete dictionary to obtain sparse representation by the idea of efficiently approximating contour-like singularities in 2-D images. The AF is a smooth low resolution function in the direction of the contour, and behaves like a wavelet in the orthogonal (singular) direction. That is, the AF is built on Gaussian functions along one direction, and on second derivative of Gaussian functions in the orthogonal direction. The structure of AF is very good at capturing the orientation of palmprint image [16]. The AF has the following general form

$$ G\left( {u,v} \right) = \left( {4{u^2} - 2} \right)\exp \left( { - \left( {{u^2} + {v^2}} \right)} \right) $$
(1)

where (u,v) is, in this case, the plane coordinate and can be obtained in the following way.

$$ \left[ {\matrix{ u \hfill \\ v \hfill \\ }<!end array> } \right] = \left[ {\matrix{ {{{1} \left/ {\alpha } \right.}} \hfill &{0} \hfill \\ 0 \hfill &{{{1} \left/ {\beta } \right.}} \hfill \\ }<!end array> } \right]\left[ {\matrix{ {\cos \theta } \hfill &{sin\theta } \hfill \\ { - sin\theta } \hfill &{\cos \theta } \hfill \\ }<!end array> } \right]\left[ {\matrix{ {x - {x_0}} \hfill \\ {y - {y_0}} \hfill \\ }<!end array> } \right] $$
(2)

Where [x 0, y 0]is the center of the anisotropic filter, the rotation θ is to locally orientation the filter along palm contours and α and β are to adapt to contour type. Parameters α, β are the anisotropic scaling factors along X-axis and Y-axis. These parameters are adjusted to adapt to contour smoothness, which is helpful for extraction local orientation of palmprint images.

A line in palmprint contains various information including type, width, position and orientation. These properties can used to compare two palmprint images by discriminating different line features. All the lines in palmprints belong to the negative lines (darker pixels in the center of the line) category. Palm lines do have a certain width for principal lines are wider(about 2-6 pixels). On top of types and width, line position is often considered as an important feature. It is worthwhile to note that lots of research results show that the orientation information is most discriminative features. And various of filters are designed to extract the palmprint orientation [7, 11, 12, 17, 26].

The choice of the Gaussian envelope is motivated by the optimal joint spatial and frequency localization of this kernel and by the presence of second derivative-like filtering in the early stages of the human visual system. It is also motivated by the presence of second derivative-like filtering in the early stages of the human visual system. Usually, β > α is set to better obtain the line orientation of palmprint images. A 3D visualization of an AF can be seen in Fig. 1. The orientation information of a local region is calculated by the competitive rule( winner-take-all rule):

$$ j = \arg \mathop{{\min }}\limits_p \iint {I\left( {x,y} \right)G\left( {x,y,\alpha, \beta, {\theta_p}} \right)dxdy} $$
(3)
Fig. 1
figure 1

a Appearance of anisotropic filter(3D);(b) Sample anisotropic filter with a rotation of 2*pi/3 radians and scales of 3 and 9

Where I represent a preprocessed palmprint image and j is called the winning index. The orientations of the twelve filters are employed, in particular, θ p is /12, where\( p \in \left\{ {0,1,2,\ldots, 11} \right\} \). Based on the competitive rule is introduced (3), we can extract the orientation information on a palm line, which is the palmprint intrinsic features and very important to distinguish different hands. The winning indexes are integer representations of the local orientations of palmprints.

Figure 2 shows three palmprint images and their directional representations. Among them, Fig. 2a are from the same palm, but were captured in different illumination conditions. Although the illumination conditions changed drastically, their directional representations Fig. 2b are still very similar. From this example, it can be concluded that the directional representations is also robust for the change of illumination.

Fig. 2
figure 2

Examples of palmprint directional representations (a) Preprocessed images. (b) directional representations

Figure 3 shows three palmprint images and their directional representations. Among them, Fig. 3a comes from Fig. 2a which adds Gaussian white noise(noise level is 25) . Although the noise made the image quality and understandability worse, their directional representations are still very similar. From this example, it can be concluded that the directional representations is also robust for the noise corruption. It also indicates that it has robust nessto different capturing environments and devices.

Fig. 3
figure 3

Plmprint images and their directional representation and directional representation s are robust for noise corruption(a) Preprocessed images. (b) directional representations

In a word, different orientations are employed to calculate the sum of product of pixel values in local area and anisotropic filters(also it is the same as the “convention” operation). And the orientation which is corresponding to the smallest amplitude response (via the competitive rule) is regarded as the pixel orientation. All the pixel orientations are called the directional representations. Also the direction representations can well reflect the essence feature of palmprint images properly. The Sum operation can be implemented very fast and the number of anisotropic filter banks is less than Gabor.

3 Feature extraction and classification methods

PCA, which it to reduce the large dimensionality of the data space to the smaller intrinsic dimensionality of feature space, has been widely used as linear feature extraction in computer vision [1]. It computes the basis of a space which is a space which is represented by its training vectors yields projection directions that maximize the total scatter across all classes. These basis vectors, actually eigenvectors, computed by PCA are in the direction of the largest variance of the training vectors. PCA is one of the most successful techniques that have been used in image recognition. PCA computes the basis of a space which is a space which is represented by its training vectors. These basis vectors, actually eigenvectors, computed by PCA are in the direction of the largest variance of the training vectors. PCA techniques, also known as Karhunen-Loeve methods, choose a dimensionality reducing linear projection that maximizes the scatter of all projected samples.

More formally, let us consider a set of N sample images {x 1,x 2,…,x N } taking values in an n-dimensional image space, and assume that each image belongs to one of c classes {X 1,X 2,…,X c }. Let us also consider a linear transformation mapping the original n-dimensional image space into m-dimensional feature space, where m < n. The new feature vectors \( {y_k} \in {\mathbb{R}^m} \) are defined by the following linear transformation:

$$ {y_k} = {W^T}{x_k}\quad k = 1,2,\ldots N $$
(4)

Where \( W \in {\mathbb{R}^{{n \times m}}} \) is a matrix with orthonormal column.

If the total scatter matrix S T is defined as

$$ {S_T} = \sum\limits_{{k = 1}}^N {\left( {{x_k} - \mu } \right){{\left( {{x_k} - \mu } \right)}^T}} $$
(5)

Where N is the number of sample images, and \( \mu \in {\mathbb{R}^n} \)is the mean image of all samples, then after applying the linear transformation W T, the scatter of the transformed feature vectors {y 1,y 2,…,y N } is W T S T W. In PCA, the projection W opt is chosen to maximize the determinant of the total scatter of the projected samples, i.e.,

$$ {W_{{opt}}} = \mathop{{\arg \max }}\limits_W \left| {{W^T}{S_T}W} \right| = \left[ {{w_1},{w_2},\ldots {w_m}} \right] $$
(6)

Where\( \left\{ {{w_i}\left| {i = 1,2\ldots, m} \right.} \right\} \) is the set of n-dimensional eigenvectors of S T corresponding to the m largest eigenvalues. These eigenvectors have the same dimension as the original images.

Sparse representation, which indicates that account for most or all information of a signal with a linear combination of a small number of elementary signals, has proven to be an extremely powerful tool for representing natural images. Finding a representation with a small number of significant coefficients can be solved as the following optimizing problem:

$$ {\widehat{x}_0} = \arg \min {\left\| x \right\|_0}\;{\mathrm{subject}}\;{\mathrm{to}}\;Dx = y $$
(7)

Where \( {\left\| . \right\|_0} \) denotes the l 0-norm, which counts the number of nonzero entries in a vector. Seeking the sparsest solution to Dx = y is a NP problem. The theory of sparse representation and compressed sensing reveals that if the solution x0 sought is sparse enough, the solution of the l 0-minimization problem is equal to the solution to the l 1-minimization problem. The optimizing problem can be solved as the following:

$$ {\widehat{x}_1} = \arg \min {\left\| x \right\|_1}\;{\mathrm{subject}}\;{\mathrm{to}}\;Dx = y $$
(8)

Given sufficient training palmprint samples of the i-th object hand class,\( {D_i} = \left[ {{d_{{i,1}}},{d_{{i,2}}},\ldots, {d_{{i,{n_{{_i}}}}}}} \right] \in {\mathbb{R}^{{m \times {n_i}}}} \), a test palmprint sample \( y \in {\mathbb{R}^m} \) from the same hand will approximately lie in the linear span of the training palmprint samples associated with object i. y = D i x i for some coefficient vector \( {x_i} \in {\mathbb{R}^{{{n_i}}}} \) .

Therefore, given a new test palmprint sample feature y from one of the classes in the training feature set, we first compute its sparse representation via basis pursuit. Usually, the small nonzero entries in the estimation associated with the columns of D from a single object class I, and can easily assign the test palmprint feature y to that class. Based on the prior sparse representation of palmprint images, one can treat the test feature can be treated as a linear combination of all training features of each object. And, one can identify the right class from multiple possible classes. It can be computed as follows: For each class i, let \( {\lambda_i}:{\mathbb{R}^n} \to {\mathbb{R}^n} \)be the characteristic function which selects the coefficients associated with the i-th class, one can obtain the approximate representation \( {\widehat{y}_i} = D{\lambda_i}\left( {{{\widehat{x}}_1}} \right) \) for the given test sample y. We then classify y based on the approximations by assigning it to the object class that minimizes the residual between y and \( {\widehat{y}_i}:\min {r_i}(y) = \left\| {y - D{\lambda_i}\left( {{{\widehat{x}}_1}} \right)} \right\| \)[23].

4 Experimental results and analysis

4.1 Experimental settings and methods

In this section, we present experiments on PolyU publicly available databases for palmprint recognition, which demonstrate the efficacy of the proposed directional representations and validate the claims of the CS. In PolyU Palmprint Database, there are 600 gray scale images captured from 100 different palms by a CCD-based device [8]. Six samples from each palm are collected in two sessions: the first three samples were captured in the first session and the other three were captured in the second session. The average time interval between these two sessions was 2 months. The size of all the images in the database was 384 × 284 with a resolution of 75dpi. By using the similar preprocessing approach described in literature [27], a central part (128 × 128) of each image is extracted for further processing. Figure 4 illustrates a ROI image cropped from the original palmprint image. The first three samples of each palm are selected for training, and the remaining three samples are used for testing. The feature vector of the input palmprint is matched against all the stored templates and the most similar one is obtained as the matching result.

Fig. 4
figure 4

a The determination of ROI. (b) A cropped ROI image of the palmprint image in (a)

The experimental results have been generated on a PC with an Intel Pentium 2 processor (2.66 GHz) and 3 GB RAM configured with Microsoft Windows 7 professional operating system and Matlab 7.10.0(R2010a). In the implementation of Gabor filters, the parameters are set as kmax=π/2, σ=2π, \( f = \sqrt {2} \), u={0,1,…11}, v={0,1,2}. For the anisotropic filter banks, the orientations of the twelve filters are employed, in particular, θ p is /12, where\( p \in \left\{ {0,1,2,\ldots, 11} \right\} \).and the scales of x-axis and y-axis are 7 and 21, the size of anisotropic filter is 23 × 23.

The Nearest Neighbor (NN) classifier employed minimum Euclidean distancebetween the query feature vector and all the prototype training data. The Support Vector Machine (SVM) classifier employed linear kernel as it gave us the best results [19]. A highly efficient algorithm suitable for large scale applications, known as the Spectral Projected Gradient (SPGL1) algorithm [3, 21], is employed to solve the BP problems. The performance of a verification method is often measured by the FAR(false accept rate), FRR (false reject rate)and EER(equal error rate) and also ROC(receiver operating characteristic) curve is plotted. To demonstrate the effectiveness of the proposed approach, the following series of comparison experiments were carried out. Additionally, the speed for every part of the proposed algorithm is compared with the traditional algorithms in the last subsection.

4.2 Palmprint verification performance

Verification, also known as one against one matching, is defined as deciding if a person is whom he claims to be by examining his palmprint. To obtain the verification accuracy of our palmprint system, each of the palmprint images was matched with all of the palmprint images in the database. A matching is noted as a correct matching if two palmprint images are from the same palm. Six samples from each palm are collected in two sessions: the first three samples were captured in the first session and the other three were captured in the second session. The total match is 30,000. The failure to enroll rate is zero. The test palmprint images is 300, and each sample is match the three training samples with the same class, the mean value of the three matching scores is acted as the final score. Therefore, the number of comparisons that have correct matching is 300, and the rest are incorrect matchings. The feature dimension is 30 in the experiments.

Figure 5 depicts the corresponding Receiver Operating Characteristic (ROC) curve, which is a plot of genuine acceptance rate against false acceptance rate for all possible operating points. From Fig. 5a illustrated the CS classifier algorithms, we can see that the PCA approach can operate at about 52 % genuine acceptance rate(GAR) and a 1 × 10−2 % false acceptance rate(FAR). At the same FAR, the GAR of Gabor approach is about 34 % higher than that of PCA and the GAR of our system is about 42 % higher than that of PCA. From Fig. 5b illustrated the SVM classifier algorithms, we can see that the PCA approach can operate at about 81 % genuine acceptance rate(GAR) and a 0.2 % false acceptance rate(FAR). At the same FAR, the GAR of Gabor approach is about 11 % higher than that of PCA and the GAR of our system is about 14 % higher than that of PCA. From Fig. 5c illustrated the NN classifier algorithms, we can see that the PCA approach can operate at about 36 % genuine acceptance rate(GAR) and a 1 × 10−2 % false acceptance rate(FAR). At the same FAR, the GAR of Gabor approach is about 22 % higher than that of PCA and the GAR of our system is about 46 % higher than that of PCA. This results show that the proposed directional representation and CS approach is comparable with previous Gabor based palmprint approaches.

Fig. 5
figure 5

Verification test results, the receiver operator characteristic curves of different approaches

Results come from different approaches, such as spatial image representations, Gabor based approach and DR(directional representations) approach, and also different classification algorithms, such as NN, SVM and CS, are compared in Table 1. From Table 1, we can see that the results of our approach are better than that of other approaches. From the Fig. 5 ROC curves and Table 1 EERs, the proposed system obtains a better performance than the traditional Gabor approaches.

Table 1 EERs of different approaches(feature dimensions: 30)

4.3 Palmprint identification performance

Identification is defined as deciding whom a person is according to his/her palmprint. The feature vector of the input palmprint is matched against all the stored templates and the most similar one is obtained as the matching result. Following these schemes, we have calculated recognition rates with the dimensions ranging from 5 to 150. Figure 6 shows the recognition rates for this experiment. As we can see from this Fig. 6, the correct recognition rate increases with the increasing of the dimension of features, and it surpasses 90 % when the dimension equals with or exceeds 25. With the increasing of the feature dimension, most of the palmprint image has been extracted. Therefore, the recognition performance trends to stable when the dimension of feature is more than 45. In general, the compressed sensing(CS) classifier has superior performance than the linear SVM and NN.

Fig. 6
figure 6

Recognition Performance (a) Recognition performance of different approaches with NN classifier (b) Recognition performance of different approaches with linear SVM classifier. (c) Recognition performance of different approaches with CS classifier

In the lower dimension, such as 10, the feature cannot represent the intrinsic the palmprint image, therefore, the NN seems better. The Fig. 6 also suggests that the recognition rate of ours has better performance than all the other approaches under the same condition. For the feature dimension is lower than 45, the directional representations based approaches has better performance than Gabor methods. When the feature dimension is larger than 45, the performance of directional representation and Gabor are nearly the same. As the dimension of the feature space increases, most of the palmprint image has been extracted. Therefore, the recognition performance trends to stable.

As illustrated in Table 2, with 40 dimensional features, CS achieves a recognition rate between 89.3 % and 97.0 % and . One the other hand, the best recognition rates achieved by NN and linear SVM are 95.7 % and 95.3 %, respectively. In general, for the same dimensional feature, the CS classifier achieves maximum recognition rate of the three classifiers. For the same classifier, the directional representation achieves a better recognition performance.

Table 2 Recognition performance under different approaches with 40 dimensional features

4.4 Computational speed

The proposed algorithm consists of three parts: (1)directional representations (2)PCA- dimension reduction method(3)CS classifier. Compared with the traditional Gabor-based algorithm, the dimension of directional representation is the same with original palmprint image, that is 64 × 64(downsample strategy is employed in the implemented process), which is much smaller the 9600 dimension of Gabor–based. The lower dimension can speed up the PCA operation. The Sum operation can be implemented very fast and the number of anisotropic filter banks is less than Gabor. From above discussion, we can infer that the time of DR implemented is less that of Gabor. Table 3 illustrates the computing time of the proposed feature exaction approach and other approaches. Form the Table 3, the computational running time of the proposed directional representations for feature extraction is nearly half of the Gabor representational approaches. Of course, they are both longer than that of PCA approach.

Table 3 Running time under different feature exaction approaches with 40 dimensional features

Table 4 illustrates the computing time under different feature classifiers with 40 dimensional features. Form the Table 4, the computational running time of the CS classification is much longer than then and linear SVM approaches. In the libsvm toolbox(libsvm-mat-3.0-1) [19], the SVM algorithm is implemented using compiled VC++ and Matlab mixed programming, therefore, the speed of SVM is faster than that of NN in our experiments. From Table 3 and Table 4, we can get that the total computational running time of the proposed approach for feature extraction and classification is 80.61 s. It should be noted that in samples in the palmprint Database is 600, therefore, the mean time of one user is about 0.13, which can meet the real time applications.

Table 4 Running time under different feature classifiers with dimensional features

The performance of approaches based on directional representations is a little better than the Gabor-based. What is important, the running time of Gabor based palmprint recognition algorithms is 1.5 times of that Directional Representations based algorithms. As the dimension of the feature space increases, directional representation has the similar performance. However, the computation burden on the directional representations is less than that of the Gabor based approach.

5 Conclusions

In this paper, a novel image representation approach, named directional representations for expressing palmprint images efficiently, is proposed. Firstly, a new representation for appearance based approach using competitive rule on the multiple anisotropy filters is presented. Compared with the Gabor representation, the proposed directional representation contains stronger discriminative information with lower computational burden, and is insensitive to illumination changes. Then, subspace based approaches, such as PCA, is used to extract the palmprint features and reduce the dimension. Finally, a compressed sensing classification is employed to distinguish different palms from different hands. Experimental results show that the proposed algorithm have better performance with a real-time implementation speed.