1 Introduction

A biometric is used for automatic identification of an individual. Biometry technology is employed to acknowledge persons by using distinctive biological characteristics [27]. Biometric authentication is commonly applied the police investigation and forensic applications. In the last decade, the use of biometry is being hyperbolic day by day for human identification and verification. Biometry has benefits over standard authentication systems like a PIN and password based systems due to its individualism, permanence and the convenience of use. Numerous biometric traits (fingerprint, iris, palmprint, finger vein, finger knuckle, face, ear, hand pure mathematics etc.) are used for verification and identification. In a unimodal system, one biometric modality is employed for authentication, e.g., palm print, fingerprint, footprint, face, knuckle, voice, etc. However, unimodal biometry systems suffer from numerous inabilities to tolerate malformed information (noise, device, environmental noise, physical appearance), hence multiple biometric modalities can be used to improve the accuracy of a biometric [6].

An attack is the presence of a fabricated article or duplicate human attribute to a biometric system [22]. There are several attacks that can be imposed to compromise the security of a system. Ratha et al. [42] described these attacks and clubbed them into eight classes as shown in Fig. 1. In the proposed method attack 1 i.e. the presentation attack is considered.

Fig. 1
figure 1

Classification of attacks

Silicone, Playdoh and Gelatin are sensible materials to generate false distinctive finger impression on scanner [15, 46]. The sample can be captured of a fingerprint either from a supporting procedure or a non-helpful method. In an exceedingly useful procedure when the real user involves in the arrangement of the fake sample whereas in a non-agreeable method when the assaulter can catch the real one unknowingly. There are three types of spoofing attacks reported on the basis of which, an unauthorized person breaks any iris biometric system. These are photo attacks, contact-lens attacks and artificial eye attacks [14]. Numerous approaches are discussed in the literature to classify the fake and real person on the basis of contact-lens attack or photo attacks using texture features. Fingerprint and iris are widely used on the basis of the fact derived from Table 1.

Table 1 Various biometric modalities performance [39]

Fingerprint and iris images show different textural characteristics between real and fake one. Spoof images show different behavior if they will be fetched with different sensing element or they are made of different fabrication [10]. The proposed system focuses on two objectives:

  1. 1.

    Investigation of robust liveness detection method using fingerprint and iris modality;

  2. 2.

    Decision level fusion to further improve the performance using multimodal biometric.

The proposed method is based on textural properties of real and fake biometric. In the proposed method, the fake fingerprint classification system is performed based on two types of texture properties. The primary category of texture uses statistical features and the second category uses macro-feature (core and delta) for fingerprint classification. To the best of our knowledge, no state-of-the-art methods supported core and delta points for fingerprint liveness detection. In the proposed method, fingerprint micro and macro texture features are consolidated and generate a single feature vector. For iris image classification, feature descriptor known as Local Binary Hexagonal Extrema Pattern (LBHXEP) [4] is used to extract the texture feature of iris. This proposed descriptor works efficiently in both types of attack i.e. printed iris attack and contact lens attack.

In the proposed method, the performance is further improved by fusing the decision of both the modalities i.e. fingerprint and iris. Several methods such as sum-rule [43] and kernel-based [5] are available for biometric information fusion at matching level and decision level. The choice of decision level fusion is effective when the individual system makes a distinctive decision about the identity of an individual. In the proposed work, we have used the standard D-S theory for fusion of fingerprint and iris at decision level.

The proposed method is evaluated on a publicly ATVS and LivDet2011 datasets. The experimental results indicate 94.7% precision to arrange between the genuine and fake one for finger print. The decision level classification derives the prominent result using D-S theory when compared with sum- rule [28] and the product-based rule function [34] with 97.8% accuracy.

The rest of the paper is ordered as follows: Section 2 introduces the related work, Section 3 discusses the basic preliminaries used in the proposed system, Section 4 presents the proposed system, Section 5 presents the experimental results and comparison with state-of-the-art methods and Section 6 has conclusion of this paper.

1.1 Related work

In last few years many anti-spoofing or liveness detection algorithms have come to light which could classify between real and fake biometric sample. These algorithms can be classified into two classes: a) Hardware based and b) Software based. The classification is depicted in Fig. 2 [13].

Fig. 2
figure 2

Classification of anti spoofing methods

Hardware-based solutions would like physical science gadgets. They are terribly expensive and are much addicted in to setting and skin state of affairs each. When we integrate the biometric with device then inclined outcome is obtained [7, 45]. Different electronic gadgets present different characteristics of same input image. Thus considering all this, we will focus on software solution supported static technique.

Abhyankar and Schuckers [3] developed a detection technique that consisting of three parts: MTA, NRFA and fuzzy k-means classifier analysis. Classification rates of 97.3%, 96.5% and 92.3% were reportable for Optical, physical phenomenon DC and Electro-optical severally. The limitation of this work was that it fully was tested on restricted dataset and wishes the validity of technique against perspiration development across the population, also, not tested on completely different environmental conditions.

A technique for detection of pretend fingerprint that supports time-series photos given by Jia and Cai [29]. A fingerprint photo from time-series sequence is captured from a fingerprint scanner. The limitation of this technique was that it had been tested only restricted datasets and one device.

LBP (Local binary pattern), wavelet energy was used for feature extraction of fingerprint photos by Nikam and Agarwal [36]. Texture details are extracted by LBP and ridge frequencies and orientation details extracted from wavelet. Three classifiers were used for classification. Fingerprint databases consisting restricted dataset of gummy fingerprints.

In this strategy, Galbally et al. [16] have extracted 10 diverse quality measures as feature vector that will be utilized for description of fingerprint images. This technique does not consider any physiological changes. Nikam and Agrawal [37] gathered unique fingerprint images and investigate neighborhood and global information. Different classifiers KNN and neural system were likewise utilized for order.

A strategy for recognition of dynamic pores on distinctive finger impression photos for fake detection is planned by Memon et al. [33]. Strategy contains of high-pass separation, association separating and thresholding. The technique was tried on four edge esteems and constraint of this strategy that it does not decide ideal limit. Ridge -strength, Ridge-continuity and Ridge-clarity are extracted from an image proposed by Galbally et al. [17]. LDA (Linear Discriminate Analysis) classifier is used to classify between real or fake. A limitation is to find the best performing features subset and does not classify the samples based on other physiological measures.

Marasco and Sansone [32] conferred a method by combining sweat and morphology-based static highlights for distinctive finger impression aliveness recognition. A morphology-based methodology will separate amongst live and counterfeit footage, whereas sweat primarily based methodologies will derive dynamic information from a solitary image. The restriction of this work is that the highlights aren’t chosen naturally for varied sensors.

A local phase quantization (LPQ) method is proposed by Ghiani et al. [20]. LPQ can distinguish spectrum differences between a “live” fingerprint and a “fake” one. They adopted the rotation invariant extension of LPQ because surface of sensor can produce different fingerprint orientation. A feature vector of 25 image quality measures given by Galbally& Marcel [19]. In this system complexity is too high and inclusions of temporal information are required.

Abhishek and Yogi [2] projected a unique technique for distinctive the phony fingerprints utilizing gelatin, Play-Doh and polymer molds and then forth. The delivery of details during a distinctive mark is attention-grabbing and utilized by the bulk of the distinctive finger impression acceptance calculations. This method was tried on the fingerprint verification competition datasets and declared precision over 85%.

Low level gradient features and texture features from Gabor wavelet using dynamic score level integration proposed by Dubey, Goh and Thing [12]. They are not able to classify cooperative fake user and non-cooperative user with real user and fail to identify the macro features. Convolutional Neural Networks (CNNs) proposed by Nogueira et al. [38] for spoof fingerprint detection. They have used four different models: two CNNs pertained on natural images, classical local binary pattern approach and CNN with random weights. They did not consider different kind of datasets with different sensors.

Yuan, C. et al. [53] extracted features in light-weight of CNN and passed to SVM classifier. PCA is used to decrease the dimension when every convolutional or pooling activity done.

Fake iris detection system is strictly based on types of attack. Photo attack will behave differently as contact lens attack. Different approaches have been proposed by researchers by considering these types of attack. Galbally et al. [18] proposed a method with quality related features and applicable only on high quality iris printed images.

Galbally et al. [19] extracted feature vector of iris image from quality measures. LDA and QDA classifier are used to classify between real and fake. This method is applicable only for photo attack. Raghavendra and Busch [41] use M-BSIF for feature extraction. It is validating only on photo attack and decision level fusion is not tested. Zernike moments are calculated by Kohli et al. [31] across the iris image and LBPV is used for textural changes. It emphasizes on textural changes, not sufficient for structural changes. Hu et al. [26] constructed local regional parts and proposed a local descriptor for low level features. They failed to construct strong regional features and overlap to each other.

Bhogal et al. [9] calculated six IQM to discriminate real from fake in presentation attacks for classification of iris. Abate et al. [1] used Hough transform for iris localization and calculated 6 metrics for 32 sections, 4 rows and 8 columns. It is tested on UPOL, UBIRIS and MICHE dataset. This approach is not suitable for fake iris detection. Wang et al. [48] proposed quality assessment method for screen content images. A local information content weighting scheme issued to derive the optimal perceptual weights. This method is suitable for screen content images but not for real or finger print images and as well as not for real and fake iris images.

When the different biometric modalities are reterived from different detector, then it may be from dissimilar scale or with dissimilar distribution. After multiple decisions from different platform, we have to take a final perception of the things by considering all separate decisions. Poh et al. [40] considered fingerprint and face images at decision level fusion. The outputs have been combined using the naive Bayes principal.

Vora et al. [47] focused on statistical analysis of different kernel methods. They used SVM based fusion of match scores for face and fingerprint biometric trait. Efficient multimodal biometric systems for face and iris fusion are proposed by Gomez-Barrero et al. [21]. They compute the similarity scores obtained by the face and iris and normalized scores are fused with a sum.

Dempster–Shafer theory used at score-level for multi-biometric detection proposed by Nguyen et al. [35] with uncertainty factors. D-S theory is implemented on score level and used for authentication. Wild et al. [49] proposed a method for fingerprint and face fusion in the presence of attacks. 1-median filtering is used at score level. Human ear and palm print are used in biometric recognition by Hezil and Boukrouche [25]. They used serial fusion method for features concatenation.

In the proposed system, statistical texture features and spatial analysis of the fingerprint pattern is utilized for fake or real classification. The approach is further improved by fusing iris modality with the fingerprint modality. In this paper, we are using decision level fusion using D-S theory and find the prominent result, when is implemented on ATVS, LivDet2011 and IIIT-D CLI dataset.

1.2 Preliminaries

In this section we will discuss the basic preliminaries used in the proposed system. Statistical texture feature i.e. micro features are derived from Gray Level Co-occurrence Matrix (GLCM) and are proposed by Haralick [8]. Haralick statistical features are good for micro textures; however, they show average performance on macro textures based on spatial analysis. Amadasun and King [24] proposed five macro texture features derived from Neighborhood Gray Tone Difference Matrix (NGTDM). D-S theory [44] is used for fusion of independent multi decision at decision level. On the basis of the final decision, proposed liveness detection approach decides the person is authentic or not. If the person is authentic then user will be accepted, otherwise rejected.

1.3 Neighborhood gray-tone difference matrix (NGTDM) [8]

NGTDM calculates the dissimilarity between a gray value and the average gray value of its neighbors within distance δ. The sum of total differences for gray level i, is stored in the matrix. Let Xgl be a set of segmented voxels and xgl(jx, jy, jz) ∈ Xgl be the gray level of a voxel at position (jx,jy,jz)then the average gray level of the neighborhood is:

$$ \overline{A_i}=\overline{A}\left({j}_x,{j}_y,{j}_z\right) $$
$$ =\frac{1}{w}\sum \limits_{k_x=-\delta}^{\delta}\sum \limits_{k_y=-\delta}^{\delta}\sum \limits_{k_z=-\delta}^{\delta }{x}_{gl}\left({j}_x={k}_x,{j}_y={k}_y,{j}_z={k}_z\right), $$
(1)

Where (kx, ky, kz) ≠ (0, 0, 0)and xgl(jx = kx, jy = ky, jz = kz) ∈ Xgland w is the number of voxels in the neighborhood that are also in Xgl.

1.4 Gray level co-occurrence matrix (GLCM) [24]

The grey-level co-occurrence matrix, one of the most known texture analysis method related to second order statistics. Each entry of the matrix will represent number of occurrence of one pixel with intensity i and another pixel with intensity jat a separation D = 1, 2, 3, ... toward the path θ = 00, 450, 900, 1350of an image I. In Equation 2, (L × M) is the set of pixels of the image ordered by row-column, # denotes number of elements in the set with intensity i and j. (×1, ×2) and (×3, ×4) are the ordered set of pixels.

$$ \left.\begin{array}{l}\begin{array}{l}\mathrm{X}\left(\mathrm{i},\mathrm{j},\mathrm{x},{0}^0\right)=\#\left\{\left(\left(\mathrm{x}1,\mathrm{x}2\right),\left(\mathrm{x}3,\mathrm{x}3\right)\right)\in \left(\mathrm{L}\times \mathrm{M}\right)\times \left(\mathrm{L}\times \mathrm{M}\right)\right.\\ {}\kern5em \left.\kern0.5em :\mathrm{x}1\hbox{--} \mathrm{x}3=0,\left|\mathrm{x}2\hbox{--} \mathrm{x}4\right|=\mathrm{D},\mathrm{I}\left(\mathrm{x}1,\mathrm{x}2\right)=\mathrm{i},\mathrm{I}\left(\mathrm{x}3,\mathrm{x}4\right)=\mathrm{j}\right\}\\ {}\mathrm{X}\left(\mathrm{i},\mathrm{j},\mathrm{x},{45}^0\right)=\#\left\{\left(\left(\mathrm{x}1,\mathrm{x}2\right),\left(\mathrm{x}3,\mathrm{x}4\right)\right)\in \left(\mathrm{L}\times \mathrm{M}\right)\times \left(\mathrm{L}\times \mathrm{M}\right)\right.\end{array}\\ {}\kern5.75em :\left(\mathrm{x}1\hbox{--} \mathrm{x}3=\mathrm{D},\mathrm{x}2\hbox{--} \mathrm{x}4=\mathrm{D}\right)\\ {}\begin{array}{l}\left.\kern6.5em \mathrm{or}\ \left(\mathrm{x}1\hbox{--} \mathrm{x}3=-\mathrm{D},\mathrm{x}2\hbox{--} \mathrm{x}4=\mathrm{D}\right),\mathrm{I}\left(\mathrm{x}1,\mathrm{x}2\right)=\mathrm{i},\mathrm{I}\left(\mathrm{x}3,\mathrm{x}4\right)=\mathrm{j}\right\}\\ {}\mathrm{X}\left(\mathrm{i},\mathrm{j},\mathrm{x},{90}^0\right)=\#\left\{\left(\left(\mathrm{x}1,\mathrm{x}2\right),\left(\mathrm{x}3,\mathrm{x}4\right)\right)\in \left(\mathrm{L}\times \mathrm{M}\right)\times \left(\mathrm{L}\times \mathrm{M}\right)\right.\\ {}\begin{array}{l}\left.\kern8.25em :\left|\mathrm{x}1\hbox{--} \mathrm{x}2\right|=\mathrm{D},\mathrm{x}3\hbox{--} \mathrm{x}4=0,\mathrm{I}\left(\mathrm{x}1,\mathrm{x}2\right)=\mathrm{i},\mathrm{I}\left(\mathrm{x}3,\mathrm{x}4\right)=\mathrm{j}\right\}\\ {}\mathrm{X}\left(\mathrm{i},\mathrm{j},\mathrm{x},{135}^0\right)=\#\left\{\left(\left(\mathrm{x}1,\mathrm{x}2\right),\left(\mathrm{x}3,\mathrm{x}4\right)\right)\in \left(\mathrm{L}\times \mathrm{M}\right)\times \left(\mathrm{L}\times \mathrm{M}\right)\right.\\ {}\begin{array}{l}\kern7em :\left(\mathrm{x}1\hbox{--} \mathrm{x}2=\mathrm{D},\mathrm{x}3\hbox{--} \mathrm{x}4=\mathrm{D}\right)\\ {}\left.\kern6.5em \mathrm{or}\ \left(\mathrm{x}1\hbox{--} \mathrm{x}3=-\mathrm{D},\mathrm{x}2\hbox{--} \mathrm{x}4=-\mathrm{D}\right),\mathrm{I}\left(\mathrm{x}1,\mathrm{x}2\right)=\mathrm{i},\mathrm{I}\left(\mathrm{x}3,\mathrm{x}4\right)=\mathrm{j}\right\}\end{array}\end{array}\end{array}\end{array}\right\} $$
(2)

The computation of GLCM at D = 1 is described as follows.

figure d

1.5 Dempster-Shafer theory (D-S theory) [44]

D-S is a mathematical approach to combine evidences based on belief function and plausibility. Dempster-Shafer theory is an approach. Each fact has a degree of support, between 0 (no support for the fact) and 1(full support for the fact). This Theory differs from Bayesian approach in a way it uses “Dempster’s rule of combination” to incorporate new information and can be considered a generalization of Baye’s rule. This theory also maintains a dual concept of weight.

Let Θ be a set of possible conclusions: Θ = {f1,f2, …, fn}. Each fi is mutually exclusive and Θ is exhaustive.

1.5.1 Frame of discernment

The frame of discernment of Θ is the set of all possible subsets of Θ = {f1, f2, f3}. Then the frame of discernment of Θ is:

$$ \left(\mathrm{\O},{\mathrm{f}}_1,{\mathrm{f}}_2,{\mathrm{f}}_3,\left\{{\mathrm{f}}_1,{\mathrm{f}}_2\ \right\},\left\{{\mathrm{f}}_1,{\mathrm{f}}_3\right\},\left\{{\mathrm{f}}_2,{\mathrm{f}}_3\right\},\left\{{\mathrm{f}}_1,{\mathrm{f}}_2,{\mathrm{f}}_3\right\}\right) $$

Ø has a probability of 0. The element in the power set has a probability between 0 and 1 and the probability of {f1, f2, f3} is 1.

1.5.2 Mass function

It is a function represented as m(A) that gives a value between [0, 1] to each focal element A. It is based on the evidence. The value 0 represents no belief and value 1 demonstrates total belief and is allocated to every subset of the power set. If mass function of an individual proposition of A is m(A) then it is given in Equation 3:

$$ \sum \limits_{A\subset \theta }m(A)=1 $$
(3)

1.5.3 Belief and plausibility of A

The belief in an element A is the sum of the masses of elements which are subsets of A is represented as Equation 4.

$$ Bel(A)=\sum \limits_{B\subset A}m(B) $$
(4)

The plausibility pl(A) is the sum of all the masses of the sets that intersect with the set given in Equation 5.

$$ pl(A)=1- Bel\left(\neg A\right)=\sum \limits_{B\cap A\ne \varnothing }m(B) $$
(5)

1.6 Belief interval of A

The confidence associated with set A is defined by the belief interval:

$$ \left[\mathrm{bel}\left(\mathrm{A}\right)\ \mathrm{pl}\left(\mathrm{A}\right)\right] $$

The probability lies between bel(A) and pl(A). Probability cannot be less than bel(A)and pl(A) represents the maximum possible value of prob(A). A small difference between bel(A) and pl(A)shows the certainty about our belief and large difference shows the uncertainty about our belief.

1.7 Dempster rule of combination

The mass of focal element is combined by using following rule of combination:

$$ {\mathrm{m}}_{\mathrm{final}}=m1\oplus m2\oplus m3\oplus m4 $$

Where

$$ \left[m(A)\oplus m(B)\right]=m(c)=\frac{\sum \limits_{A\cap B=C}m(A)m(B)}{1-\sum \limits_{A\cap B=\varphi }m(A)m(B)} $$
(6)

1.8 Support vector machine (SVM) classifier [52]

We have used Support Vector Machine (SVM) [52] as a classifier to classify fingerprints into live and fake classes since it provides an effective result for real time applications. We have a tendency to pass training feature set and testing feature set to the SVM classifier wherever it classifies to the tested fingerprint pictures into fake and live class. SVM uses different kernel mapping functions like linear, polynomial and Gaussian, etc. that transforms a feature vector into another high dimensional area. In the proposed system we use linear kernel for fingerprint and iris classification.

2 Proposed multimodal Liveness detection frameworks

Unimodal liveness detection systems are less secure and have unsatisfactory detection. New unnoticed materials and methods continuously challenged state-of-the-art spoofing detectors in a unimodal detection system. The major issue with unimodal liveness detection system is that no single technology is suitable for all the applications. Therefore, in this paper, we have proposed a multimodal liveness detection method that can be used to compensate the limitations of a unimodal liveness detection system. The proposed system has less FTA (failure-to-acquire), FTE (failure-to-enroll), FNMR (false non-match rates) and FMR (false match rate). The proposed method reduces data distortion and increased reliability.

The proposed multimodal liveness detection method is shown in Fig. 3. The proposed method consists of three modules. The first module describes a fingerprint liveness detection approach. The second module demonstrates iris liveness detection. Finally, a decision level fusion is implemented in the third module.

Fig. 3
figure 3

Framework of proposed multimodal detection system

2.1 Fingerprint Liveness detection approach

The spoof fingerprint liveness detection is taken into account as a two-class classification i.e. real or fake. Real and fake fingerprint pictures possess different textures properties [13]. In this module, a feature vector consisting of texture features is utilized for classification by a support vector machine to classify input fingerprint image. The block diagram of the fingerprint liveness detection approach is shown in Fig. 4.

Fig. 4
figure 4

Framework of fake fingerprint detection system

2.2 Preprocessing

The preprocessing step is used is to enhance the image appearance and remove the unrelated noise from an image. In this work, to improve the contrast of the input image I, histogram equalization is performed:

$$ {H}_{i,j}=\left\lfloor \left(L-1\right)\sum \limits_{n=0}^{I_{i,j}}N{H}_n\right\rfloor $$
(7)

Where H is the histogram equalized image, L is the possible intensity value and NH denote the normalized histogram.

2.3 Extraction of feature vector

  1. A)

    The spatial analysis is performed by using following global macro features from NGTDM proposed by Amadasun [8].

  1. 1.

    Resized the fingerprint image into 128 × 128pixels.

  2. 2.

    Find the different NGTDM using d = 1 for each individual finger print image.

  3. 3.

    Calculate the following two global texture features from NGTDM:

• Coarseness: The quality of being rough or harsh.

• Busyness: Level of busyness is measured in terms of spatial magnitude of intensity changes.

  1. B)

    To extract the statistical texture features from GLCM, following steps are taken:

  1. 1.

    Compute four normalized GLCM matrices from the processed fingerprint image of M = 128 × 128at distance d = 1, in the directionθ = 00, 450, 900, 1350.

  2. 2.

    Find out a single matrix by taking the averages of all the four normalized GLCM matrices.

  3. 3.

    Slide a window of size m × mpixels over the GLCM matrix of 128 × 128 with a translation step t = m/2 pixels. In our experiments we have taken m = 16. The number sub images are computed as follows:

$$ N={\left[ floor\left(\left(M-m\right)/t\right)+1\right]}^2 $$
(8)
  1. 4.

    For each sub image, calculate the following six local Haralick features [24]:

    • Energy: It deals with the smoothness of the image.

    • Contrast: Measures of local level variations of contrast.

    • Correlation: Measures of correlation between pixels in two dissimilar directions.

    • Variance: Measures contextual information i.e. dispersion around the mean of combinations of reference & neighbor pixels.

    • Homogeneity: Measures that takes high esteems for low-differentiate pictures.

    • Entropy: Measures of mediation and takes low esteems for smooth pictures.

2.4 Iris Liveness detection approach

This is a classification problem, which is used to differentiate between real and fake iris images. The block diagram is given in Fig. 5 and this module consists of the following steps.

Fig. 5
figure 5

Framework of fake iris detection system

Segmentation of iris image using integro-differential operator [11].

Daugman’s rubber sheet model used for normalized the iris image [30].

Extract iris features using LBHXBP descriptor [4].

Classification done by using SVM classifier.

In the proposed module for iris fake detection, a feature descriptor known as Local Binary Hexagonal Extrema Pattern(LBHXEP) [4] is used. The objective of the descriptor is to extract a feature vector from normalized iris image. In this descriptor, we find the extrema (maximum and minimum) values and indexes from the local Hexagonal pattern and find the average of two center pixels in hexagonal pattern. We derived the relationship between one center pixel and all 6-neighbors at distance 1 and 2. The intensity values of the local hexagonal extrema are compared with the intensity value of the center pixel. The LBHXEP over the normalized iris image M is given calculated using Equation 9.

$$ LB{H}_X EP=\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{$n$}\right.\ast {\sum}_{i=1}^n LB{H}_xE{P}_i $$
(9)

Where n is the number of sub images and the Local Binary Hexagonal Extrema Pattern (LBHXEPi) for one local pattern is calculated as shown in Fig. 6.

Fig. 6
figure 6

LBHXEPi,j Computation

3 Multimodal fusion using Dempster-Shafer theory

In a multimodal biometric system, the biometric information is captured from different biometric modalities and from different sensors. The decision level fusion strategy integrates biometric traits in an exceedingly easy as compared with feature level fusion, which typically directly integrates totally different biometric traits at the feature level. Fusion is performed at a later stage in decision level, as compared to the multi-biometric system at feature level fusion or matching score level fusion methods. In the proposed system, each biometric modality is classified independently and the final classification is done by fusing the decision of two modalities i.e. fingerprint and iris, at decision level using D-S theory. In the proposed method, we define the two frame of discernment:

$$ {\varTheta}_f=\left\{{\theta}_{fr},{\theta}_{fr}\right\} $$
$$ {\varTheta}_i=\left\{{\theta}_{ir},{\theta}_{ir}\right\} $$

Where θfrand θfi represent the real and impostor classes in finger and θir, θii represent the real and impostor classes in iris respectively. The decision level fusion is implemented by the following steps:

  1. 1)

    To convert the genuine predictive value and imposter predictive value for finger and iris into evidence in the range [0, 1].

$$ {m}_i\left({\theta}_{ij}\right)=\psi \left({S}_i\right) $$
(10)

where i corresponds to the fingerprint and iris, and j represents genuine or impostor, Si is the genuine and imposter average error rate (AER) value, by the classifier i. In the proposed method, sigmoid transform function is used and given as in Equation 11:

$$ \uppsi \left({\mathrm{s}}_{\mathrm{i}}\right)=1/\left(1+{\exp}^{-\mathrm{Si}}\right) $$
(11)
  1. 2)

    Combine all evidences for all focal elements for mf and mi using D-S rule of combination for the following frame using Equation 12.

$$ \left\{\phi, {\theta}_{fg},{\theta}_{fi},{\theta}_{ig},{\theta}_{ii},\left\{{\theta}_{fg},{\theta}_{ig}\right\},\left\{{\theta}_{fg},{\theta}_{ii}\right\},\left\{{\theta}_{fi},{\theta}_{ig}\right\},\left\{{\theta}_{fi},{\theta}_{ii}\right\}\right\} $$
$$ {m}_{fusion}\left({\theta}_{ij}\right)={m}_f\left({\theta}_{ij}\right)\oplus {m}_i\left({\theta}_{ij}\right) $$
(12)
  1. 3)

    Calculate belief and plausibility to all focal elements of finger and iris.

$$ Bel\left({A}_{fi}\right)=\sum \limits_{B\subset A}m(B) $$
(13)
$$ pl\left({A}_{ft}\right)=\sum \limits_{B\cap A\ne \varphi }m(B) $$
(14)
  1. 4)

    Proposed liveness detection is two class problem (Accept, Reject) and final result is obtained by applying threshold t:

$$ \mathrm{Result}=\Big\{{}_{\mathrm{Reject}\kern1.75em \mathrm{otherwise}}^{\mathrm{Accept}\kern1.75em t>=\left[ bel(A)- pl(A)\right]} $$
(15)

4 Results and discussion

This section describes the analysis of classification by the proposed system for liveness detection. To estimate the effectiveness and accuracy of the proposed system, we have compared the results of the proposed system with existing methods. The proposed system is implemented in MATLAB 8.3.0.532(R2016a) at Intel® core™ i7 CPU @ 2.2 GHz with 8 GB of RAM. The experimental results are depicted in the form of graphs and tables.

4.1 Dataset

The performance of the proposed system is tested on the standard datasets ATVS-FFp [17], ATVS-FIr [17], LivDet2011 [51] and IIID-CLI database [50]. As far as our knowledge is concerned there is no spoof dataset which has multi traits for the same person. In the proposed system, a virtual user is considered and organic dataset for that user is created by pairing fingerprint and iris images from the available dataset for the multimodal liveliness detection system.

4.1.1 ATVS-FFp dataset [17]

Two types of the dataset are used in the ATVS-FFp. The first dataset is prepared with the cooperation of the user and second is without the cooperation of the user. In the first dataset, index and mid fingers for the left and right hands of 17 subjects are taken. In this dataset three sensors named as Precise SC100 (S1), BiometrikaFX2000 (S2) and Yubee (S3) are used. The database consists of 816 real and spoofed image samples.

The second dataset contains 768 real images and the same number of fake images of 16 persons. For cross-validation, we have reconstituted the third dataset from the ATVS-FFp dataset, where we mix the entire sample from both cooperative and non-cooperative fingerprint images. In Table 2 structure of ATVS-FFp dataset is demonstrated.

Table 2 Structure of the ATVS-FFp

4.1.2 ATVS-FIr dataset [17]

In this dataset printed images of the real and fake iris of 50 users have been taken. It comprises 800 fake and real iris images. The real and fake samples of iris were captured using only one sensor. Table 3 represents the structure of ATVS-FIr dataset.

Table 3 Structure of the ATVS-FIr DB

4.1.3 LiveDet2011 dataset [51]

This dataset is created during spoof fingerprint detection competition held in 2011. Four sensors (Biometrika, Digital Persona, Italdata and Sagem) are used to extract the iris images. Six unique materials known as gelatin, latex, PlayDoh, silicone, ecoflex and wood are used to manufacture fake fingerprint. Structure of the LivDet2011 is shown in Table 4.

Table 4 Structure of the LivDet2011 dataset

4.1.4 IIIT-D CLI dataset [50]

In this dataset, left and right iris images of 101 are captured with 202 iris classes. Total of 6570 iris images are captured. Two iris sensors i.e. Cognet and Vista are used for extraction. Table 5 is used to represent the structure of IIIT-D CLI dataset.

Table 5 Structure of the IIIT-D CLIdataset

4.1.5 Paired dataset

This organic paired database is created for multimodal liveness detection from two different modalities i.e. fingerprint and iris modalities of different users from ATVS-FFp and ATVS-Fir datasets. A chirmeric dataset is compiled by pairing iris and fingerprint images from different datasets. This chirmeric dataset consists of 150 virtual users with 150 samples of real and fake fingerprint and iris images. Table 6 illustrates the test dataset for fusion at decision level.

Table 6 Structure of the paired dataset

4.2 Evaluation measures

The proposed method is validated using the following evaluation measures:

$$ \mathrm{Average}\ \mathrm{Error}\ \mathrm{Rate}\ \left(\mathrm{AER}\right)=\left( FAR+ FRR\right)/2 $$
(16)

Where

$$ \mathrm{False}\ \mathrm{Accept}\ \mathrm{Rate}\ \left(\mathrm{FAR}\right)=\frac{\left(\mathrm{No}.\mathrm{of}\ \mathrm{accepted}\ \mathrm{imposters}\ \mathrm{as}\ \mathrm{real}\right)\times 100}{\left(\mathrm{Total}\ \mathrm{no}.\mathrm{of}\ \mathrm{imposters}\right)} $$
(17)

and

$$ \mathrm{False}\ \mathrm{Reject}\ \mathrm{Rate}\ \left(\mathrm{FRR}\right)=\frac{\left(\mathrm{No}.\mathrm{of}\ \mathrm{rejected}\ \mathrm{reals}\ \mathrm{as}\ \mathrm{imposter}\right)\times 100}{\left(\mathrm{Total}\ \mathrm{no}.\mathrm{of}\ \mathrm{reals}\right)} $$
(18)

AER is calculated from FAR and FRR and average classification accuracy (ACA) is the criteria to assess the reliability of a biometric modality. The less value of AER and the high value of ACA denote the good performance of the system.

$$ \mathrm{Average}\ \mathrm{Classification}\ \mathrm{Accuracy}\ \left(\mathrm{ACA}\right)=\frac{\left(\mathrm{Correct}\ \mathrm{real}\ \mathrm{rate}+\mathrm{Correct}\ \mathrm{fake}\ \mathrm{rate}\right)}{2} $$
(19)

4.3 Experiments on fingerprint dataset

In the first experiment, we have analyzed the AER and ACA value for different sensors using ATVS-FFp dataset and LiveDet 2011 dataset. Table 7 illustrates the AER and ACA for cooperative, non-cooperative and cross-validation dataset on three sensors. It is concluded from the results that for a cooperative dataset, Sensor S2 performs better than other two sensors and for cross-validation Sensor S3 is better than the other two. Table 8 shows AER value and ACA value of Sensor Sagem outperform than the other three sensors.

Table 7 AER value and Accoracy for cooperative, non-cooperative and cross dataset
Table 8 AER value and ACA for LivDet2011 dataset

To further validate the performance a Receiver Operating Characteristics (ROC) curve between False Positive Rate (FPR) and True Positive Rate (TPR) for ATVS-FFp dataset is also analyzed. ROC curves are used to show in a graphical way the connection/trade-off between sensitivity and specificity for every possible test. The area under the ROC curve gives an idea about the benefit of testing. Figure shows that sensor S2 and Sensor S1 is more effective for a cooperative dataset and non-cooperative dataset respectively. Figure 8 validates the performance of the proposed system on LivDet 2011 dataset.

Fig. 7
figure 7

ROC curve between sensors of cooperative dataset and non-cooperative dataset of ATVS datset

Fig. 8
figure 8

ROC curve between Sensors of LivDet2011 dataset

Table 9 describes the relative comparison of AER between the proposed system and state-of-art methods [12, 17]on the ATVS dataset. We can strongly conclude that the proposed system is more efficient than the other two state-of-the-art systems.

Table 9 Comparison between AER on Biometrika Sensor

To further validate the performance of the proposed system, we compare the AER value of the proposed system with Galbally et al. [17] on a cooperative dataset and non-cooperative dataset. The Sensor S2 is equipped with a 32-bit RISC processor and implements very reliable symmetric encryption algorithms with 128-bit keys. Besides the resolution and the size of the sensing area, the images produced by a fingerprint scanner are characterized by other factors which conclude their contrast and geometric distortion. The texture features used in the proposed system suppress these distortions and produce a robust feature vector for effective liveness detection. Figure 9 demonstrates that the proposed system shows low AER for Sensor S2 on cooperatice dataset and low AER for Sensor S1 on non-cooperatice dataset.

Fig. 9
figure 9

Comaprison between AER of proposed system with Galbally et al. [17]

Figure 10 shows the relative comparison of AER between the proposed method and state-of-the-art methods [12, 23, 38] on the LivDet2011 dataset. It is depicted from Fig. 10 that the proposed system is competent and effective as compared state-of-the-art systems.

Fig. 10
figure 10

Comaprison between AER of proposed system wth others tested on LivDet2011 dataset

Table 10 illustrates the ACA of the different state-of-the-art liveness detection system. It is evident from the analysis that the proposed system has a higher accuracy i.e. 94.7% for both fake and real fingerprint liveness detection system, which is comparatively better than other related works.

Table 10 ACA of proposed system with state-of-the-art approaches

4.4 Analysis of multimodal liveness detection system

After fusing the decision from fingerprint and iris liveness detection, a threshold is applied to produce effective decision result. The impact of the threshold on the liveness detection performance is illustrated in Table 11.

Table 11 Impact of threshold on AER

The proposed D-S fusion approach is measured on AER,which incorporates both FAR and FRR for finger liveness detection system and iris liveness detection system. In the proposed system organic dataset is used. AER performance of unimodal detection system with multimodal detection system is shown in Table 12.

Table 12 AER performance of different approaches

The performance of the proposedmethod is also validated using ROC curve analysis as shown in Fig. 11. ROC curve represents the complete sensitivity/specificity report. It is an elementary tool for diagnostic test estimation. In a ROC curve, the true positive rate (Sensitivity) is plotted in function of the false positive rate (Specificity). Each point on the ROC curve represents a sensitivity/specificity pair. The area under the ROC curve (AUC) is a measure of how well a parameter can distinguish between two classifications i.e. real or fake. D-S based fusion approach achieve better performance over the individual modality i.e. fingerprint and iris detection system.

Fig. 11
figure 11

ROC curve between unimodal systems and multimodal system

We have further demonstrated the performance of the proposed method against quality based approaches i.e. sum-based and product-based. Figure 12 demonstrates the D-S fusion is better than the other two fusion methods i.e.sum-based and product-based.

Fig. 12
figure 12

ROC curve between D-S fusion and other quality based fusion

The performance of the proposed D-S fusion based approach in term of ACA is described in Table 13. It is concluded that the proposed system behaves better than other state-of-the-art methods and unimodal liveness system.

Table 13 ACA performance of different approaches

5 Conclusion

Liveness detections are methods to identify presentation attack by differentiating a real and fake biometric representation. In the proposed work, we have investigated the effectiveness of texture features for the liveliness detection system. The proposed method was further improved by fusing independent decision of individual modalities i.e. fingerprint and iris. The standard D-S theory is used for decision level fusion. The performance of the proposed method is tested on different standard datasets. The experimental results illustrate the effectiveness of the proposed approach and can improve the security level and reliabilityof any biometric system.