1 Introduction

The unprecedented growth of technology has connected everbody all over the globe and allowed easy sharing of multimedia content in no restriction of time and access. However, the unlimited access has posed new threats to these multimedia data where the illegitimate individuals can intercept the data in between transmission and misuse it. Various techniques like encryption, steganography,cryptography, watermarking etc are being employed to provide authenticated systems. However, recently the biometric techniques are coupled with these aforementioned approaches to enhance the ability of distinguishing authentic users from non-authentic ones. The user specific biometric traits based on behavioral or physiological characteristics of an individual helps to build more robust systems. While designing a biometric based authentication scheme, various important factors like verification accuracy rate, security, costing, robustness against attacks, computational time and scalability of the system must be considered [14]. Some of the most efficient approaches for securing these biometric systems are encryption [9, 27, 29] or watermarking where a secret information is embedded without deteriorating the cover image [11, 14, 15, 24, 30]. Various other hybrid schemes like combining encryption with biometrics [1], cryptographic biometric systems [2], steganography or watermarking biometric trait [3, 4] have already been proposed in the literature.

The major drawback of the traditional watermarking schemes is the lack of unique watermark for the rightful owner. They use logos, images, text or random noise sequences as watermarks for embedding in the cover images. These watermarks could be easily tampered or imitated to exploit the rightful ownership or destroy the authentication results creating situation of ambiguity. Thus, user specific biometric traits based or the physical or behavioral characteristics of a person like fingerprint, iris detection, face watermark, gait etc could be used as watermarks. Embedding of such user dependent traits will definitely serve as unique watermarks for authorization and provide requisite level of security to such watermarking systems. For a person claiming as rightful owner, this watermark is extracted from the watermarked cover image and compared with the other samples kept in the database. If a match is found, then it proves the rightful ownership else he is not the legitimate owner.

Many watermarking schemes coupled with biometric information have already been proposed in the literature. Most of them mainly target embedding of one biometric trait into another biometric information, so that the rightful ownership of the biometric samples stored in the databases could be proved [5, 6]. Other schemes aim at embedding of user specific biometric traits in normal images i.e. data unrelated to the user whereby after extraction, the rightful ownership can be claimed. One such watermarking algorithm based on the hybrid domain of DWT-SVD used fingerprint as watermark data. The minutiae co-ordinates matrix was embedded after SVD decomposition into the cover image [7]. Another watermarking scheme embedded binary iris codes into the normal cover images whereby the similarity was computed between the extracted watermark and other samples of user claiming ownership to build the identity of the owner [8]. Other methods aim at exploiting the multimodality where the basic idea is to embed one biometric trait into another biometric trait [10, 12, 13].

A watermarking scheme proposed in the domain of DWT, DCT and LPC (Linear Predictive Coding) embedded multiple features of different biometric traits like face, speech signature into the cover data [10]. The similarity between each extracted trait and the other original samples kept in the database was evaluated based on correlation values. However, no fusion was done to combine the individual similarity scores of the trait. Another watermarking scheme that used fingerprint and iris features to combine into one and obtain a fused watermark was proposed in [12]. This fused watermark was then inserted into the cover data. For verifying the ownership, the fused watermark was extracted and thereafter separated into two individual watermarks. The similarity comparisons were then performed between these individual watermarks and the other original samples stored for claiming the rightful ownership. Another watermarking scheme performed in the DCT domain modified the amplitude of the selected coefficients for embedding of two biometric traits. The fusion of the traits was done at the decision level to proceed further for the authorization purposes [13]. A multimodal biometric watermarking system has been proposed in the present paper. Based on the redundant wavelet transform, the proposed watermarking algorithm embeds two watermarks i.e. iris codes and the face features extracted from the face template into a chosen sub band which is randomized using the Hessenberg decomposition. The iris watermark is scrambled using Arnold chaotic map prior to embedding so as to enhance the security of the scheme. Further, the face features are embedded reversibly based on a secret key without deteriorating the quality of the watermarked image. The ownership verification is performed based on the fused biometric score level. Simulation studies are conducted to show the effect of watermarking on the verification accuracy of the proposed system. Also, the authentication evaluation based on the individual biometric traits as well as the fused score at the decision level are presented to assess its affect on the verification accuracy.

The rest of the paper is organized as follows: In Section 2, some preliminaries are given, Section 3 discusses the proposed multimodal biometric watermarking approach. Experimental results along with analysis are presented in Section 4 and finally, conclusions are stated in Section 5.

2 Preliminaries

A brief overview of the concepts used in the proposed approach is given as follows:

2.1 Redundant wavelet transform (RWT)

Though wavelet transform emerged as one of the very powerful tools for varied applications of image processing but it proved not vey optimal choice for other applications involving analysis of signals, filtering, detection, deconvolution etc. The reconstructed image from the altered wavelet coefficients contained many visual artifacts. This mainly arose owing to lack of the translation invariant property in wavelet transform.

The redundant wavelet transform (RWT) evolved as a discrete estimation of the continuous wavelet transform. Although, history has roots that RWT was proposed independently via different names like the overcomplete WT (OWT) [18], discrete wavelet frames (DWFs)[20], the shift-invariant WT (SIWT) [19], the undecimated WT (UWT)[17], and algorithme trous [16]. These names were given due to the fact that the downsampling was removed which resulted in shift invariance property of RWT with a fixed spatial sampling rate across the whole scale. Thus, resulting in size of the sub-bands same as the size of the input signal. Mathematically, RWT can be defined in terms of scaling h 𝜖 l 2(Z) and wavefilters g 𝜖 l 2(Z) of an underlying orthonormal wavelet transform. To maintain the multiresolution property at each scale, these filters should be adjusted accordingly. The upsampling can be defined as follows:

$$ x[k] \uparrow 2 = \left\{\begin{array}{ll} x[k/2] , & \text{if}\,\, k \ is \ even \\ 0 , & \text{otherwise} \end{array}\right. $$
(1)

At scale j + 1, the scaling and wavelet filters can be as:

$$ h_{j+1}[k]=h_{j}[k]\uparrow 2 , g_{j+1}[k]=g_{j}[k]\uparrow 2 $$
(2)

where h 0[k] = h[k] and g 0[k] = g[k]. Recursive implementation with the filter bank operations will result in RWT of any signal x 𝜖 l 2(Z) as:

$$ c_{j+1}[k]=h_{j}[-k] \times c_{j}[k], d_{j+1}[k]=g_{j}[-k] \times c_{j}[k] $$
(3)

where c 0 = x and j = 0, 1…J − 1. RWT at J scale is collection of sub bands result of filtering operations: X J = R W T J [x]=[c J d J d J − 1d 1] such that

$$ \Vert X^{j} \Vert^{2}= \Vert c_{J} \Vert^{2}+ \sum\limits_{j=1}^{J} \Vert d_{j} \Vert^{2} $$
(4)

2.2 Singular value decomposition

Singular value decomposition is one of the efficient ways for factorization of a square or rectangular matrix to bring up the very structure of the matrix [22]. It is very popular in image processing as it has many advantages like the singular values are very stable in nature, get very less affected by the general image processing distortions and represent the algebraic properties of the image. Singular value decomposition of a real (complex) matrix X of order m × n can be represented as:

$$ X=USV^{T} $$
(5)

where S = d i a g(σ 1,σ 1,…σ r ) and U and V are orthogonal (unitary)matrices and signify the left and right singular values. σ i represent the singular values of the matrix and r = m i n(m, n) is the rank of the matrix satisfying

$$ \sigma_{1} \geq \sigma_{2} \geq \sigma_{3}{\ldots} \geq \sigma_{r} $$
(6)

2.3 Non-linear chaotic map

To enhance the randomness and increase security aspect of the schemes, chaotic systems are often used. It possesses many useful properties like initial parameter sensitivity, non-periodicity, unpredictability etc. which make these maps as the stochastic signal generators [21]. We have employed piecewise non-linear map here to generate random sequences. Mathematically, it can be defined as follows:

F : II where I = [0, 1]

$$ F(x_{k+1})=\left\{ \begin{array}{cll} \left( \frac{1}{I_{i+1}-I_{i}}+a_{i}\right)(x_{k}-a_{i})- \frac{a_{i}}{I_{i+1}-I_{i}}(x_{k}-a_{i})^{2} & if & x_{k} \epsilon [I_{i},I_{i+1})\\ 0 & if & x_{k}=0.5 \\ F(x_{k}-0.05) & if & x_{k} \epsilon (0.5,1] \end{array} \right. $$
(7)

Where x k 𝜖(0, 1] and I i is the sub interval such that , 0 = I 0 < I 1 < …I i I n + 1=0.5. The parameter \(a_{i}\epsilon (-1,0)\bigcup (0,1)\) is the tune sequence in the i th interval such that

$$ \sum\limits_{i=0}^{n-1}(I_{i+1}-I_{i})a_{i}=0 $$
(8)

The important properties of the above map are as follows:

  • The iteration system obtained by (7) i.e. x k + 1 = F(x k ) is chaotic for all x k 𝜖[0,1]

  • The sequence \({(x_{k})}_{k=1}^{\infty } \) is Ergodic in [0,1] having the uniform probability distribution function ρ(x)=1, which further shows the uniformity of the map, i.e. the probability of each value in [0,1] is equal to be selected.

  • The sequence \({(x_{k})}_{k=1}^{\infty } \) has δ-like autocorrelation function given as

    $$R_{F}(r)=\lim_{j\rightarrow \infty} \frac{1}{j}\frac{{\sum}_{k=1}^{j}x_{k}x_{k+r}}{{\sum}_{k=1}^{j}{x_{k}^{2}}},x\geq 0$$

2.4 Hessenberg decomposition

In Hessenberg decomposition [22], any matrix A can be factorized via orthogonal similarity transformations into the form as follows:

$$ A=QHQ^{T} $$
(9)

where Q is an orthogonal matrix whereas H is an upper Hessenberg matrix, implies H = [h i j :h i j =0 w h e n e v e r i > j + 1]. This decomposition usually involves computation of Householder matrices. A Householder matrix P is the orthogonal matrix of the form:

$$ P=\frac{I_{n}-2uu^{T}}{u^{T}u} $$
(10)

where u is a non-zero vector in R n and I n is the n × n identity matrix. The main reason of using these matrices is the ability of introducing zeros. For instance, suppose x = (x 1,x 2,…,x n ) is a non-zero vector in R n and u is defined as:

$$ u=x+sign(x_{1})||x||_{2}e_{1} $$
(11)

where \(sign(x_{1})=\frac {x_{1}}{|x_{1}|}\) and e 1 is the first column of I n . It then follows:

$$ P_{x}=-sign(x_{1})||x||_{2}e_{1} $$
(12)

i.e. a vector having zeros in all but its first component. There are n − 2 steps in the overall procedure when A is of size n × n which further proves the ability of introducing zeros. At the beginning of k th step orthogonal matrices P 1,P 2,…,P k − 1 have been expressed as:

$$ A_{k-1}=(P_{1},P_{2} {\ldots} P_{k-1})^{T}A(P_{1},P_{2} {\ldots} P_{k-1}) $$
(13)

having the form:

$$ A_{k-1}=\left[\begin{array}{ccc} H_{11}^{(k-1)} & H_{12}^{(k-1)} & H_{13}^{(k-1)}\\ 0& b_{22}^{k-1} & H_{23}^{(k-1)} \\ 0& H_{32}^{(k-1)} & H_{33}^{(k-1)} \end{array}\right] $$
(14)

where \(H^{(k1)}_{11}\) is Hessenberg matrix. Let \(\tilde {P_{k}}=\frac {I_{n-k}-2u^{k}(u^{k})^{T}}{(u^{k})^{T}u^{k}}\) be a Householder matrix with the property that \(\tilde {P_{k}}b^{k-1}\) has zeros in the last nj − 1 components which further follows that the matrix \(P_{k}=diag(I_{n-k},\tilde {P_{k}})\) is orthogonal and A k is given as:

$$ A_{k}={P_{k}^{T}}A_{k-1}P_{k}=(P_{1}P_{2} {\ldots} P_{k-1}P_{k})^{T}A(P_{1}P_{2} {\ldots} P_{k-1}P_{k})\\ $$
(15)
$$ A_{k}= \left[\begin{array}{ccc} H_{11}^{(k-1)} & H_{12}^{(k-1)} & (H_{13}^{(k-1)})^{T}\tilde{P_{k}}\\ 0& b_{22}^{k-1} & (H_{23}^{(k-1)})^{T}\tilde{P_{k}} \\ 0& \tilde{P_{k}}(H_{32}^{(k-1)})^{T} & \tilde{P_{k}}(H_{33}^{(k-1)})^{T}\tilde{P_{k}} \end{array}\right] $$
(16)

where A k is Hessenberg matrix. Since, we need n − 2 steps in the overall procedure, the above (15) can be rewritten as:

$$ H=Q^{T}AQ \rightarrow A=Q^{T}HQ $$
(17)

where H = A n − 2 and Q = P 1 P 2P n − 3 P n − 2

3 The proposed methodology

The proposed scheme is based on the multimodal biometric watermarking system. It fuses the two biometric traits of iris codes and facial features into one and uses it as the decision criteria for distinguishing between genuine and impostor attempts for a claiming owner. The two watermarks i.e. iris codes and facial features are pre-processed prior to the actual embedding into the cover image. The scheme consists of four main phases: Pre-processing of biometric data before embedding, watermark embedding, watermark extraction and authentication. The details of each phase is as follows:

3.1 Pre-processing of biometric data before embedding

Two biometric watermarks: Iris as first grayscale watermark (W 1) and face features as the second binary watermark (W 2) are used in the proposed scheme. The iris images from the database are resized according to the size of the cover images and then scrambled using the arnold transformation based on a secret key to deliver it unreadable depending upon the period of the iris image. This obtained scrambled iris watermark is thus ready to be inserted. The second binary watermark is obtained from the face database images using the feature extraction algorithm as in [26]. The algorithm extracts the features based on the center symmetric local binary pattern(CSLBP) and gray level co-occurrence matrix (GLCM). The local features are captured via CSLBP and thereafter using the GLCM, the co-occurrence of the pixel intensities are encountered with. It serves as a better way of obtaining local feature descriptors for an image instead of histograms where only frequency information is captured. The feature vectors thus obtained are arranged in a matrix and operated upon by reversible XOR operation with the halftone version of the watermarked image. This halftone version is obtained via error diffusion technique and generates a secret key \(K_{W_{2}}\) that is used in the extraction phase.

3.2 Watermark embedding

The embedding phase involves embedding of two watermarks: grayscale and binary in hierarchial order. The embedding is done in such a way so that embedding of one watermark doesn’t disturb the other one. The detailed methodology is depicted in Fig. 1 and described in detail as follows:

  1. 1.

    Perform redundant wavelet transform on the cover image C upto L-level, denoted as follows: \(C_{\Psi }^{\Gamma } \), where Ψ𝜖[1,L], Γ𝜖{A, H, V, D}

  2. 2.

    Selection and randomization of a sub-band: A sub band is selected from the RWT decomposition and randomized prior to actual embedding of the grayscale watermark. The details are as follows:

    1. (a)

      Obtain two random sequences via non-linear chaotic maps using secret keys k 1 and k 2 of length M × M and N × N respectively.

    2. (b)

      Rearrange the two obtained sequences to form two arrays represented by M 1 and M 2 of same length M × M and N × N respectively.

    3. (c)

      Perform Hessenberg decomposition on the arrays M 1 and M 2 to obtain two orthogonal matrices R 1 and R 2.

    4. (d)

      Select a sub band \(C_{\Psi }^{Z}\) and randomize it via R 1 and R 2 as follows:

      $$ \tilde{C}_{\Psi}^{Z}=R_{1}C_{\Psi}^{Z}R_{2} $$
      (18)
  3. 3.

    The grayscale watermark W 1 is scrambled using Arnold cat map Y times to obtain \({W^{s}_{1}}\) which is a secure version as it contains sensitive information of the owner.

  4. 4.

    Embed the scrambled watermark \({W^{s}_{1}}\) as follows:

    1. (a)

      Perform SVD on both correlated watermark \({W^{c}_{1}}\) and chosen sub band \(\tilde {C}_{\Psi }^{Z}\) for embedding as follows:

      $$ \tilde{C}_{\Psi}^{Z}=U_{\tilde{C}_{\Psi}^{Z}} S_{\tilde{C}_{\Psi}^{Z}}V^{T}_{\tilde{C}_{\Psi}^{Z}} $$
      (19)
      $$ {W^{c}_{1}}=U_{{W^{c}_{1}}}S_{{W^{c}_{1}}}V^{T}_{{W^{c}_{1}}} $$
      (20)
    2. (b)

      Modify the singular values of the sub-band with the singular values of the watermark as follows:

      $$ S^{\prime}_{\tilde{C}_{\Psi}^{Z}}=S_{\tilde{C}_{\Psi}^{Z}}+ \alpha S_{{W^{c}_{1}}} $$
      (21)

      where α gives the watermark strength.

    3. (c)

      Perform inverse SVD to construct modified sub-band

      $$ \tilde{C}_{\Psi}^{\beta,Z}=U_{\tilde{C}_{\Psi}^{Z}} S^{\prime}_{\tilde{C}_{\Psi}^{Z}}V^{T}_{\tilde{C}_{\Psi}^{Z}} $$
      (22)
  5. 5.

    Perform inverse randomization of the sub band as follows:

    $$ C^{Z}_{\Psi}=inv(R_{1}) \tilde{C}_{\Psi}^{\beta,Z}inv(R_{2}) $$
    (23)
  6. 6.

    Inverse redundant wavelet transform upto L level is applied to obtain the watermarked image C New.

  7. 7.

    Embedding binary watermark: The binary image W 2 is embedded as follows:

    1. (a)

      Obtain a halftone version \(H_{C^{New}}\) of the watermarked image using error diffusion method.

    2. (b)

      Perform exclusive-OR operation bitwise with the halftone image \(H_{C^{New}}\) and the second watermark W 2 to obtain the extraction key \(K_{W_{2}}\).

      $$ K_{W_{2}}=XOR(H_{C^{New}},W_{2}) $$
      (24)

The embedding of second watermark is lossless as it doesn’t alter the pixel intensities before or after the embedding.

Fig. 1
figure 1

Embedding process of proposed scheme

3.3 Watermark extraction

The watermark extraction phase is depicted in Fig. 2 and detailed as follows:

  1. 1.

    Obtain the halftone version \(H^{\prime }_{C^{\prime }}\) of the suspected watermarked image C using error diffusion method.

  2. 2.

    Extract the binary watermark using bitwise exclusive operation of the secret key with the obtained halftone image.

    $$ W^{\prime}_{2}=XOR\left( H^{\prime}_{C^{\prime}},K_{W_{2}}\right) $$
    (25)
  1. 3.

    Extraction of second watermark: The extracted watermark is evaluated quantitively using various similarity measures with respect to the original binary watermark. If the similarity is greater than the predefined threshold value only then the extraction of the second watermark is proceeded else it proves violation of rightful ownership. The process for extraction is detailed as follows:

    1. (a)

      Perform redundant wavelet transform on the watermarked image C upto L-level, denoted as follows: \(C_{\Psi }^{\prime \Gamma } \), where Ψ𝜖[1,L], Γ𝜖{A, H, V, D}

    2. (b)

      Randomization of sub band: The selected sub band is randomized for extraction of the grayscale watermark as follows:

      1. (i)

        Obtain two random sequences via non-linear chaotic maps using secret keys k 1 and k 2 of length M × M and N × N respectively.

      2. (ii)

        Rearrange the two obtained sequences to form two arrays represented by M 1 and M 2 of same length M × M and N × N respectively.

      3. (iii)

        Perform Hessenberg decomposition on the arrays M 1 and M 2 to obtain two orthogonal matrices R 1 and R 2.

      4. (iv)

        Randomize the selected sub band \(C_{\Psi }^{\prime Z}\) with R 1 and R 2 as follows:

        $$ \tilde{C}_{\Psi}^{\prime Z}=R_{1}C_{\Psi}^{\prime Z}R_{2} $$
        (26)
      5. (v)

        Perform SVD on the randomized sub band \(\tilde {C}_{\Psi }^{\prime Z}\) as follows:

        $$ \tilde{C}_{\Psi}^{\prime Z}=U_{\tilde{C}_{\Psi}^{\prime Z}} S_{\tilde{C}_{\Psi}^{\prime Z}}V^{T}_{\tilde{C}_{\Psi}^{\prime Z}} $$
        (27)
      6. (vi)

        Extract the singular values of the grayscale watermark as follows:

        $$ S_{W^{\prime}_{1}}=\frac{S_{\tilde{C}_{\Psi}^{\prime Z}}-S_{\tilde{C}_{\Psi}^{Z}}}{\alpha} $$
        (28)
      7. (vii)

        Perform inverse SVD to obtain the extracted watermark.

        $$ W^{\prime}_{1}=U_{{W^{c}_{1}}}S_{W^{\prime}_{1}}V^{T}_{{W^{c}_{1}}} $$
        (29)
      8. (viii)

        Descramble the extracted watermark Yt to obtain the actual watermark \(W^{ext}_{1}\).

Fig. 2
figure 2

Extraction process of proposed scheme

3.4 Authentication process

The proposed scheme performs authentication of rightful owners of digital images via extraction of both the biometric traits. The extracted watermarks are compared with the other samples provided by the users and stored in the database for this purpose. For face biometrics, feature is computed from the suspected watermarked image using the same secret key \(K_{W_{2}}\) as used at the time of embedding. This extracted face features serve as the second watermark and compared with the features extracted from the other samples provided by the user in the database. The similarity measure used in the algorithm is the d1 distance measure. It performs better for local pattern matching as compared to other distance measure like euclidean, manhattan,canberra,chi-square etc. If the match exceeds the threshold level, the extraction of first watermark is preceded else it is categorized as impostor attempt and halted. The iris codes are extracted from the inverse process of embedding as described in the extraction phase. Once the singular values are obtained for the iris code, inverse singular value decomposition is computed to get the grayscale scrambled iris template. The obtained iris watermark is decrambled using the secret key of the arnold scrambling depending upon its time period which finally gives the first extracted grayscale watermark. Thereafter, the feature extraction algorithm is applied on the extracted watermark to collect its features which is compared with the feature obtained from the other database samples. The decision can be based on the individual scores of each biometric trait or combined multimodal score. For instance, matching scores for feature extracted from face or iris templates or rather a fused score based on the two traits.

The d1 distance measure used to account for similarity between the extracted traits and the other database samples must be as minimum as possible. It must be normalized and subtracted from one prior to actual fusion of both traits using the weighted sum rule method ([25, 26]). Similarity measures must be in the same domain prior to fusing as for present case 1 indicates perfect match and 0 as perfect mismatch. Hence, fused measure becomes as follows:

$$ S_{if}= w*S_{N_{i}}+ (1-w)*S_{N_{f}} $$
(30)

where \(S_{N_{i}}\) and \(S_{N_{f}}\) represents the normalized S i and S f values ranging in the interval [0,1] with 1 as perfect match and w as weight. The fused measure S i f is thereafter checked and if it is greater than some threshold t, it indicates authentic user else unauthentic one.

The multimodality enhances the effectiveness of the verification stage as it reduces chances of error and provides features for convenience as it’s just like a single measure comparison and no heavy computations are needed.

4 Experimental results and discussion

The performance of the proposed multimodal biometric watermarking algorithm has been tested upon SDUMLA-HMT database integrating the multiple biometric feature information into one. Whenever a ownership claim is to be resolved, the face feature vectors and iris binary templates are extracted from the suspected watermarked image and compared with the other samples of the user stored in the database. If a match is found, it is categorized as genuine claim otherwise taken as impostor attempt. The most important factor for the success of any biometric watermarking algorithm is its recognition accuracy. To validate the proposed scheme, verification based on face features, iris codes and the combined multimodal trait have been described in detail as follows:

4.1 Description of databases

For the experimental studies, biometric multimodal data of SDUMLA-HMT database has been taken. It contains data about 106 persons. We have considered only two biometric traits: face and iris images [25]. The face database contained four different kind of images varying in poses (look upward, forward, and downward), illumination conditions(normal illumination and with lamp illumination), facial expressions (smile, frown, surprise, and close eyes) and accessories (glasses and hat). The face database contains 7 × (3 + 4 + 2 + 3) × 106 = 8, 904 images. The size of images is 640 × 480 pixels with 24 bit bmp files and size of database as 8.8G. In iris images database, 10 samples (5 images per eye) for each individual was captured under infrared illumination. Hence total 1060(2 × 5 × 106) images with size as 768 × 576 were considered.

The performance of watermarking algorithms is generally based on evaluating imperceptibility measures like peak signal to noise ratio, mean square error or other metrics like normalized cross correlation metrics. But when it comes to a biometric watermarking system, then we need to ensure performance in terms of recognition accuracy. A watermarked biometric system should further enhance the security aspects of the biometric traits used without compromising in its quality and features. For the proposed algorithm, we have verified the performance based on face feature recognition, iris recognition and combined multimodality feature. The details are described as follows:

4.2 Performance evaluation using face verification

The images from the face database are fetched as input to the face detection algorithm which gives the face region as output of size 320 × 240 [31]. The face region is then employed with the feature extraction algorithm which computes the prominent local features based on the center symmetric local binary pattern and gray level co-occurrence matrix. The co-occurrence of pixel pairs is considered using different distances and directions in the obtained local pattern map [26]. Thereafter, these feature sets are matched using d1 distance measure.

4.3 Performance evaluation using iris verification

The images from iris database are resized according to the size of the cover images and thereafter, scrambled via arnold chaotic map based on secret key to render it unreadable and encrypted. It is then embedded into the sub bands of the cover image imperceptibly. For the recognition, the iris template is extracted from the watermarked image and features obtained from it are compared with the features extracted from the other samples of the user’s iris templates stored in the database. Similarity based on d1 distance measure is computed for distinguishing the genuine and impostor attempts.

4.4 Performance evaluation using multimodal biometrics

Biometric systems based on single biometric trait often face problems of non-universality and circumvention [23]. Hence, multimodal biometric systems are preferred for biometric verification where decision is based on the combined measure of multiple biometric traits considered. In the proposed algorithm, we have employed the face and iris traits as biometric features. Hence, a fused matching score based on weighted sum of the individual similarity scores of face and iris features is considered for ownership verification based on some threshold value. For a given image, if the fused score comes out to be greater than the threshold value, it is categorized as genuine else as impostor attempt, thus proving the authenticity of the claim.

4.5 Parameters considered in the algorithm

4.5.1 Threshold determination

The security of the proposed watermarking algorithm scheme depends on various factors considered in the algorithm like the threshold value which decides the limit when the extraction of the iris codes will be proceeded. The similarity between the extracted facial features is computed with respect to the other user samples stored in the database. Only, when the authentication upto a certain limit is reached, then only the algorithm will be proceeded further. That is to say, when certainity about genunity is assured upto a limit. The threshold value is determined based on the large number of experiments where biometric templates are embedded and then matched after extracting from the other database samples. The maximum similarity reached is taken up as the threshold limit because as the threshold increases, lesser impostor attempts will be falsely classified as matches.However, more genuine attempts will be falsely classified as nonmatches. Hence, an optimum threshold value will balance security with robustness.

4.5.2 Sensitivity analysis

The sensitivity of the proposed multimodal biometric system relies on the secret keys k 1 and k 2 used as initial seed values for randomization of the cover image using non linear chaotic maps. The high sensitivity to the initial conditions enhances robustness against attacks as slight changes in the key value will totally change the results. Three cases may arise: (1) when k 1 is slightly changed, (2) when k 2 is slightly changed, and (3) when both k 1 and k 2 are slightly changed. A very slight change in the secret keys will render the extracted watermarks unrecognizable.

Also, scrambling of the iris codes using arnold map t times prior to embedding again strengthens the security as unauthentic owner will not be able to decrypt the watermark even if he obtains it. These aspects build the fact that proposed multimodal biometric scheme assures security.

4.5.3 The parameter n

The parameter n represents the level of RDWT decomposition in the proposed algorithm. As the number of levels increase, the robustness enhances against attacks. However, the time complexity increases. Besides a proper level decomposition, the verification accuracy of the multimodal biometric system doesn’t get affected. The lowest frequency sub band of third level RDWT decomposition has been used for the experimental studies.

4.6 Results and discussion

To evaluate the efficacy of the proposed scheme, samples from the SDUMLA-HMT database were used.Some are shown in Fig. 3. The biometric verification was done based on the watermark data extracted from the cover images. Some sample cover images are depicted in Fig. 4. Two biometric traits that is, the iris templates and features extracted from the face image were employed as the respective watermarks. Face region was extracted using a triangle based face detection algorithm whereas iris images were resized depending on the size of the cover image. The ownership verification was done based on the fused similarity score. For the presented simulation study, 5 iris samples of right eye and 5 face images (frontal, different illumination, accessories, expressions) per subject in the dataset of 106 users were considered. To obtain the optimum value of the weight w used in the fused score, individual scores were calculated for raw biometric data from the database. The accuracy error rate was aimed to be minimized depending upon the Receiver Operator Characteristics(ROC) curves with smallest achievable values of Area Under Curve(AUC). The optimum value for the weight came out to be 0.54. The proposed system was thereafter checked for how the verification accuracy of the system gets affected by the generation, embedding and extraction of the biometric traits i.e. the iris codes and the face features into the cover images. The extracted watermarks were thereafter compared with the other biometric samples of the user kept in the database and evaluated based on the individual as well as the fused matching scores. The d1 distance was used as the criteria for matching of the iris templates as well as the face features extracted from the face biometric data.The total number of genuine and impostor attempts were equal to 5∗4∗106=2120(number of intraclass comparisons) and 106∗105=11130(number of interclass comparisons) respectively.

Fig. 3
figure 3

Sample images of faces and Iris used as watermarks from SDUMLA-HMT database

Fig. 4
figure 4

Test images used as cover images in the proposed scheme

The verification accuracy of the proposed system based on the individual traits, Iris templates S i and face features S f as well as the fused one S i f have been presented through ROC curves in Figs. 56 and 21. As can be observed from the Fig. 5, the verification accuracy is almost at the same level before and after watermarking as can be inferred from the ROC curve and Equal Error Rate(EER) almost at the same level. Now, for the attacks scenario, the EER slightly increases for the histogram equalization attack, and thereafter keeps on increasing for sharpening,gaussian blur, resizing, pixilation, wrapping, salt and pepper noise and finally JPEG compression. The verification accuracy of the system keeps on deteriorating as the EER increases. Other attacks have also been tested with like horizontal flipping, contrast adjustment, row-column deletion, vertical flipping,rotation and SPIHT compression. The variation plot is depicted in Fig. 5.

Fig. 5
figure 5

Result of Histogram Equalization Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

For assessment of the verification accuracy based on the face features, the ROC curves are plotted in Fig. 6. The verification accuracy keeps on deteriorating as the different attacks are applied. The order of deterioration goes as for histogram equalization, sharpening, gaussian blur, resizing, pixilation, wrapping, salt and pepper noise and JPEG compression. The EER increases in this order signifying the decrease in verification accuracy. The overall EER decreases for each of the attacks when the fused matching score is considered as can be observed in Fig. 21. The proposed scheme has been tested for variety of cover images. Results for some of the test images in terms of ownership verification accuracy evaluated based on PSNR, NCC, AUC and EER have been presented in Tables 12345 and 6. The value of scaling factor (α) for these experimental results was taken to be 0.035. The scheme has been tested under different attack scenarios with results presented for one of the test cover images and grayscale watermark in Figs. 7891011121314151617, 1819 and 20. The binary watermark is in form of feature vectors extracted from the face. Hence, its quantitative evaluation is presented under aforementioned attack scenarios in Table 2 based on NCC.

Table 1 PSNR values of test images
Table 2 Evaluation of extracted watermarks based on NCC against various attacks
Table 3 Results for Lena image based on Iris template, Face features and Fused traits for proposed approach
Table 4 Results for Baboon image based on Iris template, Face features and Fused traits for proposed approach
Table 5 Results for Pepper image based on Iris template, Face features and Fused traits for proposed approach
Table 6 Results for Cameraman image based on Iris template, Face features and Fused traits for proposed approach
Fig. 6
figure 6

Result of Gaussian Blur Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 7
figure 7

Result of Sharpening Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 8
figure 8

Result of Salt & Pepper Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 9
figure 9

Result of Resizing Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 10
figure 10

Result of JPEG Compression Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 11
figure 11

Result of Pixelation Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 12
figure 12

Result of Wrapping Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 13
figure 13

Result of Row-Column Deletion Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 14
figure 14

Result of Rotation Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 15
figure 15

Result of Horizontal Flipping Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d)Extracted Watermark

Fig. 16
figure 16

Result of Vertical Flipping Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 17
figure 17

Result of Contrast Adjustment Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 18
figure 18

Result of SPIHT Compression Attack (a) Original Cover Image (b) Gray Scale Watermark (c) Attacked Watermarked Image (d) Extracted Watermark

Fig. 19
figure 19

ROC curve for proposed scheme based on analysis of Iris matching score S i

Fig. 20
figure 20

ROC curve for proposed scheme based on analysis of Face features matching score S f

In the proposed multimodal verification scheme, the redundant discrete wavelet domain is chosen for embedding the biometric traits. First of all, the reason behind choosing RDWT instead of DWT was that the size of the sub bands of RDWT remains same as size of the cover image whereas in DWT, it keeps on decreasing significantly after every decomposition. So, RDWT provides sufficient capacity for embedding of watermarks. Two biometric traits are used as watermarks and embedded in different regions of the cover image without affecting each other. The iris template used as the first watermark is rendered unreadable before embedding based on secret key so that no one could retrieve the actual iris template without possessing the secret key. Furthermore, the randomization of a selected sub band prior to embedding based on secret keys again enhanced security. The second watermark which is extracted from the face biometric in the form of face features is embedded reversibly without affecting the watermarked image generated by the embedding of first watermark based on secret key \(K_{W_{2}}\).

The comparison of the proposed scheme with the other existing state of art approaches is an integral part of the success of an approach. First of all, the proposed approach exploits the redundant wavelet transform domain for embedding of watermarks which provides enough capacity. Secondly, as compared to existing schemes [7, 8] based on single biometric trait, be it iris codes or face features, the proposed scheme is based on integrating multiple biometrics into one scheme. Moreover, the affect of watermarking on the verification accuracy of the proposed biometric system has been evaluated based on the extracted watermark data not just measures done in terms of imperceptibility of the watermarked images [10, 12]. Different attack scenarios and their affect on verification accuracy have been assessed. Variety of cover images independent of the biometric traits have also been tested upon for the appropriateness of the proposed authentication scheme with results tabulated in Tables 36 for different cover images. Multimodal biometric authentication scheme based on fingerprint and iris data has been proposed in the literature [13]. It used Daughmans model and hamming distance as the similarity evaluation techniques. The major difference between their scheme and our proposed scheme is that, first of all we used face features and iris codes as the biometric traits. Secondly, we have employed d1 distance as the evaluation criteria. Thirdly, the authentication in our scheme was based on a fused score whereas in their scheme it was at decision level. The decision level authentication required setting of two different threshold values for fingerprint and iris codes based on different conjunction rules. It will be very tough for a user with little experience and will directly affect the verification accuracy of the scheme. However, in our proposed scheme it is very convenient to use as the authentication will be based on a fused score level which will act just like using a unimodal biometric system (Figs. 56 and 21).

Fig. 21
figure 21

ROC curve for proposed scheme based on analysis of fused scores of both traits S i f

Another work based on the fused score has been presented in [28]. However, it is based on fingerprint and iris templates used as biometric traits. They used Daughmans model and hamming distance for construction and evaluation of iris codes whereas for fingerprint, the minutiae points were extracted and matched in terms of their co-ordinates and orientations to obtain a matching score. These scores were then normalized and combined to get a fused score. The objective of the proposed scheme was to show how the verification accuracy of a biometric watermarking system increases upon usage of more than one biometric traits for recognition and the fusion of the evaluation metrics at the score level facilitates distinguishing the genuine and the impostor attempts. The accuracy of the system also depends upon the raw biometric data quality of the databases used. In our algorithm, we have used our own methods for embedding and extraction of biometric watermarks followed by subsequent evaluation by a fused score. In future, we will test it for bigger data sets and fuse more biometric metric to improve upon the verification accuracy of the system.

5 Conclusion

A multimodal biometric authentication scheme has been proposed in this paper. It incorporates two watermarks: one grayscale based on iris template and the other binary based on facial features extracted via feature extraction algorithm based on center symmetric local binary pattern and gray level co-occurrence matrix.The embedding of watermarks in the redundant discrete wavelet transform provided enough capacity for embedding without deteriorating the visual quality of the watermarked images. The performance of the algorithm is evaluated from the user’s verification point of view based on false acceptance rate, false rejection rate, equal error rate, area under curve and corresponding ROC curves. From the experimental results it was proved that fusion of biometric traits enhanced user’s verification accuracy and increased its robustness against different attack scenarios.