Keywords

1 Introduction

The unibiometric system that is based on a single source of information suffers from the problems like lack of uniqueness, non-universality, and spoofing attacks. On the contrary, a multibiometric system harnesses relevant information obtained from multiple biometric cues. A strategic combination of these relevant information obtained from multiple biometric cues may overcome some of the problems of unibiometric systems [13].

Our concern is to combine several unibiometric systems to achieve a multibiometric system that meets the characteristics of a robust system i.e., optimum recognition accuracy and less falsifications [46]. In order to achieve these characteristics the matching scores obtained from different unibiometric systems need transformation and mapping before their fusion. The objective of transformation and mapping operations that refers the normalization process in the biometric terminology, is to supplement the information received at the matching score level of the biometric systems, so that the performance of the combined system improves. Therefore, score normalization is an intrinsic problem. It plays a peculiar role in transforming and mapping the heterogeneous scores of distinct biometric cues into a homogeneous scale.

In literature, the normalization techniques has been found congenial in transforming the heterogeneous score to a homogeneous scale. An evaluation of normalization techniques of matching scores in multibiometric systems has been done by Singh and Gupta [7]. They reported the performance of linear-tanh-linear(LTL) and four-segments-piecewise-linear (FSPL) are better than min-max (MM), z-score and tanh normalization techniques. They also found that MM and z-score normalization techniques are susceptible to outliers. Therefore, it is needed to devise robust and efficient normalization technique that achieves optimum accuracy results.

In [7], let \(O^{T}_{k} = \{r^{T}_{k_{1}}, r^{T}_{k_{2}}, ..., r^{T}_{k_{N}}\}\) be the set of true scores of N individuals and \(O^{I}_{k} = \{r^{I}_{k_{1}}, r^{I}_{k_{2}}, ..., r^{I}_{k_{n}}\}\) be the set of impostor scores of those individuals where, \(n = N\times (N-1)\) for biometric cue k. The composite set of matching scores is denoted as \(O_{k}\) (i.e., \(O_{k} = O^{T}_{k}\cup O^{I}_{k}\) and \(|{O^{T}_{k}\cup O^{I}_{k}}| = N + n = N^{2}\)).

The distance scores (\(r^{'}_{k_{i}}\)) of user i for biometric cue k can be converted into similarity scores in the typical scale, suppose it should be [0, 1] using the formula:

$$\begin{aligned} r_{k_{i}} = \dfrac{max(O^{T}_{k}, O^{I}_{k})-r^{'}_{k_{i}}}{max(O^{T}_{k}, O^{I}_{k})- min(O^{T}_{k}, O^{I}_{k})} \end{aligned}$$
(1)

whereas \(r_{k_{i}}\) is the similarity scores of biometric cue k . Otherwise, if the distance scores lies in the range \([min(O_{k}), max(O_{k})]\) then they are simply converted to similarity scores by subtracting them from \(max(O_{k})\) (e.g., \(max(O_{k}-r^{'}_{k_{i}}\))). The precise summarization of these normalization techniques which transform the raw scores in the typical range of [0, 1], including double-sigmoid (DS), piecewise-linear (PL) are rendered in Table 1.

Table 1. Summary of the existing normalization techniques

This paper proposes two new normalization techniques and evaluated their performance by fusing two face recognition methods in uncontrolled environments, namely Fisherface and augmented local binary pattern (A-LBP) [813]. The description of techniques are given in Sect. 2. A short discussion of fusion techniques is found in Sect. 3. The effect of normalization techniques on recognition performance achieved by a multibiometric system is reported in Sect. 4. Finally, the conclusions are outlined in Sect. 5.

2 Proposed Score Normalization Techniques

This section proposes two new normalization techniques that transform heterogeneous scores to homogeneous scores. The new formulations of normalizing the matching scores are named as: (i) Modified-linear-tanh-linear (MLTL) which is formulated over tanh and linear-tanh-linear (LTL) normalization techniques along with the conversion of linear function into sigmoid function and (ii) Four-segments-double-sigmoid (FSDS) cleaves the regions of true and impostor scores into four segments and map each segment using piecewise sigmoid functions.

Fig. 1.
figure 1

Proposed score normalization techniques (a) Modified-linear-tanh-linear (MLTL), and (b) Four-segments-piecewise-sigmoid (MSPS).

2.1 Modified-Linear-Tanh-Linear (MLTL)

This normalization technique reinforce the strength of the characteristic resulted from tanh and linear-tanh-linear (LTL) function as illustrated in Fig. 1(a). Normalization function of it corresponds the non overlap region of the impostor scores to a constant value 0 and non overlap region of the true scores to a constant value 1. The overlapped region between \( O^{I}_{k}\) and \( O^{T}_{k}\) is mapped to a sigmoid function using tanh and LTL evaluator as,

$$\begin{aligned} n_{k_i} = \left\{ \begin{array}{ll|} \qquad \quad \qquad 0 \qquad \qquad \qquad \text{ if } r_{k_i} < min(O^{T}_{k}), \\ \qquad \quad \qquad 1 \qquad \qquad \qquad \text{ if } r_{k_i} > max(O^{I}_{k}), \\ \dfrac{1}{(1+\exp (-2 * (0.1 * z)))} \qquad \qquad \text{ otherwise }. \end{array} \right. \end{aligned}$$
(2)

where \(z = \dfrac{r_{k_i} - \mu {O^T_k}}{\delta {O^T_k}}\); and \( \mu _{O^T_k} \) , \( \sigma _{O^T_k} \) are respectively the mean and standard deviation of the true matching scores of biometric cue k . The \(n_{k_{i}}\) is the normalized scores of biometric cue k .

2.2 Four-Segments-Double-Sigmoid (FSDS)

FSDS normalization technique cleaves the regions of true and impostor scores into four segments and map each segment using piecewise sigmoid functions as illustrated in Fig. 1(b). A reference point \(t_k\) is chosen between the overlapping regions of \(O^{T}_{k}\) and \(O^{I}_{k}\). The scores between two extremities of the overlap region are mapped using two sigmoid functions separately in the range of [0, 1] towards left and right of \(t_k\) accordingly as,

$$\begin{aligned} n_{k_i} = \left\{ \begin{array} {ll|} \qquad \qquad \quad 0 \qquad \>\quad \qquad \qquad \qquad \text{ if } r_{k_i} < min(O^{T}_{k}), \\ \dfrac{1}{(1+\exp (-2 * (0.1 * z)))} \qquad \qquad min(O^{T}_{k}) \le r_{k_i} \le t_{k}, \\ \dfrac{1}{(1+\exp (-2 * (0.5 * p)))} \qquad \qquad t_{k} < r_{k_i} \le {max(O^{I}_{k})}, \\ \qquad \qquad \quad 1 \qquad \>\quad \qquad \qquad \qquad \text{ if } r_{k_i} > max(O^{I}_{k}) \end{array} \right. \end{aligned}$$
(3)

where \(z = \dfrac{r_{k_i} - \mu {O^T_k}}{\delta {O^T_k}}\) and \(p = 2*\left( \dfrac{r_{k_i} - min(O^T_k)}{{max(O^I_k) - min(O^T_k)}}\right) -1\), the \( t_{k} \) is the threshold of biometric cue k .

Table 2. Summary of the existing fusion techniques

3 Fusion Techniques

Kittler et al. [14], have developed a theoretical framework for reconciling the evidence achieved from more than one classifier schemes. These fusion rules are, such as sum, max, min, and product. Two more different fusion strategies namely strategy A and strategy B have been evaluated by Singh and Gupta in their studies [7]. In order to use these schemes, the matching scores are converted into posteriori probabilities conforming to a true user and an impostor. They consider the problem of classifying an input pattern Z into one of m possible classes based on the evidence presented by R different classifiers. Let \(\mathbf {x_{i}} \) be the feature vector provided to the \( i^{th} \) classifier. Let the outputs of the respective classifiers be \( p(w_{j}|\mathbf {x_{i}}) \), i.e., the posteriori probability of the pattern Z belonging to class \(w_{j}\) given the feature vector \(\mathbf {x_{i}} \). Let \(c\in \{1, 2, . . . , m\}\) be the class to which the input pattern Z is finally assigned. Whereas in verification (one to one map) the value of m is 2 and in identification (one to many) the value of m is \( n-1 \). The following fusion rules have been simplified by Jain et al. [15] for computing the value of class c that are given Table 2.

Table 3. Performance accuracies (%) of normalization techniques under different fusion criterions on AT & T (ORL) face dataset.

4 Experimental Results

The efficacy of the proposed normalization techniques are tested on fusion of the two face recognition methods in uncontrolled environments on AT & T (ORL) face dataset [16]. The images of this dataset suffers from the variations, such as pose, facial expression, and eye glasses. A total of 400 images are used to recognize 40 distinct individuals from the dataset. The system is trained for independent dataset composed of 40 true scores and \( 40\times 39 \) (i.e., 1560) impostor scores, whereas the test image is selected randomly from the given images for each individual and the performance is computed. The threshold value \(t_{k}\) is computed as the median of overlapped true and impostor scores. The performance of the proposed normalization technique is analyzed using equal error rate that is an error where the likelihood of acceptance is assumed to be same as to the likelihood of rejection of the people who should be correctly verified. This error is subtracted from 100 to compute the recognition accuracy. The performance of the proposed normalization techniques are also verified by the receiver operating characteristic (ROC) curves. The ROC curve is a two dimensional measure of classification performance that plots the likelihood of the true acceptance rate (TAR) against the likelihood of the false acceptance rate (FAR).

The recognition accuracies achieved by the score normalization techniques are rendered in Table 3. The accuracy values (%) for our proposed score normalization techniques i.e., FSDS (MLTL) are found better than other existing normalization techniques. For example, these values are 99.62(98.11), 97.21(96.89), 97.79(97.56), 99.62(98.11), 99.55(97.98), 99.65(98.1), respectively for the fusion techniques, such as sum, max, min, product, strategy A and strategy B.

Fig. 2.
figure 2

Receiver operating characteristic curves show the performance of different normalization techniques of matching scores obtained from Fisherfaces and A-LBP methods using different fusion criterions on AT & T (ORL) face dataset.

The receiver operating characteristic curves of the proposed score normalization techniques are plotted in accordance with their fusion techniques i.e., sum, max, min, product, strategy A, and strategy B as shown in Fig. 2. For example using sum rule, the proposed techniques FSDS (MLTL) render the TAR value of 80 % (83 %) at 0 % of FAR. The TAR value reaches to 100 % at 0.5 % (1.2 %) of FAR for FSDS (MLTL) normalization technique. These values of TAR are far better than the other existing score normalization techniques as shown in Fig. 2(a). Next, under max rule of fusion the proposed technique FSDS (MLTL) shows the TAR value of 80 % (83 %) at 0.2 % of FAR. The TAR value reaches to 100 % at 3.3 % (4.4 %) of FAR for FSDS (MLTL) normalization technique. These values of TAR are better than the other existing normalization techniques as shown in Fig. 2(b).

Similarly, using min rule of fusion the proposed technique FSDS (MLTL) renders the TAR of 83 % (87 %) at 0 % of FAR. The TAR value reaches to 100 % at 1.9 % (2.1 %) of FAR for FSDS (MLTL) normalization technique. These values of TAR are far better than the other existing normalization techniques as shown in Fig. 2(c). The proposed technique FSDS (MLTL) shows the TAR value of 85 % (83 %) at 0 % of FAR using product rule of fusion. The TAR value reaches to 100 % at 0.9 % (1.4 %) of FAR for FSDS (MLTL) normalization technique. The reported values of TAR are better than the other existing normalization techniques using product rule as shown in Fig. 2(d).

For fusion strategy A, the normalization technique FSDS (MLTL) shows the TAR of 77 % (75 %) at 0 % of FAR. The TAR value reaches to 100 % at 0.8 % (1.7 %) of FAR for FSDS (MLTL) normalization technique. These values of TAR are better than the other normalization techniques as shown in Fig. 2(e). Same results are also reported for fusion strategy B e.g., the FSDS (MLTL) normalization technique reported the TAR of 83 % (80 %) at 0 % of FAR. The TAR reaches to 100 % at 0.7 % (1.3 %) of FAR for FSDS (MLTL). The reported values of TAR are found better than the other normalization techniques using the fusion strategy B as shown in Fig. 2(f).

The recognition accuracy results of the suggested techniques i.e., MLTL and FSDS indicate that these score normalization techniques can contribute a peculiar role in the design of a robust multibiometric system.

5 Conclusion

This paper has presented two novel techniques of score normalization namely, modified-linear-tanh-linear (MLTL) and four-segments-double-sigmoid (FSDS). The performance of these proposed score normalization techniques has been evaluated and the fusion of face recognition methods Fisherface and A-LBP. The performance of the proposed score normalization techniques have found better than the existing min-max, tanh and linear-tanh-linear (LTL) normalization techniques. This evaluation of score normalization techniques of matching scores insinuates that the proposed techniques may play an important role in evaluating the performance of a multibiometric system.