1 Introduction

In the current digital age, the ubiquity and the wide reliance on the modern technologies has helped in facilitating counter-theft procedures. Finding such counter-theft solutions is becoming a more stressful issue with the current increase in the crime rates. One technology example is to utilize the electronic fingerprint verification systems that can be used with different automated systems such as ATMs, smart buildings and transportation systems, etc. Because of this, automated fingerprint identification is receiving continuously growing interest from both academic and commercial parties.

The importance of focusing on fingerprints, as a main biometric tool to uniquely identify people, stems from many reasons. The first one is that no two individuals (even identical twins) have the same fingerprints. Another reason is that fingerprints do not change with age. Finally, each finger of a single individual has a different fingerprint. This fact can be utilized to identify not only the person who touched a certain object, but also the finger used by that person.

As shown in Fig. 1, a fingerprint is formed from the following parts.

  • Ridge Endings: the terminals of the ridges.

  • Bifurcations: the point of dividing ridge into two branches.

  • Dots and small ridges.

  • Islands: random point with no connectors.

  • Pore: a white point surrounded by black point.

  • Bridges: small ridges joining two longer adjacent ridges.

  • Crossovers: two ridges which cross each other.

  • Core: center of the fingerprint.

  • Deltas: points, at the lower part of the fingerprint, surrounded by a triangle.

Fig. 1
figure 1

A Fingerprint clarification

Fingerprints can be classified according to their shapes into three main types: Whorl, Loop and Arch. Each class has its ridge flow and pattern. Figure 2 shows these three types with their existence percentages. Loop fingerprint are the most common with a percent of 60–65% of all fingerprints. Arch fingerprints, which are rare, account for only 5% [58].

Fig. 2
figure 2

Fingerprint types

A typical Automatic Fingerprint Identification System (AFIS) usually consists of many stages that include: image acquisition, fingerprint segmentation, image preprocessing (enhancement), feature extraction, and matching result [17]. Below, we briefly define each step.

  • Image Acquisition: The stage of image acquisition is one of the most important factors that contribute in the success of the recognition system. Here, there are several factors to be taken into account before choosing the device that will be used to capture fingerprints in terms of performance, cost and size of the resulting image. In this study, we focus on the fingerprint capture phase only.

  • Segmentation: The basic idea in this step is to separate the image of the finger from the background. It is an initial step of any recognition systems [31, 39, 56, 59].

  • Preprocessing: This phase includes some modifications to the image with the aim of enhancement. This would increase the accuracy of the recognition system.

  • Feature extraction and Person Verification: At last, similar fingerprints are retrieved. So, only this set of fingerprints (source) are compared with the target fingerprint [7, 18, 44, 57].

Fingerprints are acquired in different ways. For non-automated (or non-online) systems, fingerprints are usually acquired by stamping inked fingertips on a special paper. If these fingerprints need to be stored on a digital device, an image of the paper is taken using a scanner for example. On the other hand, fingerprints are acquired using fingertips scanners for automated systems. Scanners may use different technologies such as optical, capacitive, thermal, and ultrasonic. These scanners usually need the fingertip to be pressed or dragged across some surface. This may reshape or squeeze the resulting fingerprint which may lead to incorrect identification. Also, this makes it impossible to cross use different finger tips scanners in acquiring and verification process as different scanners may results in different reshaping results.

As most modern automated systems are equipped, or can be easily equipped, with high quality digital cameras, a recent direction of research started to study the possibility of acquiring fingerprints using these cameras. According to the studies in [2, 19, 30, 40, 43, 44, 55], this approach is promising and the results are acceptable.

To design different techniques and study the impact of different factors related to automated fingerprint recognition systems, a large fingerprint image set is needed [23]. There are many factors that may affect the accuracy of the recognition process. Most digital cameras perform JPEG compression on the images they capture before storing them. JPEG compression can be applied using different factors, called JPEG quality, which is represented by a number between 1 and 100. The higher the JPEG quality factor the better the resulting image clarity and the larger the image size. If compression is not applied, image sizes become very large and storing these images becomes very expensive. For example, the size of an image captured by 16 mega pixel camera (that represents a pixel with 8 bits) is 16 megabytes. While compressing this image using JPEG compression with quality factor 50 may result in an image of size 200 kilobytes. Another factor that may impact the fingerprint recognition accuracy is the complexity of the background in the image.

To study the impact of any image quality factor on the finger recognition process, a very high quality fingerprint image set is needed. Preferably, the images need to be in what is known as their raw format. If this image set is available, we can easily apply any image distortion factor, with different levels, and calculates the accuracy of recognition algorithm using images distorted with these different levels. This study can be used to find a relationship between distortion levels and recognition accuracy. For example, we can compress the fingerprint images with different JPEG quality factors to study the relationship between the compression level and the recognition accuracy.

In the study [1], a survey was conducted about the available fingerprint image sets and the acquisition devices used to collect them. The conclusion in that study is that the available fingerprint datasets acquired using digital cameras are all taken with built in degradation factors due to specific lighting, different background, low camera quality, image compression, etc. In other words, the available fingerprint databases are far from being suitable for testing the impact of different customizable degradations. In this work, we build a digital camera acquired fingerprint set in perfect conditions that can be customized easily and can be used for testing the impact of different degradations on the accuracy of different fingerprint recognition algorithms. We also experiment with a fingerprint recognition system to study the impact of some factors such as image compression and subject gender.

The rest of this paper is organized as follows. Section 2 discusses the related works on the impact of image compression on various recognition tasks. Section 3 gives an overview of the most common devices used for online fingerprint acquisition. In Section 4, we describe briefly some of the most common fingerprint databases, while a detailed description of our dataset is given in Section 5. The fingerprint recognition system we use is discussed in Section 6 and the experiments and their results are discussed in Section 7. Finally, a conclusion of the paper along with future directions of this work are given in Section 8.

2 Related work

Image quality is very important for the biometric identification [8]. Unfortunately, usually, the better the image quality, the larger the image size which may require more storage space. In this section, we discuss a number of studies concerning the importance of image quality and image compressing operations and their impact on system performance.

Many studies in the literature analyzed digital image compression impact on image quality in general [45, 52]. On the other hand, only few studies analyzed image compression impact on biometric recognition algorithms other than fingerprint recognition [8, 11, 36].

The iris of the eye is one of the biometric features that are commonly used. Iris recognition depends entirely on the quality of the captured iris image. In 2008, Daugman, John, and Cathryn Downing [8] studied the effect of compressing images on the performance of iris recognition systems. They presented an approach that combines region-of-interest isolation with JPEG and JPEG2000 compression at intense levels. Results showed that it is possible to compress iris images to less than 2000 bytes with little impact on the performance identification. Based on previous ratios change in order of bits in the iris code templates are small. According to the study, the impact was between 2% to 3%. In the summary of their study, they concluded that the compression ratio 150 : 1 from original image does not affect the system’s performance. Image features and details remain without being affected negatively.

In [38], the authors presented the impact of age to identify people through fingerprints, they divided the images into four different age groups, and they concluded that the details extracted depend on the age group. With each age group, the skin is different in terms of smoothness and dryness. Age group affected the performance of the system significantly. The results showed that the images of people aged 60 years and older achieved better performance. The researchers used an optical sensor to collect their database. Moreover, they used an available commercial assessment tool that counts number of minutiae.

The conclusion in [38] is also confirmed by Modi at el. in [37]. They presented research about measuring the impact of the quality of the fingerprint image on identification of two samples of two different ages: the first sample age is between eighteen and twenty-five and the second sample age is 62 and more. In their study, they used two matching algorithms; Neurotechnologija VeriFinger and bozorth3. Moreover, they used fingerprint Image Software (NFIS) to analyze the quality of the images in there experimental results. False Matches were used for confirming that the old sample fingerprints has the highest number of minutiae compared to the fingerprints of the younger sample.

Other studies were focused on comparing different methods for compression images. In particular studies, they designed a fusion system that combined more than one way to identify individuals, like the combination between the face and the fingerprint that presented in [11, 36]. These studies measured the impact of compression on the images when using many common compression algorithms such as jpeg and jpeg2000. The main difference between their work and ours is that they were interested in comparing between different compression methods only.

3 Fingerprint acquisition devices

This section gives an overview of the most common devices used for online fingerprint acquisition. Many technologies have been utilized in these devices. Namely, optical, capacitive, thermal, and ultrasonic.

3.1 Optical sensor

Optical fingerprint devices are the oldest and the most commonly used devices to capture fingerprints. Figure 3aFootnote 1 clarifies how this sensor works. The fingerprint image is captured by placing a finger on one side of a prism, a light source that emits light vertical on the second prism side, and a camera that is placed in front of the third prism side. The light collected by the camera is then converted into a digital signal where bright areas reflect the valleys and dark areas reflect the ridges. This sensor was used to acquire a large number finger print databases, such as FVC2000 (DB1, DB3) [35], FVC2002 (DB1 and DB2) [33], FVC2004 (DB1 and DB2) [34], FVC2006 (DB2) [5]. Figure 3b is an example of a commercial optical sensor. The main advantages of optical sensors are temperature insensitivity, low cost, and high resulting fingerprint resolution. On the other hand, it suffers form many disadvantages. As optical sensor depends on light reflection, as a result, its functionality may suffer from lighting environment. Also, optical sensors must be of a certain large size to be able to generate a high quality image [50]. Moreover, as the finger should be placed in contact of some plate (plastic or glass), the residue of the previous fingerprint on the sensor may impact the quality of the current fingerprint image as they may overlap. Furthermore, this residue can be used to steal the fingerprint.

Fig. 3
figure 3

Optical technology

One of the commercial examples that uses this technology is shown in Fig. 3b is the SecuGen Hamster IV.Footnote 2 it is one of the common devices in fingerprint systems, which have been certified by the FBI and STQC. Beside SecuGen Hamster IV, other optical devices were used to collect the IIIT-D Multi-sensor Optical and Latent Fingerprint [49].

3.2 Capacitive sensor

This device is based on the principle of measuring the electric capacity of the human skin. Fingerprints are placed on a very thin plate that contains a set of micro-capacitors. Skin of a fingerprint contains ridges and valleys. The capacitive sensor basically measure the capacity of the skin of the finger. The capacity of ridges is different form the capacity of valleys because the capacity depends highly on the distance of the skin from the micro-capacitor plate. A simple prototype of a capacitive sensor is shown in Fig. 4a.Footnote 3 This sensor was used to capture a number of fingerprint datasets that includes FVC2000 (DB2) [35] and FVC2002 (DB3) [33].

Fig. 4
figure 4

Capacitive technology

An advantage of capacitive sensors are small in size and therefore they are cheaper to produce. They also consumes low energy [24]. This advantage is particularly useful in the battery powered devices. It also insensitive to environmental effects such as sun light. Many manufacturers, such as Sony, Fujitsu, Hitachi, and Apple, have adopted this technology in their fingerprint recognition systems.

3.3 Synthetic fingerprint generation

This method automatically generates artificial fingerprints based on a model. This model usually is controlled by a set of parameters which can be changed to produce different fingerprints in terms of clearness, area, centered or not, rotation, the difficulty of the recognition of the fingerprint, and many more. This method can be used to build very large fingerprint databases in short time with almost no cost. Also, the generated artificial fingerprints may be tuned to simulate natural fingerprints. This method was used in some fingerprint databases such as FVC2000 (DB4) [35], FVC2002 (DB4) [33], FVC2004 (DB4) [34], FVC2006 (DB4) [5].

3.4 Thermal sensor

This sensor is based on pyro electric material. This material has the ability to change temperature differences to different voltages. By the difference between the temperatures of the ridges and the valleys, the voltage generated for each one is different from the other. One of the most common thermal sensors is “Atmel FingerChip” which is used to acquire many well known fingerprint databases. Namely, FVC2004 (DB3) [34] and FVC2006 (DB3) [5].

This technology is not very common commercially. One of the difficulties faced when dealing with this technology is that the thermal image should be captured in a very short time. That is because when a finger is placed over the pyro electric material, the temperature difference between the areas under the ridges and the valleys will be initially detectable. Eventually, after a short period of time, the temperature difference vanishes as the adjacent areas heat each other. As this method depends on body temperature, it is very hard to be deceived (Fig. 5).

Fig. 5
figure 5

Thermal technology

3.5 Ultrasound sensor

This device is based on the principle of sending high frequency sound waves, and then capturing the reflection from the finger surface. A clarification of how this technology works is shown in Fig. 6. The advantage of this technology is its ability to pass through dust, dirt, and skin oil that may cover the finger surface. For this reason, fingerprint ultrasound sensing may be used to capture high quality images if if the finger is dirty. On the other hand, this technique has many disadvantage. For example, it is very expensive and may take long time to capture the fingerprint image. QualcommFootnote 4 are utilizing ultrasonic technology in their new processor (snapdragon 820) security features.

Fig. 6
figure 6

Clarification of how ultrasonic sensors wok

3.6 Digital cameras

Many digital camera types are used in the literature of fingerprint systems. We categorize them into webcams [2, 19, 20, 40, 43], mobile phone cameras [9, 26, 28, 30, 44], and advanced digital cameras [22, 55]. Webcams are digital cameras that are usually low in resolution and very cheap in price. These cameras comes usually built-in personnel computers for video chatting purposes. Because these cameras have low resolution, it is very hard to get the fingerprint details from an image. Therefore, many researches proposed many image pre-processing techniques to overcome this problem.

In the last few years, mobile phones cameras have witnessed a huge improvement which have enabled them to be used to acquire fingerprints. Current resolution of mobile phones have reached 8 mega pixels on average. However, there are a set of criteria that should be considered when taking fingerprint images using these devices, such as lighting, distance between the finger and the camera, the appropriate focus that should be chosen, and the image preprocessing steps.

Advanced digital cameras, also, have been studied and used to acquire fingerprint images. These cameras produces very high resolution images. In these cameras, image acquisition is highly flexible. They can also apply advanced image enhancement and compression algorithms.

4 Fingerprint databases

In this section, we describe briefly some of the most common fingerprint databases:

FVC2000 database

is used in the First International Fingerprint Verification Competition. The database consists of four parts; DB1, DB2, DB3 and DB4 [35]. DB1 and DB2 contain images of 110 fingers with eight impressions per finger (880 images in total). The images of up to four fingers per each participant where taken. Fingerprints of the fore and the middle fingers of both hands are taken. Half of the fingerprints are for males. The ages of the individuals in these two databases are 20-30 years old. DB1 was acquired using a low-cost optical sensor (Secure Desktop Scanner) while DB2 was acquired a low-cost capacitive sensor (TouchChip by ST Microelectronics). According to [35], these two databases cannot be used to evaluate the accuracy of the biometric algorithms. That is because no cautions were taken to guarantee the quality of the images and the plates of the sensor were not systematically cleaned after each use In DB3, the fingerprints are acquired from 19 volunteers, of whom 55% are males. The images of six fingers are taken from each participant (thumb, fore, and middle of left and right hands). The ages of the individuals are between 5 and 73. DB3 images were acquired using an optical sensor (DF-90 by Identicator Technology). DB4 was automatically generated using synthetic generation.

FVC2002 database

was used in the Second International Fingerprint Verification Competition [33]. It consists of four parts; DB1, DB2, DB3 and DB4. The number of images is the same as in FVC2000. In this database, four different types of technologies were used to acquire the fingerprints in the four different databases (i) an optical sensor “TouchView II” by Identix, (ii) an optical sensor “FX2000” by Biometrics, (iii) a capacitive sensor “100 SC” by Precise Biometrics, and (iv) synthetic fingerprint generation. 90 individuals participated in this database collection. In DB1, DB2, and DB3, images of different 30 participants were collected. Images of 4 fingers (the fore and the middle fingers of both hands) of each participant were collected. No any extra effort were taken to guarantee the quality of the images as the perpos of these images are just to be used in competition. DB4 was synthetically generated.

FVC2004 database

was used in the Third International Fingerprint Verification Competition [34]. It consists of four parts DB1, DB2, DB3 and DB4. Each one of first tree parts contains images of different 30 persons. These persons are 24 years old on average. Fingerprints of the fore and the middle fingers of both hands were taken from each person. For each part, 120 fingers were considered with 12 impressions per finger for a total of 1440 images. In this database, four different techniques of of fingerprint acquisition were used: (i) an optical sensor (CrossMatch V300) for DB1, (ii) an optical sensor (URU 4000) by Digital Persona for DB2, (iii) a thermal sweeping sensor (Atmel FingerChip) for DB3, and (iv) synthetic generation (SFinGe v3.0) for DB4.

FVC2006 database

was used in the Fourth International Fingerprint Verification Competition [5, 12] consists of four parts: DB1, DB2, DB3 and DB4. Each one of these four parts contains the images of 150 fingers, 12 impression of each finger. The four parts were acquired using Capacitive (Electric field) sensor, Optical Sensor, Thermal sweeping Sensor, and SFinGe v3.0, respectively. This fingerprint data set is a part of a large biometric database collected within the European Project BioSec [4].

CASIA-FingerprintV5 database

[6] contains 500 subjects. Each subject contributed 40 fingerprint images of his thumb, index, middle, and ring fingers of both his left and right hands, with 5 images per finger. Different finger images defer in rotation or pressure. The database contains 20,000 fingerprint images in total. All fingerprints are stored as 8 bit gray-level BMP images. The finger print images were captured using the URU4000 optical fingerprint sensor in one session.

IIIT-D multi-sensor optical and latent fingerprint

[49] contains 19200 images acquired from 100 subjects using different sensors. Table 1 contains the main details about this database.

Table 1 IIIT-D multi-sensor optical and latent fingerprint

IIIT-D Latent database

[47] consists of images of all 10 fingers for 15 subjects. Latent fingerprints were taken using Canon EOS camera with resolution 4752x3168. Fingerprints were taken with different effects (dry, wet and moisture skin) as well as different backgrounds (tile and ceramic plate) with a total number of 1046 images. Those images are freely available.

Low-cost Equipment database

[40] consists of 400 fingerprints of 6 participants captured using two webcams: Microsoft Lifecam VX3000 and Philips SPC630NC.

Smartphone Camera database

[44] consists of 100 different fingerprints taken from 25 participants. Fingerprints of the index and the middle fingers of both left and right hands were taken. Fingerprints were captured using iPhone 4, Samsung Galaxy s1, and Nokia N8 mobile phones.

Camera-based fingerprint database

[28] consists of fingerprints of 13 different subjects. Three samples were taken for each finger of the participating subjects. The total is 390 fingerprint images. The images were captured using Nokia 6700 classic and HTC magic mobile phones.

5 Description of our database

As shown in Section 4, none of the available fingerprint databases are suitable for our study goal which is to study the impact of digital image compression on fingerprint recognition. For this reason, we built our own high quality digital fingerprints database. To capture high quality fingerprint images, we used a digital single-lens reflex (DSLR) to capture the images and we made sure that the camera stores the images as raw images to avoid any compression. We use a Canon EOS 750D (also known as Canon EOS Rebel T6i) camera, which is fast and easy to use. Canon t6i provides high-resolution options and auto focus. In addition, it is equipped with Wi-Fi technology, which enables the users to easily access the captured images.

Our database, the Raw Fingerprint Image Database (RFID), contains (6,000) fingerprint images of 500 high school students and teachers, 90% of them aged between 13 and 18. Half of the volunteers were males, and the other half were females.

We have designed a high table to help the volunteer and the photographer fix their hands on and to install on different wooden backgrounds. The volunteers are asked to fix their hands on this high table to guarantee the quality and the clearance of fingerprints in the image. Twelve images were taken for each volunteer. Eight images were taken for the thumb and one image was taken for each of the index, middle, ring, and pinkie fingers of the volunteers’ right hand. The images of the index, middle, ring, and pinkie fingers are captured in front of a wooden white background plate. For the thumb, we took eight pictures, six of them with a white background and two of them with colored backgrounds. The two colored backgrounds were blue and violet. Table 2 summarizes the fingerprint capturing environment parameters and Fig. 7 clarifies the environment of our database collection.

Table 2 Environment of the used camera in database collection
Fig. 7
figure 7

Database collection environment

The images are stored in CR2 format. We use a file naming convention that encodes the number and gender of the volunteer and the name of the finger. Figure 8 shows file naming convention in details. For example, the file (P016_MTh8.CR2) represents the image of the thumb with colored background for the 16th male volunteer. Figure 9 below shows some examples of the annotated dataset.

Fig. 8
figure 8

File naming convention

Fig. 9
figure 9

Examples of the annotated dataset

5.1 Difficulties faced while collecting data

The database collection process lasted for three months (from December 2015 until March 2016). During this period, many challenges were faced. Most notably are the following:

  • The process of capturing clear images of fingerprints required many trials because of involuntary movement of the finger.

  • Dealing with uncompressed images requires large storage capacity, where one image size is greater than 27 MB.

  • Adjusting the lighting and distance requires considerable effort and some experience.

  • There are people who have soft fingerprints, which makes it difficult for us to capture a detailed image of their fingerprints. This forced us to re-capture some images several times.

  • Because most of volunteers are students, their fingers are sometimes dirty and have ink stains. This resulted in poor fingerprint image quality. We used wet wipes to clean their fingers before image acquisition.

  • The collection of the data took place in a county in the Middle East and North Africa (MENA) region. Unfortunately, this region has witnessed several notorious police states, which created fear in minds of the public about security-related issues. Collecting fingerprints is something that would always raise serious suspicions. We had to explain thorough our project and its scholarly value. We also presented the subjects with firm assurances that there is no hidden agenda or malicious intent behind this process. Some were convinced with our arguments while others simply refused to participate.

6 Our fingerprint recognition system

To study the impact of image compression on fingerprint recognition, we must have a fingerprint recognition system in hand to perform some experiments. In this section, we describe the fingerprint recognition system that we put together for this purpose.

A system that identifies people through fingerprint, must perform several successive steps that are mainly (1) converting the image jpeg, (2) convert the image to grayscale, (3) segment the fingerprint from the image, (4) converted the finger image into binary image, (5) extract features from the finger image, and (6) compare features of the image with other images in the system to find a match.

Any automated identification system is designed in two main phases, training and testing. In the training phase is about processing a database of fingerprint images and represent them with some useful features. This phase passes through several steps which are (1) database collection, (2) image preprocessing, (3) feature extraction, and (3) storage of these features The second phase is called testing, where a query fingerprint image (the testing image) is put through the same preprocessing and feature extraction steps of the training phase. Then the query image is compared with all the database images (training images). If a query image matches one of the database images with a match that is above some threshold, the result query image is considered a match. Otherwise, the system will reject the test image. This step is called identification/recognition.

Fingerprint matching in our framework is not carried out using direct image-to-image comparison since this will need long time and large memory size [25]. Furthermore, dirt or distortion in images might lead to a mismatch between two images of the same fingerprint. Hence, this method is not practical; instead, the matching stage in most fingerprint systems is based on certain marks and features in the fingerprint [46].

If there are two fingerprints with the same terminations and bifurcations, so that, they form the same shape and the same dimensions, then, there is a high possibility that they belong to the same person. Recognition systems do not need the entire minutia recorded for both persons, but a certain number of detail is enough to be compared. This number of details varies depending on the fingerprint identification system and the quality of the original image. In the following sections, we illustrate in details our methodology starting from converting the captured raw files to JPEG images with different QFs and ending with determining the best QF that produces the highest accuracy. Figure 10 shows the framework stages, which will be explained in details next in this section.

Fig. 10
figure 10

Our framework phases

6.1 Building the database

For an individual to be recognized by an automated fingerprint recognition system, he should be first registered to the systems. Figure 11 illustrates the registration of an individual in the proposed system. The following subsections details the registration process.

Fig. 11
figure 11

Illustration of the templates registration

6.1.1 Image conversion and compression

At the registration stage, the fingerprint images are captured by a Canon camera and saved without any compression as raw images. Particularly, in a the used Canon camera version, raw images are saved with CR2 extension. CR2 is also called digital negative. Specialist photographers prefer to use this type of file, because it retains many details of the image. However, it is not widely used because the image size is very large (in our database, a single image reaches up to 30 MB in size). This makes it difficult for such images to be stored and uploaded to the Internet. The next step is to convert the CR2 images to JPEG using different QF ranging from 1 to 100. JPEG stands for Joint Photographic Experts Group, which is usually used as a name of the a standard for still image compression algorithm. It compresses the image to ensure the preservation of quality and reduce the storage size of the image. It is widely used in many imaging-related applications, where smaller sizes help in storing, transferring and uploading the images on the Internet [21].

To conduct this conversion, we used a special library for this purpose called FreeImage. FreeImage is an open source graphics library, written in C language and used in image processing; it is compatible with Linux, OSX and Windows. It is widely used due to its speed and simplicity. Floris van den Berg created it and it has been under development for more than 9 years. It is worth mentioning that this library is free and not included within the automatically installed libraries in different programming languages or instructional machines. Therefore, it needs to be added to our source code, to make sure that it is compatible with our system.Footnote 5 Examples of converted images in different QFs are shown in Fig. 12.

Fig. 12
figure 12

Fingerprint images after conversion in different qualities

6.1.2 Segmentation

After database collection, each image must be segmented into two regions: foreground (finger region) and background. This step is called segmentation. There are several methods for segmentation. One method separates the object from the background using the color information, was first proposed by Hassanat et al. in 2015 [15]. This method is simple, efficient and deals with digital color images and thus we rely on it for our segmentation phase. The method utilizes a novel fusion of color space models based on neural networks to separate the foreground from the background. In the training step, users click ten times on the foreground (finger region), and another ten clicks on the background. The system stores the pixels that are clicked on and other eight neighboring pixels. After that, it saves these pixels to be the input layer of the neural network. Consequently, the neural network learns how to identify the ROI from the background. This way, we identify the finger area, and then store the new image, which contains only the finger area. Figure 13 show a sample output of this stage.

Fig. 13
figure 13

Image segmentation result

6.1.3 Binarization

After taking the segmented images, we must recognize the minutia (ridges and valleys), by converting the image into two colors, white and black. This step is called binarization. One of the important methods for binarization of colored images is Khalil’s method [26, 28]. This algorithm performs image enhancement as follows. First, it converts the colored image to grayscale image to reduce the problems of different illuminations in the image. This problem results in non-uniform components in the image. After converting the image to grayscale as shown in Fig. 14a, a simple normalized image is required as shown in Fig. 14b. It utilized the algorithm proposed by Hiew et al. [16]. The normalization is a pixel-wise procedure. According to Khalil’s method, the normalization is used to reduce the illumination problems through calculating the variance and mean as shown in (1). It is worth noting that the normalization step does not affect the structure of the minutiae whether it is ridge or valley.

$$ G(x,y)=\frac{f(x,y)- m_{f} (x,y)}{ \sigma_{f} (x,y)} $$
(1)

where, f(x, y) expresses the grayscale fingerprint image, mf(x, y) is the mean of f and σf(x, y) expresses the standard deviation of the segmented image. The spatial smoothing is used for evaluating mf and σf. After that, two main steps are performed. The first is Discrete Fourier transform (DFT)-based fingerprint ridge analysis, where the ridge analysis method is utilized to detect the ridge area. The ridge information is extracted by performing a high-pass filter based on the DFT on input images. The second step is the Short Time Fourier Transform (STFT) method, which is used for image enhancement. The first step consists of four sub-steps as follows. At first, a Gaussian high-pass filter G(u, v) is applied as a scalar product of the matrices on the input normalized images. It was created through a padding size (P x Q) as shown in Fig. 14c, and filter width of Fourier transform according to the following equations:

$$ D_{0}= 0.05 \times P $$
(2)
$$ D(u,v)=\sqrt{(u-\frac{P}{2})^{2} +(v- \frac{Q}{2} )^{2} )} $$
(3)
$$ G(u,v)= 1-e^{\frac{-D^{2}(u,v)}{2{D_{0}^{2}}}} $$
(4)

where, P and Q: padding size block. D0: is the cut-off frequency, D(u, v) is the distance between coordinates (u, v) and the center of frequency at (0, 0).

Fig. 14
figure 14

Binarization stages

In these equations, the size of input image is used to determine the padded size as 2 × M × N, where M and N is the size of input image. The appropriate width of the filter generated during the experimental results in [10] which equals 5% of the padding size, P. The Fourier domain is used to transform the input image F(X, Y ) into the frequency domain (DFT domain) (see (6)). Results from the previous steps are reversed into the spatial domain using the matrix H(u, v) (see (7)).

$$ F(u,v)= \sum\limits_{x = 0}^{M-1}\sum\limits_{y = 0}^{N-1}f(x,y) e^{-i2\pi(\frac{ux}{M}+ \frac{vy}{N}) } $$
(5)
$$ H(u,v)=F(u,v)G(u,v) $$
(6)
$$ h^{\prime}(u,v)= \sum\limits_{x = 0}^{M-1}\sum\limits_{y = 0}^{N-1}H(u,v) e^{i2\pi(\frac{ux}{M}+ \frac{vy}{N}) } $$
(7)

where M and N is the original image dimension.

In the fourth and final phase of this step, the image is cropped to dispose of the previous padding. Then a simple threshold is used to extract the ridge information as defined in (8) and (9). Algorithm 1 illustrates the entire algorithm.

$$ I_{ridge} = \{I_{ridge}(i,j) \in h^{\prime}| i = 1,2, ..., I_{h^{\prime}}, j = 1, 2, ..., I_{W} \} $$
(8)
$$ I_{ridge} = \left\{\begin{array}{ll} 1,& I_{ridge}(i,j)>0.005 \times \max(I_{ridge})\\ 0,&otherwize \end{array}\right. $$
(9)
figure e

According to Khalil’s method, the first step is to identify the ridge area (as shown in Fig. 15, whereas the second step is to enhance the image obtained from the ridge analysis step. The step of enhancing the image is required to obtain highly accurate results. For this reason, the Short Time Fourier Transform (STFT) analysis is used to improve the image. At first, the image that contains information of the ridge is divided into overlapping windows. The purpose of this overlap between windows is to preserve the coherence between the lines of the ridges and to get rid of the cluster impact that naturally happens because of applying the window-by-window method. However, there is a procedure preceding the application of the STFT analysis, which is the Fourier spectrum of the block F(r, 𝜃), where 𝜃 is the orientation. is a random value. r is ridge frequency, r is presumed to consist of random values. Then, the (ROI) is identified by calculating two functions: the marginal density function p(𝜃),p(r) and the probability density function p(r, 𝜃).

$$ P (r,\theta)= \frac{\mid F(r,\theta) \mid^{2}}{{\int}_{r}{\int}_{\theta}\mid F(r,\theta)\mid^{2}} $$
(10)
$$ P(r) ={\int}_{\theta}P(r,\theta)d \theta $$
(11)
$$ P(\theta) ={\int}_{\theta}P(r,\theta)d r $$
(12)
Fig. 15
figure 15

Ridge image resulted from Algorithm 1

After applying the STFT analysis in small windows one by one, the results from this procedure are a Ridge Orientation Image, Energy Image (EI), and Ridge Frequency Image (RFI). The next step is smoothing the orientation image to get the coherence image based on the vector averaging as in (10) where 𝜃 is the orientation. Moreover, r denotes ridge frequency. Coherence image is responsible for the elimination of false spots, which appear because of the ridge flow ends at the small window boundary. As shown in the following equations.:

$$ E(\theta) = 0.5tan^{-1}\left( \frac{{\int}_{\theta}P(r,\theta)d\theta}{{\int}_{r}P(r,\theta)dr}\right) $$
(13)
$$ E(r) =\int{\int}_{r}P(r).r dr $$
(14)

To get the region mask, the common thresholding Otsu’s method is applied.

$$ E(x,y)= \log \{{\int}_{r} {\int}_{\theta}\mid F(r,\theta)\mid^{2} \} $$
(15)

The ridge orientation, the coherence and the frequency images are used to generate two filters; angular filter and radial filter, see (16) and (17). A filter window of 16 × 16 is used to smooth the image. Eventually, the image that the filters have been applied on, will be rebuilt with a new enhanced image as shown in Fig. 16.

$$ H(r)= \sqrt{\frac{(rrBW)^{2n}}{(rrBW)^{2n}+(r^{2}+r^{2}bw)^{2n}}} $$
(16)
$$ H({\Phi})= \left\{\begin{array}{ll} \frac{\cos^{2}\pi\mid{\Phi}-{\Phi} c\mid}{2{\Phi} BW},& \mid{\Phi}-{\Phi} c\mid \le {\Phi} BW\\ 0, &otherwize \\ \end{array}\right. $$
(17)

where, rBW is the radial bandwidth, ΦBW is the angular bandwidth, and ΦC is the mean orientation. Algorithm 2 summarizes the steps of converting ridge image to enhanced image:

figure f
Fig. 16
figure 16

Enhanced Image Resulted From Algorithm 2

6.1.4 Fingerprint feature extraction technique

After converting the image into a binary image, the next step is window direction estimation, followed by implementing some morphological operations. The previous two steps are to extract the (ROI) [27]. For estimating the window direction, the image is divided into blocks, where the size of a block is WxW. According to [3, 27], the best values of W is 16. To carry out these steps, in each window in an image, the gradient values are calculated on the X-axis and the Y-axis by using two Sobel filters for all pixels. After that, the value of the least square approximation is calculated based on the following equation:

$$ tan2 \beta=\frac{2\sum\sum (g_{x}\times g_{y})}{\sum\sum ({g_{x}^{2}}-{g_{y}^{2}})} $$
(18)

where, gx and gy are the gradients taken along x-axis and y-axis, respectively.

The orientation of ridges is estimated according to the equation above. The equation is easy to interpret where the gradient values gx and gy represent the cosine and sine values, respectively. So, the tangent value of the window direction is estimated using the following formula:

$$ tan2\theta=\frac{2sin\theta cos\theta}{cos^{2}\theta - sin^{2}\theta} $$
(19)

After estimating each window, any window that does not carry any information around the edges and grooves is rejected based on the following formula:

$$ E = \frac{2\sum\sum (g_{x}\times g_{y})+\sum\sum ({g_{x}^{2}}-{g_{y}^{2}})}{W\times W\times \sum\sum ({g_{x}^{2}}-{g_{y}^{2}})} $$
(20)

The output of these steps is called the direction map that is shown in Fig. 17.

Fig. 17
figure 17

A binerized fingerprint (left) along with its direction map (right)

The next step is to perform some morphological operations. These operations are of the most common operations in image processing. They are mainly two operators; Open and Close. Both of these operators depend on the sequence of two operations. The first is called erosion which is an operation used to shrink or thin objects in binary images. The second is called dilation, which is an operation used to grow or thicken objects in binary images. The opening operation uses erosion then dilation, but the closing operation uses dilation then erosion. The fingerprint area is obtained in addition to its boundaries, which are obtained by the subtraction of the two previous operators: open and close. Figure 18 shows the output of this step.

Fig. 18
figure 18

Region of Interest and its boundaries

The next step is thinning. This step is used to remove small parts of the pixels. These redundant pixels of ridges are eliminated by dividing the image into windows with M × M matrix, where M is dimension of matrix M = 3 [13]. This operation plays an important role to get the best accuracy [14]. In each window, the sliding neighborhood algorithm used to calculate the centroid. The centroid pixel location is changed for each window. After finding the centroid, the summation is performed for each pixel of the thinned matrix to obtain a new matrix consisting of minutiae points. Then subtracts 1 from the computed sum, which gives the value of the pixel. After this step, delete the redundant and singular pixels that may arise after the previous step. These conditions can be summarized as follows:

  1. 1.

    Divide the image into 3 × 3 blocks.

  2. 2.

    Thin each line of ridge to its center pixel.

  3. 3.

    Remove redundant and singular pixels.

Many algorithms can be used for this task. According to the study presented recently in 2015 by Hannah et al. [13], Zhang-Suen’s Algorithm and Hall’s algorithm are two of the most common techniques. According to [13], the former algorithm is more accurate and is more efficient than the latter. This is why we use Zhang-Suen’s algorithm in this work. Algorithm 3 presents the basic steps of the thinning algorithm. Any pixel that lies on the edge of the foreground regions is converted to its neighboring background. Also, any pixel that has more than one foreground neighbor is removed along the ridge line. This procedure repeated until we reach convergence.

figure g

The next step is removing the H breaks and spikes. This step is considered as another morphological operation to eliminate some isolated points and spikes, which are usually called the H breaks, where any single-pointed ridge or single-pointed break in a ridge is considered as noise and should be removed. This step can be coded in MATLAB using the following commands;

bwmorph (binaryImage, ’hbreak’, k)

bwmorph(binaryImage, ’clean’, k)

bwmorph(binaryImage, ’spur’, k)

Where bwmorph applies a specific morphological operation to the binary image, binaryImage is the image name, ’clean’ is th option to remove isolated pixels (individual 1’s that are surrounded by 0’s), such as the center pixel in this pattern, ’spur’ to Removes spur pixels.

For example:

0 0 0 0

0 0 1 0

0 1 0 0

1 1 0 0

becomes

0 0 0 0

0 0 0 0

0 1 0 0

1 1 0 0

‘hbreak’ is used to unconnected H-connected pixels. For example:

1 1 1

0 1 0

1 1 1

becomes

1 1 1

0 0 0

1 1 1

The next step is marking minutiae points. The output image from the previous step is used to determine the minutiae. Simply a thinned image is divided into windows of size 3 × 3. If the value of the central pixel equals to 1 and surrounded by at least three neighbors with value 1, then it is considered as bifurcation. If the value of the central pixel is equal to 1 and surrounded by only one neighbor with value 1, it is considered as termination. This technique is denoted by Crossing Number (CN), and it is applied commonly to extract the minutiae [53, 54].

The CN is calculated according to following equation:

$$ CN(Pxl) = \frac{1}{2{\sum}_{i = 1}^{8}\mid P_{i}-P_{i + 1}\mid} $$
(21)

Where, Pxl is the pixel value (zero or one).

If number of 1s = 0 then, isolate the point,

If number of 1s = 1, then it is a termination point,

If number of 1s = 2, then it is a connective point,

If number of 1s = 3 then it is a bifurcation,

If number of 1s = 4 then it is a crossing point.

See Fig. 20.

For example, Fig. 19 defines a block from an image. According to the previous equation, CN = 1/2 × (1 + 1 + 1) = 1.5. Then, this block is defined as bifurcation.

Fig. 19
figure 19

A block of an image (3 × 3 window)

Figure 20 illustrates a special case that a genuine branch is triple counted. From Fig. 20, the ridgelines area is represented by ones and the background area is represented by zeros. We color pixels according to their types where green indicates bifurcation and blue indicates termination. The yellow pixels represent a special case called triple counting and is considered as a bifurcation.

Fig. 20
figure 20

Different pixel types

A post step is performed to reduce the number of false minutiae. False minutiae may be generated in the image due to factors such as noisy images, and image artifacts produced by the thinning process. Hence, after the minutiae is extracted, it is necessary to perform a post step in order to validate the minutiae.

The impact of false minutiae is significant on the accuracy of the system. In our work, we focus on eliminating any false minutia by calculating the average distance between two neighbor ridges.

Table 3 shows the various types of false minutiae.

Table 3 Various types of false minutiae

There are three important rules to remove false minutia:

  1. 1.

    If the value of the distance between bifurcation and termination is less than D and the two minutia are on the same ridge as shown in Minutiae 1 in Table 3, then reject minutia.

  2. 2.

    If the value of distance between two bifurcations is less than D and the two minutia are in the same ridge as shown in Minutiae 2 and Minutiae 3 of Table 3, then reject minutia.

  3. 3.

    If two terminations are equal to distance D and their directions are coincident with a small angle variation, and no other termination is located between the two terminations as shown in Minutiae 4, Minutiae 5 and Minutiae 6 of Table 3, then reject minutia.

It is worth mentioning that the removal of the false minutia has important positive implications on the accuracy of the matching. Moreover, the reduction of the number of minutiae reduces the time complexity.

The next step is the unification of minutiae representations. Because of the diversity of the fingerprint image capturing conditions, the fingerprints should be captured in a standard way to be able to identify the minutiae features. These standards are related to x-coordinate, y-coordinate, the orientation of the associated ridges. Regarding bifurcation, the calculation of the orientation needs special treatment, where the bifurcation point contains three ridges and their own x-y coordinates. To resolve this issue, we use the method of Shobhraj et al. [51]. According to Shobhraj method, the orientation between two terminations is implemented using the following procedure:

  1. 1.

    Track a ridge segment starting from the point that termination and length is D.

  2. 2.

    Sum up all the x-coordinates (sx) of points on this ridge segment.

  3. 3.

    Get sx then divide the above summation with D and determine sy using the same technique [42]. \(s_{x} = \frac {\sum x-coordinates of points in the ridge segment}{D}\) \(s_{y} = \frac {\sum y-coordinates of points in the ridge segment}{D}\)

  4. 4.

    Finally, determine the direction using Equation 22:

$$ Tan^{-}1 \left( \frac{sy-ty}{sx-tx}\right) $$
(22)

6.2 Fingerprint identification

After th registration, the fingerprint features are stored in the form of a template, with (.dat) extension, in the database. After that, the system should be able to identify the registered fingerprint. This stage is called fingerprint identification/recognition. The identification process depends on the minutiae features matching. After extracting and saving the minutia from the query fingerprint and stored as vector, this vector is matched against the stored minutiae victores in the database. The similarities between the minutiae are calculated using the minutiae match algorithm. Based on this, it is determined whether two images are for the same person or not. In this work, we use the alignment-based algorithm proposed in [32]. This algorithm goes through two stages: the alignment stage, where the two minutiae are aligned, and the match stage, where the similarity score of the two minutiae is computed.

The first stage is the alignment stage. When two fingerprint images are compared, the selection of the minutia of each image is selected, and the similarity of the two ridges related to two reference points of the minutiae is calculated. The algorithm computes all similarity scores and chooses the highest similarity for the match.

Let us assume that ridge is related to each minutia in the form of a series of x-coordinates (X1, X2...XN) of points on the ridge. For further illustration, we show an example. Consider that I1 and I2 are the two minutiae obtained by the following (23) and (24):

$$ I_{1} = {m_{1}, m_{2}, ..., m_{M} } $$
(23)
$$ I_{2} = {m^{\prime}_{1}, m^{\prime}_{2}, ..., m^{\prime}_{M} } $$
(24)

where mi = (xi,yi,𝜃i). A sample point is taken along the ridge (L) where L defines the average inter-ridge length. In order to obtain the similarity score between the minutiaes, we use the following equation.

$$ S =\frac{{\sum}_{i = 1}^{m} xiXi}{\sqrt{{\sum}_{I=\theta}^{m} xi^{2}Xi^{2}}} $$
(25)

where (xi xn) , (Xi XN) are the x coordinates for each minutia and m is minimum of the n and N value. If the similarity score is greater than the threshold value, then execute next step, else, compare the next new two ridges.

The second stage is Match Stage. If we assume that almost all minutiae are found in the same position and direction, then the minutiae can be compared to obtain similarity scores. However, the minutiae are exposed to some problems, where it is impossible to have the same minutiae at the exact position and direction. This problem is solved by surrounding each minutia pattern with a bounding box. Thus, the comparison will no longer be between the exact minutiae location and direction, but rather between the location and orientation within the surrounding boxes.

7 Experiments and results

7.1 Evaluation methodology

We have selected a random sample of 100 volunteers (50 males and 50 females; aged between 15 and 18 years) and captured their fingerprint images. After that, we converted each raw image to a compressed image using JPEG compression with different QFs ranging from 1 to 100. This resulted in 10,000 images (100 images per volunteer).

The first experiment was conducted with the hundred random volunteers on QFs 10,20,30,...,100. The main objective of this experiment is to determine the best tradeoff between compression and fingerprint recognition accuracy.

While capturing images of the sample, the disparity in images clarity, regarding to volunteer’s gender sparked my attention. As details of the male fingerprint was clearer than females, this observation led me to conduct my experiments on separate male and female samples, and record the results. This may help to make a decision in this regard.

One criterion related to the reliability of the matching process is related to the difference between the highest similarity value between the query image and the database images and the second highest similarity value. So, we study the results of this criterion. If the differences are little, then the results are not reliable.

Another criterion to judge the effectiveness of the used system is how it handles unknown fingerprints. A good system should produce very low similarity scores in such cases. So say that a recognition system is authorized, and can be used effectively, you must identify the fingerprint even changed his image. System users need match differently every day with original images stored within the system. To check the system, we take a sample of 50 images on behalf of session 1 and 50 other images of the same people but in a different session. After that, we measure the system’s accuracy using images from different sessions.

Our experiments involve processing the obtained images in different ways, each with its own software requirement. For example, we compress the images using different QFs. We use C# code for this task. We also segment the compressed images to get rid of background. We use MATLAB code to conduct this task. A binarzation step is applied using MATLAB code. The matching algorithm, which includes feature extraction as well as similarity score computation, is also conducted with MATLAB.

7.2 Impact of using different training and testing QFs

In the first set of experiments, we aim at measuring the impact of using different QFs for the training as well as the testing sets. For the training dataset, we randomly choose 100 volunteers and take their fingerprint images with QFs ranging from QF10 to QF100.

For the results shown in Fig. 21, the experiments were conducted using 100 images in different QFs ranging from 10 to 100. Each figure represents the results of the experiments on a different quality factor specified in the training dataset. In these figures, we note that the accuracy of the fingerprint recognition system increased proportionally with the quality factor. In Fig. 21j, the experiment was conducted regarding to QF100 in training dataset. Here we note that the framework achieved satisfying accuracy for testing QFs of 30 and above. For training QFs of 50 and above, the achieved accuracy between 80% and 100% for testing QFs of 40 and above. However, when as we decrease the training QFs from 30 to 20 to 10, the accuracy range gradually decreases from 70–80% to 60–70% to 30–40%. I think that these ranges are not generally acceptable for fingerprint recognition systems.

Fig. 21
figure 21

Fingerprint recognition accuracy when using training sets of different quality factors

7.3 Using low testing QFs

The previous set of experiments revealed that while moderate and high testing QFs resulted in high recognition accuracy, low testing QFs (QF30 and below) presented more interesting trends. So, it is worth it to study such QFs in more detail especially since it is reasonable to assume that training fingerprint images might be captured and stored with high QF whereas testing (query) fingerprint images might be captured and stored with low QF. In this experiment set, the training dataset contained fingerprint images of 100 volunteers with QF100. As for the testing images, they have QFs ranging from 10 to 30.

Figure 22 shows the results of this experiment. The Figure shows that the accuracy increases as the testing QF goes from 10 to 22. However, for testing QFs between 22 and 30, the accuracy saturates on about 75%. This accuracy is acceptable for certain applications such as when using fingerprint recognition to record the attendance of students in classes or employees in an organization.

Fig. 22
figure 22

Fingerprint recognition accuracy for low testing QFs

7.4 Considering gender

During the sample collection phase, it has come to our attention that there are differences between male and female fingerprints in terms of clarity. This motivated us to further explore this issue in a separate set of experiments in order to determine whether fingerprints of different genders are affected differently by the different QFs. To the best of our knowledge, this issue has not been addressed before.

For this experiment, we randomly choose a set of 100 volunteers (spread equally among the two genders). Figure 23 shows the results of the experiments. The figure shows that there is no difference in the system accuracy on testing QFs 60 and above. However, if we look at the results for testing QFs below that, we find that the accuracy for male fingerprints is significantly higher than that for female fingerprints. Most interestingly are the results for testing QF 10 where the accuracy for male fingerprints reached 50% while the accuracy for female fingerprints is only 10%. One justification for such disparity is the typical lifestyle of members of each gender in the region from which the volunteers hail. Specifically, we are talking about how women do the housework and are continuously exposed to erosion factors that affects the fingerprints negatively. This is a result of several causes including: washing pots, big chemical detergent exposure leading to the smoothness of their fingerprint. Unlike males who are characterized by performing tough jobs making their fingerprints clearer.

Fig. 23
figure 23

Impact of Gender on QF

7.5 Difference between highest and second highest similarity scores

As mentioned earlier, the fingerprint recognition system works in a “nearest neighbor” manner in the sense that it calculates the similarity between the test image and all training images and then it chooses the training image that is closest to the test image. It would be interesting for such systems to study how reliable is the decision they make by considering the how close is the system to choose the second nearest neighbor. Specifically, consider the training image with the highest similarity (call it i1) and the training image with the second highest similarity (call it i2). If the similarity score of i1 is very close to that of i2, then the system cannot be considered reliable.

Figure 24 shows the results of this experiment. The figure shows that with higher QF the differences in similarity score between image with maximum similarity and image with second maximum similarity becomes higher, which means that the recognition system is more reliable. The figure also shows that the differences for male fingerprints are higher than those for female fingerprints, which supports the observation made in the previous experiment about the system higher effectiveness in matching male fingerprints compared with female fingerprints.

Fig. 24
figure 24

Differences in similarity score between image with maximum similarity and image with second maximum similarity

7.6 Using unseen testing images

In all of the previous experiments, the testing images exist in the training dataset, but with different QFs. However, in realistic fingerprint recognition scenarios, the query (testing) image might be for someone whose fingerprint images are in the training database, but the testing image itself does not appear in the training database. For this purpose, we consider a training dataset of 50 volunteers where each image is stored in three different QFs: QF30, QF50 and QF70. As for the testing images, they are for the same volunteers, but they are not the same as the training images (the training and testing images were each taken in a separate session). Also, for the testing images, we consider more QFs: QF10, QF20, QF100. Figures 2526 and 27 show the results of this experiment. The figures show the same trend in the sense that the accuracy is low for QF10 and, as the testing QF is increased, the accuracy exhibits sharp increase followed by a quick saturation after QF30. This conforms with the previous experiments.

Fig. 25
figure 25

Accuracy between two different sessions on QF30

Fig. 26
figure 26

Accuracy between two different sessions on QF50

Fig. 27
figure 27

Accuracy between two different sessions on QF70

7.7 Rejecting unauthorized people

In all of the previous experiments, the query (testing) image is for someone whose fingerprint images are already in the training database. So, we are actually testing the system’s ability to correctly identify people already in the database. This means that system’s output is a true positive or a false negative (known as Type II error). To use more technical terms, the accuracy that we have been measuring so far is known as the sensitivity or recall of the system, which measures the percentage of true positives to the summation of true positives and false negatives. However, such errors are not the only errors faced in realistic scenarios. In a false positive (known as Type I error), the system mistakenly identifies a person as being in the database where in fact he/she is not. To account for such errors, we use a measure called specificity, which computes the percentage of true negatives to the summation of false positives and true negatives. Now, to use this measure, it does not make sense to keep using the “nearest neighbor” approach discussed in the previous section. Instead, we must introduce a minimum threshold on the similarity such that if the highest similarity score is below this threshold, then the testing image must be rejected (i.e., not matched with any training image).

For this experiment, we consider a training dataset of 50 volunteers where each volunteer has two images taken on two different sessions and each image is stored in three different QFs: QF30, QF50 and QF70. As for the testing images, they are for the six volunteers different in the session from the volunteers of the training dataset, but from same quality factor. Table 4 shows the similarity score for each of the six testing images. The results show that low QFs gives higher similarity scores, however, the similarity scores are still very low to be on any concern. In fact, the table shows that using a low similarity threshold of 60%, which is considered to be very low in the literature, can guarantee that our system would have a specificity of 100%.

Table 4 Similarities for 6-peoples out of the systems with 3 different QFs:

8 Conclusions and future work

In this paper, we have surveyed the literature for the most common fingerprint acquisition devices and the different available fingerprint image sets. We have noticed that the latest trend in fingerprint acquisition is to use different types of digital cameras and specially the use of mobile phone cameras. We have also noticed that the available fingerprint datasets acquired using digital cameras are all taken with build in degradations due to specific lighting, different background, low camera quality, image compression, and many more. In this paper we built our own high quality fingerprint image set that can be used to study the impact of different factors on the accuracy of a fingerprint recognition accuracy. Based on the results of our experiments, we concluded that the area of image acquisition must be given more attention than what it has now. We must consider the image quality with no less importance than other main stages of biometrics framework such as feature extraction and matching techniques. Through our pilot study, we focus on choosing the best QF of image that provides no loss of information that may be important in determining the identity of the individual. Moreover, there is no need to keep the large size of the image when it is enough to use a certain percentage of the size of this image to identify the individual. Our experiments prove that using (20%–30%) of the original image size is good to extract enough features with no loss of important information. In this work, we built a full fingerprint recognition system. In the future, we can study other applications such as the effect of image compression on identification through iris recognition, hand geometry, etc. Another future direction could answer the question: Can we compress children (under 10 years) fingerprint with the fingerprint information preserved? This is important because I notice during dataset collection that fingerprints of children are not sufficiently clear.