Keywords

1 Introduction

For example, biometrics includes fingerprints and retina scans. Biometric systems have the best track record for person authentication. Because biometrics is difficult to steal, it may be used instead of passwords and smart cards. Behavioral traits are often paired with physical traits to create biometric IDs. Physiological qualities of the body include fingerprints, palm veins, DNA, face recognition, iris, and so on [1].

Other categories included speaking voice, walking motion, etc. Biometrics helps safeguard people's health in two ways: physically and behaviorally. Biometrics has entered our everyday lives. Although iris is certified for precise biometric systems. Many people believe that biometrics is the proper technique to identify someone. The iris is the annular eye area between the white sclera and the black pupil [2].

Iris texture is very random and distinct, making it unlikely to be unique across two iris patterns or two eyes. Circular, ridged, crypts, freckles, furrows, and zigzag patterns are among the most common iris texture traits. Its reputation for dependability helps in distinguishing individuals. Until his death, the iris pattern remained unchanged. Because recognition techniques are not susceptible to spoofing assaults, they are regarded safer. Again, altering its patterns requires more risk than altering fingerprints, which are simpler to manipulate. Eyelashes, reflections, and eyelids may all interfere with iris-based identification systems. For example, Aadhar card access, banks, and current apps employ iris-based identification [3] (Fig. 1).

Fig. 1
An image of the human eye depicts the outer structure of the iris. The innermost dark center is labeled pupil. The layer surrounding the pupil is labeled the iris. The border of the iris is labeled limbus. The white portion of the eye is the sclera. The fold of skin above the eye is labeled eyelid.

The outer structure of human iris

This study encompasses the following: Sect. 2, the study of related Studies. In Sect. 3, there is the proposed iris recognition system, and in Sect. 4, there are the experimental results and analysis. Finally, the study's future evolution is summarized in the last section.

2 Review of Related Studies

S. Arora and M. S. Bhatia an iris scan may be used to verify a person's identification using deep learning. The circular Hough transform and CNN is used to locate an iris. Then, using a Softmax classifier, the dataset is classified into one of 224 classifications. The accuracy of their proposed system relies on the choice of hyper-parameters and optimization methods.

To discover the iris border in the eye picture, Reddy [4] used Daugman’s rubber-sheet model, followed by a circular Hough transform and edge detection. The recommended system is then trained using a CNN to identify iris characteristics. This classifier uses Softmax to classify 224 iris datasets and 108 CASIA V1 Iris datasets. Thus, the system's overall performance and deep-learning training and optimizations. We met our targets with 98% success and 95.4% accuracy.

Alaslni and Elrefaei [5] suggested CNNs (VGG-16s) for iris recognition model training. They used a pre-trained VGG-16 to successfully train on the IITD and CASIA Iris datasets. However, the suggested solution does not work with other publicly available datasets or solve other biometric issues.

CNN's iris recognition system was developed by Kien Nguyen et al. They tested DenseNet on two datasets, LG2200 and CASIA-Iris-Thousand, and found that it outperformed AlexNet, VGG, Google Inception, and ResNet. The authors did not explain why the AlexNet model was less accurate.

Shervin Minaee et Deep CNN features for iris recognition: a conceptual study They developed a system using VGG Net characteristics and got decent results. They studied data from CASIA-Iris-1000 and IIT Delhi using two well-known databases. This study uses a modified object recognition framework. Their findings would have been more accurate if they had used an architecture optimized for iris detection.

Aziz Nazri Ali and Sue Chin Yow His idea was to use deep learning to develop an iris scanner. In addition to SVM, Bayesian optimization, and data augmentation, this study uses CNN and data augmentation. The study's dataset is CASIA V1. The results are noteworthy; however, the approaches are not suited for other publicly accessible datasets.

3 The Proposed Iris Recognition System

In Fig. 2, the proposed iris recognition system uses a denseness for feature extraction and classification. The iris recognition technology will be developed in three areas: It includes both picture capture and preprocessing.

Fig. 2
A flowchart of the iris recognition system. Dataset; pre-processing: image enhancement, discreet wavelet transform, image segmentation, and image normalization; feature extraction: convolutional neural network and Softmax classifier. If it is a match, then recognition otherwise it is considered unrecognition.

Flow diagram of the system employed in this study

3.1 Image Acquisition

After capturing a picture of the iris, it may be preprocessed to improve clarity. CASIA-Iris-1000 [6], interval 12 [7], Delhi 10 [8], and the UBIRIS Database [9] were used in this investigation. The table has not been preprocessed Table 1.

Table 1 Presents the information on images on each dataset before preprocessing and is known as the information on images before preprocessing

3.2 Preprocessing

The four methods depicted in Fig. 3 are used before processing iris images. Because collected pictures are packed with noise, they must be processed before being fed to segmentation. Because brightness is not uniform and noise must be reduced, a histogram equalization is used to enhance the picture. Use unsharp mask to sharpen your picture to bring out textures and characteristics.

Fig. 3
A flow diagram of the process in iris systems. The process is as follows: a human eye is captioned input; the borders of the iris and pupil are highlighted and the image is captioned segmentation followed by normalization; a C N N model for feature extraction followed by neural network classification which gives the final output.

The process involved in iris systems

3.2.1 Histogram Equalization

Equalization of histograms: A histogram equation is applied to increase the contrast and brightness of images. A linear and equal distribution of colors is obtained by examining the color distribution in the image and choosing a conversion function to modify the light values. The number of repetitions for each color is calculated, and then, the number of repetitions for each color is calculated and done. According to Fig. 4, the values are divided by the sum of the colors in the image to yield new color values.

Fig. 4
Images of the eye with area graph to depict the color distribution. Values are estimated. Original image outline: dark shaded throughout (50, 100), (175, 1400), (180, 1000), (200, 1400). Output image outline: dark shaded bars at (0, 400), (40, 400); light shaded at (80, 1400), (125, 1000), (150, 1200); dark shaded at (225, 400), (250, 300).

The histogram equation color distribution

3.2.2 Unsharp Mask Filter

To highlight the image's texture and detail, the unsharp mask is applied. By improving the resolution of the image at a tiny scale, it can substantially increase the look of detail. It removes an image that has been smoothed out of the original image. You first generate a copy of the original image that is very fuzzy and then you remove the blurry copy from the original image, leaving only the clearer image behind. The remaining difference is known as a mask. Next, you apply the mask to the original image. Completing the original image with the mask, rather than masking the original, leaves behind a clearer end result.

3.2.3 Reducing Image Dimensions and Preserving Important Features

We can help with preprocessing, a method for reducing picture size and speeding up the matching process. Our first DWT was run on the normalized iris picture. The DWT transformation, shown in Fig. 5, adds four sub-band frequencies to the iris image: LL, LH, HL, and HH. The iris low-frequency coefficients are encoded in the LL sub-band [10], which is used for iris segmentation.

Fig. 5
Images of iris depicting D W T transformation. The first image of a normal iris is labeled L L. The second one labeled L H is light with just an iris outline. The third is a dark image labeled H L with a faint iris outline. The fourth image labeled H H is a completely faded one.

Iris template using one-level DWT

Wavelet discretization: A wavelet transformation reduces noise by decomposing data or images into wavelet coefficients [11]. The inverse discrete wavelet transformations are utilized to reconstruct the picture [12]. Figure 5 shows the resultant frequency sub-bands (LL, LH, HL, and HH). Iris features may be characterized using low-frequency coefficients, LL sub-band, allowing us to employ LL frequency sub-band for feature reduction [10].

3.2.4 Iris Segmentation

The next step in iris identification is iris segmentation, which involves processing a digital eye picture to determine the iris area. Figure 1 illustrates two concentric circles as an iris area approximation. The inner-circle depicts iris/sclera; the outer circle represents iris/pupil. The Hough transformation circular iris area.

The Hough transformation computes geometric parameters (lines and circles) in an image. The Hough transform may be used to identify the iris and pupil center coordinates and radius. Commonly used to identify object shapes based on class names. An edge map is created by calculating the gradients (first derivatives of intensity values) in an eye picture. To obtain the greatest values for circle parameters, neighboring areas on the circle at various radii are evaluated, and votes are cast [13]. This equation computes the center’s coordinates and radius:

$$ X_C^2 + Y_C^2 - Y^2 = 0 $$
(1)

The circle’s radius (r) is the biggest number, and the circle's center coordinates (XC, YC) are determined using Hough space edge points. Vertical derivatives (gradients) are used to identify the iris-sclera border. To remove the eyelids' effect from the Hough transform, only vertical gradients may be employed. Some of the circle's boundary pixels are unnecessary. Less edge points to vote on improve accuracy and efficiency [13] (Fig. 6).

Fig. 6
An image of an eye undergoes Hough transformation. The stages are as follows: an original image of the eye depicts the pupil, iris, and sclera clearly, the edge map is a diagram of the eye with the pupil and the iris, the horizontal edge map depicts no iris but pupil, and the vertical edge map presents faint boundaries of the iris and the pupil.

Segmentation using Hough transformation

3.2.5 Normalization

After successfully segmenting the iris area from an eye picture, the following step is to transform it into a fixed-size rectangular block. Normalization creates an iris with regular proportions. These measures are consistent; therefore, photographs of the same iris will have comparable properties [14]. This work uses Daugman's rubber-sheet model to normalize the iris image.

Rubber-sheet model: As the name implies, the rubber-sheet model transforms the iris into a rectangular block in non-concentric polar coordinates, which is the most prevalent kind [15]. To calculate the comparable location on the polar axis (r), Daugman created a rubber-sheet model. The angular resolution is calculated by dividing the number of radial lines by the number of data points. The radial resolution is the number of data points. Using Eq. 2,

$$ I \left[ {x\left( {r,\theta } \right), y\left( {r,\theta } \right)} \right] \to I \left( {r,\theta } \right) $$
(2)

(x, y) is the iris position, and (r) is the Cartesian and normalized polar coordinates. rranges from Rp to Rl, while varies from 0 to 2. The boundary points are utilized to calculate x(r) and y(r).

The array’s horizontal and vertical dimensions correspond to angular and radial resolution, respectively (Fig. 7).

Fig. 7
A diagram of the rubber-sheet model of Daugman. Two concentric circles followed by a rightward arrow pointing to a rectangle. The inner and outer circles are labeled R p and R l and an arrow between them is labeled r. An arrow from r points to theta. The breadth of the rectangle is labeled r. The left and right ends of the length are 0 and 2 pi.

Rubber-sheet model of Daugman

3.2.6 Deep Learning for Iris Recognition

After preprocessing, we use a Softmax classifier to extract and classify CNN features. The DenseNet201 model is used in this work to simulate the CNN architecture. This is the top categorization layer; it is linked. The Softmax classifier then gets the fully connected layer's output and computes the probability distribution across N classes. This section also discusses the efficient recognition of the training process, network architecture, and input picture size. Keeping over fitness at bay throughout training is vital.

3.2.6.1 Training Methodology

The database is used to build the training, validation, and testing sets. 80% of the database is for training, and 20% is for testing. The validation set also contains parameters with a low validation error. The Softmax classifier divides the training set into mini-batches and calculates the training errors for each. The validation set is used after each epoch to test the present architecture's validity. The present architecture's cost value and validation error are then determined. This study's beginning learning rate is 0.001. The dropout approach [16] avoids overfitting. Each node and its connection are ignored in each training iteration by 0.2. After the training phase, the best model is chosen based on performance recognition.

3.2.6.2 Network Architecture

DenseNet is a CNN model created by Huang et al. and won the ILSVRC-2016 competition in 2017. An already trained and configured convolutional neural network model called DenseNet201 is utilized for the iris recognition [17] in this method. Each of the four dense blocks and three transition layers in DenseNet201 is shown in Fig. 8. To enhance the capabilities of the DenseNet201 bottleneck output features, a flattening layer and a dense layer are applied, and then, a Softmax layer is applied on top.

Fig. 8
An illustration of the DenseNet architecture. It has 6 columns and 14 rows. Columns are titled layers, output size, DenseNet 121, DenseNet 169, DenseNet 201, and DenseNet 264.

Shows DenseNet architecture

3.2.6.3 Input Image Size

The size of the input image influences the speed and accuracy of the network. To test the system's performance, three input image sizes (128 × 128 pixels, 64 × 64 pixels, and 32 × 32 pixels) are employed. Normalization and histogram equalization followed by three layers of DWT are used to create these photos. As the iris size is lowered, the iris patterns will become completely undetectable; nevertheless, for the larger size, larger memory requirements and the higher level of complexity computations could pose a challenge.

4 Experimental Results and Analysis

It is tested with four iris datasets: IITD Iris Databases [8], UBIRIS.v1 [9], CASIA-Iris-Thousand [6], and CASIA-Iris-Interval [7]. This set of iris photographs uses pupil dilation, eyelid/lash occlusion, a tiny eyelid shadow, and specular reflection. Table 2 lists the dataset's specifications.

Table 2 The specifications of the used datasets

A laptop with a Core i7 CPU operating at 2.8 GHz runs the system and its stages, which are implemented using Python 3.8.

Initially, the investigations used the iris image before preprocessing. In this case, we use DenseNet to simulate the eye's iris in its normal size.

In the second case, the iris image is normalized. The normalizing step sets the radial and angular resolutions to 224. The normalized picture is 224 pixels wide and 224 pixels height, perfect for dense network input.

Case 3: Use the iris picture after DWT and normalization. The picture size changes from DWT 128-by-128 to 64-by-64 to 32-by-32.

There will be a random split of each picture set into training and test data. Images make roughly 80% of training data and 20% of testing data.

This is followed by feature extraction and classification using a pre-trained CNN model. DenseNet was pre-trained to identify 1000 classes [17]. We used the following database types to test the system on all participants in the dataset:

With the same dataset as the class dataset, we pre-trained a DenseNet, using a tuning approach that modifies outputs in the final fully connected layer (FC) based on the class count.

Our study used DenseNet2016 pre-trained with varied epoch counts. The complete training set traversed by the training process is one epoch [18]. The initial learning cycle's learning rate is 0.001. The mini-batch option has a batch size of 32. In computing the gradient of the loss function and adjusting weights, the words mini-batch and subset are interchangeable. Gradient descent iteratively modifies network parameters (weights and biases) [19, 20] (Table 3).

Table 3 The iris image databases' recognition accuracy

These are the database accuracy and training time before and after segmentation of the ocular image, as well as after utilizing the first level of DWT. Because DWT reduces the picture size while maintaining important features, it reduces training time and increases accuracy. So our method improves accuracy without adding training time.

5 Conclusions

A pre-trained CNN was utilized to build an effective iris recognition system (DenseNet201). Then, an iris image is segmented using a circular Hough transform and normalized using Daugman's rubber-sheet model. The normalized picture is loaded into the pre-trained CNN (DenseNet201). The approach works well on public datasets. We will evaluate the suggested system's performance on other iris datasets and other biometric recognition tasks in the future.