TSP Cmes 23674
TSP Cmes 23674
DOI: 10.32604/cmes.2023.023674
ARTICLE
Smita Khade1 , Shilpa Gite1,2, *, Sudeep D. Thepade3 , Biswajeet Pradhan4,5, * and Abdullah Alamri6
1
Symbiosis International (Deemed University), Symbiosis Institute of Technology, Pune, 412115, India
2
Symbiosis Centre for Applied Artificial Intelligence, Symbiosis International (Deemed University), Pune, 412115, India
3
Computer Engineering, Pimpri Chinchwad College of Engineering, Pune, 411044, India
4
Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), School for Civil and Environmental
Engineering, Faculty of Engineering and Information Technology, University of Technology Sydney,
Sydney, New South Wales, 2007, Australia
5
Earth Observation Center, Institute of Climate Change, University Kebangsaan Malaysia, UKM, Bangi, Selangor, 43600, Malaysia
6
Department of Geology & Geophysics, College of Science, King Saud University, P.O. Box 2455, Riyadh, 11451, Saudi Arabia
*Corresponding Authors: Shilpa Gite. Email: [email protected]; Biswajeet Pradhan. Email: [email protected]
Received: 09 May 2022 Accepted: 13 September 2022
ABSTRACT
Contactless verification is possible with iris biometric identification, which helps prevent infections like COVID-19
from spreading. Biometric systems have grown unsteady and dangerous as a result of spoofing assaults employing
contact lenses, replayed the video, and print attacks. The work demonstrates an iris liveness detection approach
by utilizing fragmental coefficients of Haar transformed Iris images as signatures to prevent spoofing attacks for
the very first time in the identification of iris liveness. Seven assorted feature creation ways are studied in the
presented solutions, and these created features are explored for the training of eight distinct machine learning
classifiers and ensembles. The predicted iris liveness identification variants are evaluated using recall, F-measure,
precision, accuracy, APCER, BPCER, and ACER. Three standard datasets were used in the investigation. The main
contribution of our study is achieving a good accuracy of 99.18% with a smaller feature vector. The fragmental
coefficients of Haar transformed iris image of size 8 ∗ 8 utilizing random forest algorithm showed superior
iris liveness detection with reduced featured vector size (64 features). Random forest gave 99.18% accuracy.
Additionally, conduct an extensive experiment on cross datasets for detailed analysis. The results of our experiments
show that the iris biometric template is decreased in size to make the proposed framework suitable for algorithmic
verification in real-time environments and settings.
KEYWORDS
Iris images; liveness identification; Haar transform; machine learning; biometric; feature formation; ensemble
model
This work is licensed under a Creative Commons Attribution 4.0 International License,
which permits unrestricted use, distribution, and reproduction in any medium, provided the
original work is properly cited.
324 CMES, 2023, vol.136, no.1
1 Introduction
Automatic human access to a system has become relatively simple in the digital era. Confirmation
of the user’s identification is critical for automated system access. Biometric authentication systems
employ biometric features to confirm the identification of a user [1]. Compared to conventional
password-based traditional verification systems, the biometric system has a distinct advantage. It
minimizes the need to memorize a passcode, pin, or keep a card in hand [2]. Biometric authentication
can be thought of as an additional layer of authentication for security-critical cyber applications
and existing traditional authentication procedures. Today, there are indeed different businesses for
biometric systems. The majority of the sectors seem to be rising quickly. According to www.statista.
com (accessed on 24th June 2022), the business for contactless biometric technologies is forecasted to
increase by roughly 30.15 billion US dollars by 2027, while the overall biometric sciences industry is
forecasted to touch 19.08 billion US dollars in 2021 [3]. Iris is widely employed in the verification and
validation of people. In most applications, it uses, because of its complex textures [4] and distinctive
features, such as the UIDAI project for citizen identification in India, the Amsterdam airport, and the
Canada-US border onon-US [5].
In comparison to fingerprint and face authentication, iris authentication delivers an additional
steadfast contactless user verification. The contactless method aids in the prevention of diseases like
COVID-19 [6]. Despite the iris having a distinct textural structure, the imposter might falsify it [7].
Table 1 shows the iris presentation attacks used that are found in the literature [8].
Individuals frequently assault the biometric system to get admittance to another person’s creden-
tials or to conceal their accurate individuality. The iris identification system can be readily fooled by
means of alternative contact lenses (which can be transparent, textured, colored) [10], replaying the
video, or using a print attack [9]. As a result, understanding the risk and susceptibility is critical for
safeguarding the biometric system [14]. The complicated risk of biometric spoofing is minimized by
assessing the liveness of biometric features prior to authentication [15]. The main objective of this
study is to identify iris liveness detection with reduced feature vectors.
The main objective of this study is to identify iris liveness detection with reduced feature vectors.
Following are the novelty and the main contributions of this paper.
CMES, 2023, vol.136, no.1 325
Using pupil dynamics, the author Czajka [19] created the liveness identification system. The pupil
reaction is tested in this system using rapid changes in the intensity of light. In the case of the eye
reacting to the variations in light intensity, it is alive; otherwise, presentation attack image. In [19],
non-linear and linear SVM is employed to categorize natural reactions and impulsive oscillations. The
system’s shortcoming is that it measures a variety of functions that take time. There are inaccuracies
in the observation because the data utilized in this research does not include any details from elderly
adults.
Author Fang et al. [20] apply many data augmentation methods to generate variability. The
strategy-level and the score-level combination of fusion methods are used for Iris PAD. Bassi et al. [21]
detected PAD using NIR, visible domain, cross-datasets and cross-spectrum datasets. Authors con-
cluded that Cross-PA and cross-datasets are still challenging, as EER values above 20% in most of
cases.
A technique to detect Accurate Ocular Regions was created by Naqvi et al. [22]. This solution
uses deep neural network variants. The system’s evaluation considers publicly available databases.
Kimura et al. [23] developed a CNN-based liveness detection System that improves model accu-
racy by modifying hyperparameters. “Attack Presentation Classification Error Rate [APCER]” and
“Bonafede Presentation Classification Error Rate [BPCER]” metrics are taken to assess the system’s
performance. The hyperparameters are all studied in this work. This method is solely effective against
prints and contact lens attacks. Author [24] studied multiple transfer learning models to detect the iris
liveness and concluded that EfficeintnetB7 gives highest classification accuracy.
Only a few studies were found to be robust against all sorts of spoofing assaults [3,25]. Most of
the studies used a higher size feature vector. Based on these findings, it can be believed that there is a
necessity for a classifier or ensembles for the detection of every sort of spoofing assault.
3 Proposed Iris Liveness Detection Utilizing Fragmental Energy of Haar Transformed Iris Images
The iris recognition system is prone to a range of security threats. Because of these flaws, the
system is less trustworthy for robust authentication applications. The study employs fragmental
energy of Haar modified iris images to attempt iris liveness detection. These fragmental energies were
employed as features to detect whether the iris was real or fake. Because of these characteristics, the
suggested methodology does not require any pre-processing, such as segmentation, normalization, or
localization, which are commonly employed by methods presented in the literature. These fragmental
energies were employed as features to detect whether the iris was real or fake. Because of these
characteristics, the suggested methodology does not require any pre-processing, such as segmentation,
normalization, or localization, which are commonly employed by methods presented in the literature.
These fragmental energies make the suggested technique faster and more accessible [26]. Resizing
the iris image to 256 ∗ 256 is the sole pre-processing performed in the proposed framework. The iris
liveness detection process depicted in Fig. 1 is a block diagram. There are three phases in the proposed
system. Resizing [pre-processing] of iris images, feature formation, and classification with iris liveness
identification.
CMES, 2023, vol.136, no.1 327
Figure 1: Block diagram of the projected iris liveness detection employing fragmental energy of Haar
transformed iris images
3.1 Pre-Processing
The importance of iris pre-processing in iris liveness detection cannot be overstated. Two iris pre-
processing techniques are used in the suggested algorithm. Because images are obtained using three
standard datasets, each dataset stores images of different sizes. We normalized the original 256 ∗ 256
images in pre-processing to ensure they remained intact throughout the experiment. At the same time,
photographing various datasets with various sensors, some (LG, Content, Vista) acquired images in
RGB format, while others [LG, Dalsa] acquired grayscale images. The images were then converted to
grayscale to keep their originality.
In the Haar transform iris image, the left highest corner has the higher energy and crucial
information, as shown in Fig. 2. This results in considerable energy compression in a limited count
of high energy coefficients. As a result, these are the preferred feature vector elements. To construct
feature vectors for proposed iris liveness detection, 256 ∗ 256, 128 ∗ 128, 64 ∗ 64, 32 ∗ 32, 16 ∗ 16, 8 ∗ 8, and
4 ∗ 4. Pixels are used to capture the high-energy portion of Haar transformed iris image coefficients.
Figure 2: Proposed fragmental energy-based feature creation approach for liveness identification from
cosine transformed iris images
These feature vectors support the reduction of the size of feature vectors. As a result, iris-liveness
detection is speedier. The compacted high energy improves iris liveness detection accuracy in these
low-frequency coefficients. These high-energy features are then employed for training the ML models
working to detect iris liveness.
4 Experimentation Setup
The investigative results of the proposed method are discussed in this section. The experiments
were performed using an Intel (R) Core (TM) i3-6006U CPU @ 2.0 GHz, 12 GB RAM, and 6a
464-bit operating system with MATLAB R2015a as a programming platform. Clarkson LiveDet2013
(Clarkson 2013), LiveDet2015 (Clarkson 2015), and IITD Combined Spoofing datasets (IIITD CSD)
were used to explore the suggested approach to iris liveness detection.
Table 2: Sample of images used for exploration from Clarkson 2013 dataset
Sensor Image category Sample images Count of images taken Count of images
for exploration training taken for testing
Table 3: Sample of images used for an experiment from Clarkson LiveDet2015 dataset
Sensor Image Sample images Count of images Count of images
category taken for training taken for testing
Patterns 349 82
Live 166 NA
LG Printed 270 NA
(Continued)
CMES, 2023, vol.136, no.1 331
Table 3 (continued)
Sensor Image Sample images Count of images Count of images
category taken for training taken for testing
Patterns 303 NA
Table 4: Sample of images used for an experiment from IIITD combined spoofing dataset
Sensor Image category Sample Images Count of images Count of images
taken for training taken for testing
(Continued)
332 CMES, 2023, vol.136, no.1
Table 4 (continued)
Sensor Image category Sample Images Count of images Count of images
taken for training taken for testing
5 Experimentation Setup
The benchmark datasets for all feature size variants are taken to test the proposed iris liveness
detection method. Performance measurements such as accuracy, F-measure, precision, and recall are
considered for testing versions of the proposed iris liveness detection technique.
CMES, 2023, vol.136, no.1 333
95 256 X 256
Accuarcy In %
128 x 128
90
64 x 64
85
32 x 32
80
16 x 16
75
8x8
4x4
Classifiers
From Fig. 4, it has been noted that the performance improves as the size of the feature vector
is reduced from 256 ∗ 256 to 8 ∗ 8 and then begins to deteriorate with feature vector size 4 ∗ 4.
This demonstrates that the fragmental coefficients of Haar transformed iris images provide more
outstanding iris liveness recognition capabilities while maintaining a small feature vector size, proving
the importance of the suggested method. The highest average accuracy, 95.94% achieved by 16 ∗ 16
fragmental coefficients.
92.67
4x4
95.68
95.94 8x8
AVG 93.73 16 x 16
90.73 32 x 32
89.74
64 x 64
93.33
128 x 128
86.00 87.00 88.00 89.00 90.00 91.00 92.00 93.00 94.00 95.00 96.00 97.00 256 X 256
Accuracy in %
Figure 4: The performance assessment for the Clarkson 2013 dataset by averaging the specific
fragmental coefficients in iris liveness detection
Table 5 appraises the performance of specific ML classifiers and ensembles of classifiers for
iris liveness detection tested on the Clarkson 2013 dataset. From Table 5, it can be noted that NB
classifiers give the highest average ACER, whereas the lowest average ACER, 2.63% achieved by
SVM+MLP+RF ensembles of classifiers.
334 CMES, 2023, vol.136, no.1
Table 5: ML classifier’s performance evaluation in the proposed iris liveness detection approach for
Clarkson 2013 dataset using an average of % accuracy, % APCER, % BPCER, and % ACER values
Classifiers/EOC AVG APCER BPCER ACER
Random forest 95.83 4.1 3.38 3.74
SVM 95.63 4.3 3.54 3.92
J48 91.00 8.89 8.12 8.505
NB 88.67 11.04 9.92 10.48
SVM+NB+RF 95.07 4.53 3.63 4.08
SVM+RT+RF 96.47 3.43 3.09 3.26
SVM+RF+MLP 96.93 3.05 2.21 2.63
Note: The highest performance is represented in bold.
98
Accuarcy In %
256 X 256
128 x 128
96
64 x 64
32 x 32
94 16 x 16
8x8
92 4x4
Classifiers
From Fig. 6, it can be seen that the performance improves as the feature vector size is compacted
from 256 ∗ 256 to 8 ∗ 8 and then begins to deteriorate with feature vector size 4 ∗ 4. This demonstrates
that the fragmental coefficients of Haar transformed iris images provide more excellent iris liveness
recognition capabilities while maintaining a small feature vector size. The highest average accuracy,
96.80% achieved by 8 ∗ 8 fragmental coefficients.
CMES, 2023, vol.136, no.1 335
96.60
4x4
96.80
8x8
96.63
AV 16 x 16
96.00
G 32 x 32
95.62
64 x 64
94.97
128 x 128
96.40 256 X 256
Figure 6: Performance assessment by averaging the specific fragmental coefficients in iris liveness
detection for IIITD CSD
Table 6 provides the performance assessment of specific ML classifiers and ensembles of classifiers
in the projected iris liveness detection explored on the IIITD CSD dataset. It is noted from the table
that Decision Tree (J48) classifiers give the highest average ACER, whereas the lowest average ACER,
2.29%, is achieved by SVM+NB+RF ensembles of classifiers. The majority voting technique generates
ensembles of classifiers, so they provide the best classification accuracy.
Table 6: Performance evaluation of ML classifiers in the proposed iris liveness detection for IIITD
CSD dataset with an average of % accuracy, % APCER, % BPCER, and % ACER values
Classifiers/EOC AVG APCER BPCER ACER
Random forest 96.75 3.23 2.18 2.71
SVM 95.30 4.69 3.78 4.24
J48 95.49 4.42 3.14 3.78
NB 96.74 3.25 2.19 2.72
SVM+NB+RF 97.37 2.59 1.98 2.29
SVM+RT+RF 96.92 2.89 2.45 2.67
SVM+RF+MLP 96.47 3.51 2.67 3.09
Note: The highest performance is represented in bold.
iris liveness recognition capabilities while maintaining a small feature vector size. The highest average
accuracy, 99.58% achieved by 8 ∗ 8 fragmental coefficients.
100
95
Accuarcy In %
256 X 256
90
128 x 128
64 x 64
85
32 x 32
80 16 x 16
8x8
75 4x4
Classifiers
98.91
99.58 4x4
99.53 8x8
AVG 92.66 16 x 16
89.66 32 x 32
88.67 64 x 64
92.26 128 x 128
256 X 256
82.00 84.00 86.00 88.00 90.00 92.00 94.00 96.00 98.00 100.00 102.00
Accuracy in %
Figure 8: Performance assessment by averaging the specific fragmental coefficients in iris liveness
detection for Clarkson 2015 dataset
Table 7 gives the performance assessment of specific ML classifiers and ensembles of classifiers
in the projected iris liveness detection explored on the Clarkson 2015 dataset. It is noted from the
table that Decision Tree (J48) classifiers give the highest average ACER, whereas the lowest average
ACER, 3.2%, is achieved by SVM+RF+MLP ensembles of classifiers. The majority voting technique
generates ensembles of classifiers, so they provide the best classification accuracy.
Table 7: Performance evaluation of ML classifiers in the proposed iris liveness detection for Clark-
sonclarkson 2015 dataset with an average of % accuracy, % APCER, % BPCER, and % ACER values
Classifiers/EOC AVG APCER BPCER ACER
Random Forest 93.49 6.06 5.98 6.02
(Continued)
CMES, 2023, vol.136, no.1 337
Table 7 (continued)
Classifiers/EOC AVG APCER BPCER ACER
SVM 96.02 3.07 3.54 3.305
J48 93.32 5.78 6.64 6.21
NB 91.34 8.29 8.61 8.45
SVM+NB+RF 95.21 3.97 4.03 4
SVM+RT+RF 95.28 4.15 4.71 4.43
SVM+RF+MLP 96.70 3.31 3.09 3.2
Table 8 (continued)
Train datasets IIITD
Test datasets Clarkson 2013 Clarkson 2015
Metric Accuracy APCER BPCER ACER Accuracy APCER BPCER ACER
Random Forest 94.71 5.25 5.12 5.19 100.00 0.00 0.00 0.00
SVM 100.00 0.02 0.01 0.02 100.00 0.00 0.00 0.00
J48 90.65 9.31 9.45 9.38 99.73 0.27 0.27 0.27
NB 94.30 5.66 5.76 5.71 100.00 0.00 0.00 0.00
SVM+NB+RF 99.19 0.77 0.57 0.67 100.00 0.00 0.00 0.00
SVM+RT+RF 96.74 3.22 3.26 3.24 100.00 0.00 0.00 0.00
SVM+RF+MLP 100.00 0.03 0.04 0.04 100.00 0.00 0.00 0.00
AVG 96.51 3.49 3.45 3.47 99.96 0.04 0.04 0.04
Note: The highest performance is represented in bold.
In this section, the results of cross datasets performances are explained. The first scenario, where
model train on Clarkson 2015 datasets was evaluated on the Clarkson 2013 and IIITD test datasets.
The evaluation results are presented in Table 8, where bold digits indicate the highest accuracy. From
Table 8, we observed that Clarkson 2015 gives a lower ACER of nearly zero percentage. However,
Clarkson 2013 datasets do not perform well and give a high ACER of around 74%.
Fig. 9 shows performance evaluation on cross datasets. It can be seen that our model outperforms
in IIITD and Clarkson 2015 datasets, however, shows low performances for Clarkson 2013 dataset.
One possible reason for this is Clarkson 2013 dataset has a smaller number of images compared to the
other two datasets.
Table 9 represents the performance comparison of fragmental coefficients across all datasets used
for implementation with an average of percent accuracy, percent precision, percent recall, and percent
F-ratio values. The highest performance is represented in bold and underlined. From Table 8, it can
be seen that reducing the number of higher energy coefficients from 128 ∗ 128 to 8 ∗ 8 improves
performance since the common part is reduced and discriminative is emphasized more.
6 Discussion
The proposed experiment was performed using the Haar transform. The fundamental goal of
experimenting with the transform domain is to learn more about how the image is split into low and
340 CMES, 2023, vol.136, no.1
high-energy parts, reducing the feature vector size and speeding up retrieval. The Haar statistic is used
to transform data [27].
By applying Haar transform on Iris images, high energy coefficients of transformed iris images
sized 256 ∗ 256, 128 ∗ 128, 64 ∗ 64, 32 ∗ 32, 16 ∗ 16, 8 ∗ 8, and 4 ∗ 4 do generate feature vectors for
the projected iris-liveness detection. The procedures outlined in Section 3.2 are used to create the
feature vector. Seven distinct ML and ensembles of classifiers are trained using these features. These
classifiers are trained using the tenfold cross-validation method to detect presentation attacks. Three
benchmark datasets are taken for testing: Clarkson 2013, Clarkson 2015, and the IIITD combined
spoofing database. These three datasets explain in Section 4.1. Accuracy, Precision, Recall, and F-ratio
and ISO standard metrics APCER, BPCER, and ACER are utilized to compare the performance of
all the variants of the suggested approach. Section 4.2 describes several performance measures.
As stated in Sections 5.1 and 5.2, feature extraction using Haar has shown outstanding average
classification accuracy. For the Clarkson 2013 dataset, the highest noted iris liveness detection
accuracy comes around 98.10%, with 8 ∗ 8 fragmental coefficients obtained using the classifier RF.
The highest average accuracy, 95.94% was achieved by 16 ∗ 16 fragmental coefficients, whereas the
maximum average accuracy was 96.31%, and the average ACER was around 2.63%, achieved by
SVM+MLP+RF ensembles of classifiers. For Clarkson 2015, the highest accuracy achieved was
99.90% by using an SVM+RF+RT ensemble classifier. For IIITD CSD, the uppermost accuracy of
97.94% was obtained by using the RF classifier. The uppermost average accuracy was 96.80% achieved
by 8 ∗ 8 fragmental coefficients, whereas the utmost average accuracy was 97.05%, and average ACER
was around 2.29%, achieved by SVM+NB+RF ensembles of classifiers. The findings show that our
suggested approach distinguishes between the bonafide and presentation attack images artifacts using
the Haar transform approach. Table 10 shows a comparison of the suggested strategy to recent studies
in this area.
Table 10: The comparison of the prevailed methods with the proposed approach
Author/year Feature Classifiers Performance Outcome [%] Dataset
generation metrics
Arora et al. CNN VGGNet Accuracy (ACC) Acc = 97.98 IIITD
(2021) [34] FAR
LeNet Acc = 89.38
Comparison Omran et al. IRISNet, [KNN, SVM, Accuracy (ACC), Acc = 96.43 IIITD
with same (2020) [35] CNN NB, DT Precision,
datasets F-Measure, and
Recall
(Continued)
CMES, 2023, vol.136, no.1 341
Table 10 (continued)
Author/year Feature Classifiers Performance Outcome [%] Dataset
generation metrics
Fang et al. ResNet50, NA APCER, BPCER ACER = 10.55, IIITD
(2022) [20] VGG16 and ACER ACER = 18.53
MobileNetv3 ACER = 11.41
Wang et al. CNN-Joint CNN, SDH Accuracy (ACC) Acc = 90.71 PolyU bi-spectra
(2019) [37] Bayesian,
CNN-SDH
Compared to similar current techniques based on fragmented energy, the Haar transformation
better discerns between real and artificial artifacts. The results reveal that the proposed method reduces
classification error and gets better accuracy when compared to earlier ways of detecting presentation
attacks using an iris liveness detection. Table 10 summarizes this information. The proposed strategy
outperforms some recent existing studies. As many recent studies used the different train and test
datasets, so we partition Table 10 into two parts, comparing with same datasets and different datasets.
Even though some studies outperform, our approach achieved this performance with reduced feature
vector size (only with 64 features).
While implementing this study, we faced a few challenges-one of them was getting access
to datasets with permissions for use in experimentation. The second challenge we faced was the
explorations of the proposed method with the number of test runs was time-consuming task.
342 CMES, 2023, vol.136, no.1
The limitations of the study are: (i) it was applied only two pre-processing techniques like resizing
and converting an image into grayscale; (ii) this experiment is limited to image size 256 by 256 and
only grayscale images were used during this study.
7 Conclusion
The paper proposed a new method for determining iris liveness. Until now, several approaches
have relied on pre-processing, such as iris segmentation, localization, and normalization; however,
this method of iris liveness detection is computationally intensive. The suggested method employs
Haar transforms on iris images to address this issue, obtaining fragmental coefficients as feature
vectors. The Haar transformed iris image fragmental coefficients are used to train various ML and
ensemble algorithms. Seven criteria are considered to compare the performance of variants of the
suggested approach. Various metrics such as accuracy, precision, recall, f-measure, APCER, BPCER,
and ACER are used to check the performance of the models. Presentation attack images are detected
with 98.10% accuracy in the Clarkson 2013 dataset. The best accuracy for IIITD-CSD was 97.94%.
The experimental results prove the effectiveness of the projected method for detecting iris spoofing
attacks. The study’s main contribution is achieving a good accuracy of 98.10% with lesser feature
vector size by using the fragmental coefficients of the Haar transformed iris image of size 8 ∗ 8
utilizing a RF algorithm with reduced featured vector size. The reduction in considered feature vector
size of iris images with improved accuracy of liveness detection is achieved by exploiting the energy
compaction property of Haar transform in the proposed method. The method is tested on three
available benchmark datasets for validation of results in a generic form. The cross-dataset validations
are performed to prove the worth of the proposed method. The main limitations of this study are as
follows: only two pre-processing techniques were applied, such as resizing and converting an image into
grayscale. Moreover, this experiment is limited to image size 256 by 256 and only grayscale images are
used. In future work, this framework may be extended with the best performance features. Currently,
the presented work is limited to the exploration of explored Haar transform features only. However, the
hybridization of transform using Haar, DCT and Kekare transforms would be an exciting exploration
in the future. Moreover, the proposed framework may be applied for the liveness detection of other
biometric traits, like face, fingerprints, etc. The best performance features a level fusion of fragmental
coefficients of Haar may be added to this framework in future work.
Author Contributions: Data curation: Smita Khade; Writing original draft: Smita Khade; Supervision:
Shilpa Gite, Biswajeet Pradhan; Project administration: Shilpa Gite, Biswajeet Pradhan; Conceptual-
ization: Sudeep Thepade; Methodology: Sudeep Thepade, Shilpa Gite; Validation: Biswajeet Pradhan;
Visualization: Sudeep Thepade, Smita Khade, Shilpa Gite, Biswajeet Pradhan; Resources: Biswajeet
Pradhan, Abdullah Alamri; Review & Editing: Sudeep Thepade, Biswajeet Pradhan; Funding acqui-
sition: Biswajeet Pradhan, Abdullah Alamri.
Funding Statement: The Centre for Advanced Modelling and Geospatial Information Systems
(CAMGIS), Faculty of Engineering and Information Technology, the University of Technology
Sydney, Australia, has funded the research. This research is also partially supported by the Researchers
Supporting Project No. RSP-2021/14, King Saud University, Riyadh, Saudi Arabia.
Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the
present study.
CMES, 2023, vol.136, no.1 343
References
1. Khade, S., Thepade, S. D., Ambedkar, A. (2018). Fingerprint liveness detection using directional ridge
frequency with machine learning classifiers. 2018 Fourth International Conference on Computing Commu-
nication Control and Automation (ICCUBEA), pp. 1–5. Pune, India.
2. Khade, S., Thepade, S. D. (2019). Fingerprint liveness detection with machine learning classifiers using
feature level fusion of spatial and transform domain features. 2019 5th International Conference On
Computing, Communication, Control and Automation (ICCUBEA), pp. 1–6. Pune, India.
3. Khade, S., Ahirrao, S., Phansalkar, S., Kotecha, K., Gite, S. et al. (2021). Iris liveness detection for
biometric authentication: A systematic literature review and future directions. Inventions, 6(4), 65. DOI
10.3390/inventions6040065.
4. Su, L., Shimahara, T. (2019). Advanced iris recognition using fusion techniques. NEC Technical Journal,
13(2), 74–77.
5. Kaur, B., Singh, S., Kumar, J. (2019). Cross-sensor iris spoofing detection using orthogonal features.
Computers & Electrical Engineering, 73(1), 279–288. DOI 10.1016/j.compeleceng.2018.12.002.
6. Khade, S., Ahirrao, S., Thepade, S. (2020). Bibliometric survey on biometric iris liveness detection. Library
Philosophy and Practice, 1–29.
7. Nguyen, K., Fookes, C., Jillela, R., Sridharan, S., Ross, A. (2017). Long range iris recognition: A survey.
Pattern Recognition, 72(9), 123–143. DOI 10.1016/j.patcog.2017.05.021.
8. Khade, S., Gite, S., Thepade, S. D., Pradhan, B., Alamri, A. (2021). Detection of iris presentation attacks
using feature fusion of Thepade’s sorted block truncation coding with gray-level co-occurrence matrix
features. Sensors, 21(21), 7408. DOI 10.3390/s21217408.
9. Kaur, J., Jindal, N. (2019). A secure image encryption algorithm based on fractional transforms and
scrambling in combination with multimodal biometric keys. Multimedia Tools and Applications, 78(9),
11585–11606. DOI 10.1007/s11042-018-6701-2.
10. Choudhary, M., Tiwari, V., Venkanna, U. (2019). An approach for iris contact lens detection and classifi-
cation using ensemble of customized DenseNet and SVM. Future Generation Computer Systems, 101(1),
1259–1270. DOI 10.1016/j.future.2019.07.003.
11. Chen, Y., Zhang, W. (2018). Iris liveness detection: A survey. 2018 IEEE Fourth International Conference on
Multimedia Big Data (BigMM), pp. 1–7. China.
12. Trokielewicz, M., Czajka, A., Maciejewicz, P. (2016). Human iris recognition in post-mortem subjects:
Study and database. 2016 IEEE 8th International Conference on Biometrics Theory, Applications and Systems
(BTAS), pp. 1–6. USA.
13. Fathy, W. S. A., Ali, H. S. (2018). Entropy with local binary patterns for efficient iris liveness detection.
Wireless Personal Communications, 102(3), 2331–2344. DOI 10.1007/s11277-017-5089-z.
14. Gupta, R., Sehgal, P. (2016). A survey of attacks on iris biometric systems. International Journal of
Biometrics, 8(2), 145–178. DOI 10.1504/IJBM.2016.077833.
15. Khade, S., Thepade, S. D. (2018). Novel fingerprint liveness detection with fractional energy of cosine
transformed fingerprint images and machine learning classifiers. 2018 IEEE Punecon, pp. 1–7. Pune, India.
16. Agarwal, R., Jalal, A. S., Arya, K. V. (2020). A multimodal liveness detection using statistical tex-
ture features and spatial analysis. Multimedia Tools and Applications, 79(19), 13621–13645. DOI
10.1007/s11042-019-08313-6.
17. Thavalengal, S., Nedelcu, T., Bigioi, P., Corcoran, P. (2016). Iris liveness detection for next generation
smartphones. IEEE Transactions on Consumer Electronics, 62(2), 95–102. DOI 10.1109/TCE.2016.7514667.
18. Hu, Y., Sirlantzis, K., Howells, G. (2016). Iris liveness detection using regional features. Pattern Recognition
Letters, 82, 242–250. DOI 10.1016/j.patrec.2015.10.010.
19. Czajka, A. (2015). Pupil dynamics for iris liveness detection. IEEE Transactions on Information Forensics
and Security, 10(4), 726–735. DOI 10.1109/TIFS.2015.2398815.
344 CMES, 2023, vol.136, no.1
20. Fang, M., Damer, N., Boutros, F., Kirchbuchner, F., Kuijper, A. (2022). The overlapping effect and fusion
protocols of data augmentation techniques in iris PAD. Machine Vision and Applications, 33(1), 1–21. DOI
10.1007/s00138-021-01256-9.
21. Li, Y. H., Aslam, M. S., Harfiya, L. N., Chang, C. C. (2021). Conditional wasserstein generative adversarial
networks for rebalancing iris image datasets. IEICE Transactions on Information and Systems, 104(9),
1450–1458.
22. Naqvi, R. A., Lee, S. W., Loh, W. K. (2020). Ocular-net: Lite-residual encoder decoder network for accurate
ocular regions segmentation in various sensor images. 2020 IEEE International Conference on Big Data and
Smart Computing (BigComp), pp. 121–124. Busan, Korea.
23. Kimura, G. Y., Lucio, D. R., Britto Jr, A. S., Menotti, D. (2020). CNN hyperparameter tuning applied to
iris liveness detection. arXiv preprint arXiv:2003.00833.
24. Khade, S., Gite, S., Pradhan, B. (2022). Iris liveness detection using multiple deep convolution networks.
Big Data and Cognitive Computing, 6(2), 67. DOI 10.3390/bdcc6020067.
25. Khade, S., Gite, S., Thepade, S. D., Pradhan, B., Alamri, A. (2021). Detection of iris presentation attacks
using hybridization of discrete cosine transform and haar transform with machine learning classifiers and
ensembles. IEEE Access, 9, 169231–169249. DOI 10.1109/ACCESS.2021.3138455.
26. Vyas, R., Kanumuri, T., Sheoran, G., Dubey, P. (2019). Recent trends of ROI segmentation in iris biometrics:
A survey. International Journal of Biometrics, 11(3), 274–307. DOI 10.1504/IJBM.2019.100842.
27. Thepade, S. D., Mhaske, V. (2015). New clustering algorithm for vector quantization using hybrid Haar
slant error vector. 2015 International Conference on Computing Communication Control and Automation,
pp. 634–640. USA.
28. Yambay, D., Doyle, J. S., Bowyer, K. W., Czajka, A., Schuckers, S. (2014). LivDet-iris 2013-Iris liveness
detection competition 2013. IEEE International Joint Conference on Biometrics, pp. 1–8. USA.
29. Yambay, D., Becker, B., Kohli, N., Yadav, D., Czajka, A., (2017). LivDet-iris 2015 – Iris liveness detection
competition 2017. 2017 IEEE International Joint Conference on Biometrics (IJCB), pp. 733–741.
30. Kohli, N., Yadav, D., Vatsa, M., Singh, R., Noore, A. (2016). Detecting medley of iris spoofing attacks using
DESIST. 2016 IEEE 8th International Conference on Biometrics Theory, Applications and Systems (BTAS),
pp. 1–6. USA.
31. Gupta, P., Behera, S., Vatsa, M., Singh, R. (2014). On iris spoofing using print attack. 2014 22nd
International Conference on Pattern Recognition, pp. 1681–1686. USA.
32. Yadav, D., Kohli, N., Doyle, J. S., Singh, R., Vatsa, M. et al. (2014). Unraveling the effect of textured contact
lenses on iris recognition. IEEE Transactions on Information Forensics and Security, 9(5), 851–862. DOI
10.1109/TIFS.2014.2313025.
33. Fang, M., Damer, N., Boutros, F., Kirchbuchner, F., Kuijper, A. (2021). Cross-database and cross-attack iris
presentation attack detection using micro stripes analyses. Image and Vision Computing, 105(12), 104057.
DOI 10.1016/j.imavis.2020.104057.
34. Arora, S., Bhatia, M. P. S., Kukreja, H. (2020). A multimodal biometric system for secure user identification
based on deep learning. International Congress on Information and Communication Technology, pp. 95–103.
Singapore.
35. Omran, M., AlShemmary, E. N. (2020). An iris recognition system using deep convolutional neural network.
Journal of Physics: Conference Series, 1530(1), 012159.
36. Das, P., McFiratht, J., Fang, Z., Boyd, A., Jang, G. et al. (2020). Iris liveness detection competition (livdet-
iris)-the 2020 edition. 2020 IEEE International Joint Conference on Biometrics (IJCB), pp. 1–9. Houston,
TX.
CMES, 2023, vol.136, no.1 345
37. Wang, K., Kumar, A. (2019). Cross-spectral iris recognition using CNN and supervised discrete hashing.
Pattern Recognition, 86(2), 85–98. DOI 10.1016/j.patcog.2018.08.010.
38. Cheng, Y., Liu, Y., Zhu, X., Li, S. (2019). A multiclassification method for iris data based on the
hadamard error correction output code and a convolutional network. IEEE Access, 7, 145235–145245. DOI
10.1109/ACCESS.2019.2946198.
39. Chatterjee, P., Yalchin, A., Shelton, J., Roy, K., Yuan, X. et al. (2019). Presentation attack detection using
wavelet transform and deep residual neural net. International Conference on Security, Privacy and Anonymity
in Computation, Communication and Storage, pp. 86–94. Cham.