Explainable Edge AI A Futuristic Computing Perspective
Explainable Edge AI A Futuristic Computing Perspective
Explainable Edge
AI: A Futuristic
Computing
Perspective
Studies in Computational Intelligence
Volume 1072
Series Editor
Janusz Kacprzyk, Polish Academy of Sciences, Warsaw, Poland
The series “Studies in Computational Intelligence” (SCI) publishes new develop-
ments and advances in the various areas of computational intelligence—quickly and
with a high quality. The intent is to cover the theory, applications, and design methods
of computational intelligence, as embedded in the fields of engineering, computer
science, physics and life sciences, as well as the methodologies behind them. The
series contains monographs, lecture notes and edited volumes in computational
intelligence spanning the areas of neural networks, connectionist systems, genetic
algorithms, evolutionary computation, artificial intelligence, cellular automata, self-
organizing systems, soft computing, fuzzy systems, and hybrid intelligent systems.
Of particular value to both the contributors and the readership are the short publica-
tion timeframe and the world-wide distribution, which enable both wide and rapid
dissemination of research output.
This series also publishes Open Access books. A recent example is the book
Swan, Nivel, Kant, Hedges, Atkinson, Steunebrink: The Road to General Intelligence
https://link.springer.com/book/10.1007/978-3-031-08020-3
Indexed by SCOPUS, DBLP, WTI Frankfurt eG, zbMATH, SCImago.
All books published in the series are submitted for consideration in Web of Science.
Aboul Ella Hassanien · Deepak Gupta ·
Anuj Kumar Singh · Ankit Garg
Editors
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature
Switzerland AG 2023
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
v
vi Preface
vii
viii Contents
Prof. Aboul Ella Hassanien is Founder and Head of the Egyptian Scientific Research
Group (SRGE) and Professor of Information Technology at the Faculty of Computer
and Artificial Intelligence, Cairo University. Professor Hassanien is an ex-dean of the
Faculty of Computers and Information, Beni Suef University. Professor Hassanien
has more than 800 scientific research papers published in prestigious international
journals and over 40 books covering such diverse topics as data mining, medical
images, intelligent systems, social networks, and smart environment. Prof. Hassanien
won several awards, including the Best Researcher of the Youth Award of Astronomy
and Geophysics of the National Research Institute, Academy of Scientific Research
(Egypt, 1990). He was also granted a Scientific Excellence Award in Humanities from
the University of Kuwait for the 2004 Award and received the Scientific University
Award (Cairo University, 2013). Also, he was honored in Egypt as the best researcher
at Cairo University in 2013. He also received the Islamic Educational, Scientific and
Cultural Organization (ISESCO) Prize on Technology (2014) and received the State
Award for excellence in engineering sciences 2015. He was awarded the Medal of
Sciences and Arts of the first class by the President of the Arab Republic of Egypt,
2017.
Deepak Gupta received a B.Tech. degree in 2006 from the Guru Gobind Singh
Indraprastha University, Delhi, India. He received an M.E. degree in 2010 from
Delhi Technological University, India, and Ph.D. degree in 2017 from Dr. A. P.
J. Abdul Kalam Technical University (AKTU), Lucknow, India. He completed his
post-doc from the National Institute of Telecommunications (Inatel), Brazil, in 2018.
He has co-authored more than 207 journal articles, including 168 SCI papers and
45 conference articles. He has authored/edited 60 books, published by IEEE-Wiley,
Elsevier, Springer, Wiley, CRC Press, DeGruyter, and Katsons. He has filed four
Indian patents. He is the convener of the ICICC, ICDAM, ICCCN, ICIIP, and DoSCI
Springer conferences series. He is Associate Editor of Computer and Electrical
Engineering, Expert Systems, Alexandria Engineering Journal, Intelligent Decision
Technologies. He is the recipient of the 2021 IEEE System Council Best Paper Award.
He has been featured in the list of top 2% scientist/researcher databases worldwide.
ix
x About the Editors
Abstract AI is a broad term used in computer science to mimic the human mind.
Machine Learning is also a part of AI and is used in various fields of decision-
making by using different algorithms. In the last few decades, AI has gained more
popularity in the field of Industry, Medicine, Education, Defense, etc. Based on the
algorithm AI is categorized into: Interpretable AI and Explainable AI (XAI). In inter-
pretable AI, experts have problems in understanding the working of the model. They
have no idea of working behind the algorithm so from here the term explainable AI
came into existence. The term Explainable AI is coined by DARPA. This is also
known as Explainable Machine Learning. The term came into existence to provide
transparency in the process of decision-making by different Machine learning algo-
rithms. In the current scenario, XAI is gaining popularity because of its transparent
working. XAI is used in every field where AI can be used but with some modifica-
tion or by adding some techniques of XAI as “SHAP (Shaply Additive exPlanations),
DeepSHAP, DeepLIFT, CXplain, and LIME”. The main goal of this chapter is to
provide a brief overview of XAI by covering almost every aspect of XAI. The chapter
is divided into sections, in the first section brief introduction is discussed. Related
work concerning medicine is stated in Sect. 2. Later other sections will cover princi-
ples, techniques, current state of art, benefits, and, applications. This chapter infers
that XAI is a great development in AI due to its transparent nature. The chapter also
addressed the challenges in XAI, as well as the field’s future potential.
Abbreviations
1 Introduction
Various industrial areas have adopted emerging technologies, and Artificial Intelli-
gence (AI) is at the root of them [1]. While the development of AI can be addressed
back decades, there is general accord that smart systems with learning, reasoning, and
adaptive skills are of critical significance. AI approaches are attaining remarkable
standards of accomplishment by attempting to address extremely hard computations
as a result of these capacities, considering these approaches crucial for future growth
[2]. Furthermore, intelligent agents such as Google Assistant, Siri, and others are
technologies increasingly integrating into individual personal lives. AI models, on
the other hand, have become a black box whose internal workings are not completely
transparent due to the increasing complexity of its algorithms. The conceptions of
explainability and interpretability emerge as a result of this trade-off. Models before
2015 have been viewed as examples of Interpretability AI due to abstraction of their
whole internal working. XAI was introduced by DARPA in 2015 to provide a thor-
ough knowledge of how AI models work. The XAI initiative intends to develop a
set of machine learning approaches that generate more comprehensible models by
retaining excellent learning performance. AI is at the base of several activity areas
that have incorporated modern information technology [3].
Explainability in AI is a discipline that has applications in a range of areas,
including medical treatment, business operations, privacy, economic and ethical
considerations, automated driving, smartphones, and AI for developers [4]. All of
the applications will be explained later in detail in the application section.
The most frequently used terminologies in XAI are:
• Understandability: Describes a model’s ability to convey its function—how
it works—to an individual despite having to elucidate its inner core or the
computational means by which the model processes data internally [5].
• Comprehensibility: If deployed in tandem with machine learning techniques,
comprehensibility alludes to learning algorithm’s capacity to express its learned
facts in a way that is understood by individuals.
• Interpretability: It is described as the potential to express or convey information
to an individual in a comprehensible way.
• Explainability: Explainability is related to the concept of description as a connec-
tion among individuals and a selection provider. The selection provided by the
selection provider can be understandable to individuals [6].
• Transparency: A model is said to be transparent if it can be understood on its own.
1 Explainable Artificial Intelligence: Concepts and Current … 3
The word “machine learning (ML)” was coined from the term “artificial intel-
ligence”. Both of these are interconnected and can be used in the same applica-
tion. Even though their functioning models and algorithms are identical. ML has
increased in popularity in both research and industry, owing to the effectiveness of
its numerous algorithms, especially in the medical field. In certain applications, a
computer glitch or a misinterpretation has some impact, but in the medical profes-
sion, it is linked to human life. Early diagnosis of the disease is often crucial for
patients’ recovery or to reduce the risk of disease from progressing to more severe
stages. Explainable Machine Learning and XAI can be used interchangeably.
The main objective of this chapter is to provide a brief overview of XAI. The
chapter is divided into sections. Section 1 provides an introduction to XAI. Section 2
discussed the current finding of the researchers. Section 3 stated the principles of
XAI. Later sections discussed Models, Current progression, Applications, Benefits,
and Challenges of XAI.
2 Literature Review
Number of work has been done in the field of health care using the XAI model. Some
of the latest research work has been stated in the section to provide a precise overview
of XAI in healthcare. The Research work in the section mainly concentrates on the
medicinal sector.
In Quellec et al. [7] proposed a unique explanatory AI approach for the classifi-
cation of multi-label images for the prediction of diabetes retinopathy. For the clas-
sification, CNN and Pixel-level algorithms were used. Indian Diabetic Retinopathy
Image Dataset was used and Python was used for the experiment. The accuracy of
the proposed model was 93%.
In Wu et al. [8] propose a counterfactual multi-granularity graph supporting facts
extraction method to extract supporting facts from the irregular EMR. The approach
used by them was suitable for Lymphedema. Techniques used by the researcher
were GNN, Counterfactual reasoning. Dataset was acquired from the Local hospital
in china. The result of the anticipated model was 99% for each parameter (precision,
recall, and F1-score value).
In Kavya et al. [9] developed a model for the diagnosis of allergy. Techniques used
in the model were KNN, SVM, C5.0, MLP, AdaBag, RF, and Condition prediction.
Dataset was taken from the allergy testing centers of South India and the experiment
was performed using R-Studio. The result of the anticipated model was 86%, and
75% for Accuracy and Sensitivity respectively.
In Amoroso et al. [10] developed an EXI system based on adaptive dimension
reduction to predict breast cancer. For the classification clustering technique was
used. Dataset was taken from Italy.
In Dindorf et al. [11] design a pathology-independent classifier that presents expla-
nations for the categorization findings and projection probabilities. Two datasets were
used and acquired, the first dataset was part of the dissertation project of Friederike
4 K. Kangra and J. Singh
Werthmann and the second was part of the dissertation project of Claudia Wolf. One-
class SVM, binary RF, and LIME approaches were used. The result of the classifier
were F1-80 ± 12%, MCC-57 ± 23%, BSS-33 ± 28%.
In El-Sappagh et al. [12] designed a model for detecting Alzheimer’s disease
using XAI which was precise and explanatory. In the proposed model two layer
RF, SHAP, and Fuzzy techniques were used. Alzheimer’s Disease Neuroimaging
Initiative dataset was used. The result of the proposed model was at the First layer:
accuracy: 93%, F1-score: 93%, and at the Second layer: Accuracy: 87%, F1-score:
87% respectively.
In Peng et al. [13] developed an XAI model to predict Hepatitis. In the proposed
model LR, DT, KNN, SVM, RF, SHAP, LIME, and partial dependence plots were
used. Dataset was taken from UCI and the experiment was performed using Python
3.6.4. The result showed an accuracy of 91%.
In Sarp et al. [14] proposed a CNN-based XAI model to classify chronic wounds.
CNN and LIME techniques were used. Dataset was taken from ekare Inc. and python
3.6 was used for the experiment. The results of the anticipated model were Precision,
Recall, and F1-score scored 95%, 94%, and 94% respectively.
In Tan et al. [15] develop a deep learning model based on high-resolution computed
tomography scans of the temporal bone to diagnose fenestral otosclerosis. The tech-
nique in the proposed model was LNN. Dataset was taken from the institutional
committee of Eye and ENT Hospital, Fudan University. The result of the anticipated
model was AUC, Sensitivity, and Specificity scored 99%, 96%, and 98% respectively.
From the literature, it has been seen that XAI is used in healthcare to predict and
classify the number of diseases like Diabetes, Spine, Chronic wounds, etc. Different
XAI models such as LIME, SHAP, Fuzzy, Deep attributes, etc. are used to clas-
sify disease. By using the XAI model results of the traditional models have been
increased as seen in the literature. Using these models Accuracy has been increased
to a remarkable extent. So, it can be concluded that XAI has more potential and its
potential should be utilized in the field of health care.
3 Principles
The principles are the collection of rules that should follow by any XAI model. There
are three principles of XAI (see Fig. 1).
4 Models
This section of the chapter covers XAI techniques to provide a precise overview. “XAI
is a research field on ML interpretability techniques whose aims are to understand
ML model predictions and explain them in human and understandable terms to build
trust with stakeholder”. To understand the ML model XAI comes with different
models.
To understand the XAI model user need to understand two terms as shown in
Fig. 2.
To understand the working of the BB, WB is used. XAI came into existence due
to trade-off between “interpretability and explainability. Two approaches: Integrated
6 K. Kangra and J. Singh
Black Box
White Box
SHAP
and explainability, on the other hand, are challenges that must be appropriately
addressed.
Interpretability: The transparent proxy model approach identifies an inter-
pretable model that estimates the black-box model’s predictions generally.
Explainability: For each prediction, these algorithms typically provide a
description in the form of feature significances for concluding. In [21], sensi-
tivity analysis and layer-by-layer relevance propagation were used to explain
deep learning model predictions in terms of input variables.
To make the BB model explainable XAI approaches came into existence as
shown in Fig. 4.
explanation of the Shapley values is not always straightforward if the model is not
additive. Furthermore, while SHAP is model agnostic, optimizing the SHAP algo-
rithm for all model types is not always simple or effective.
CNN’s have their class activation maps (CAMs). The per-class weighted linear
sum of subjective features available at diverse geographical places in a picture is
represented by CAMs [22]. Global average pooling is applied to the last convolutional
feature map of a network on the output layer. The input features for a fully linked
layer, which is then output via a loss function, are then pooled with feature maps.
The portions in the input image that have a stronger effect on the CNNs’ choice are
emphasized per class and apparent through a heatmap visualization by projecting
the output weights back to the earlier convolutional layer. Pre-trained networks and
networks that do not follow the stated fully convolutional network design are not
eligible for CAMs. Additionally, the completely attached layer and map scaling
can cause spatial information to be lost. Grad-CAM and Grad-CAM ++ are two
expansions of the original CAM model that aim to improve the explainability of
CNNs.
Grad-CAM [23] generalizes class activation mapping (CAM) to any random CNN
architecture without training. Any target class’s gradients are delivered to the last
convolutional layer, which computes a significance score based on the gradients. A
heatmap illustration of the Grad-CAM, like other approaches, shows which portions
of the input image were most essential in the CNN’s judgments. GradCAM, on
the other hand, only creates coarse-grained conceptions and is unable to describe
numerous cases of identical objects in a single image.
To circumvent these limitations, Grad-Cam ++ [24] uses the weighted average
of the gradients. Feature-oriented methods reveal where a choice is being achieved
from the perspective of the input, but they lack at human-level description of how
and why the model arrived at those conclusions.
Global attribution mappings techniques on a global scale [25] can interpret as neural
network’s (NN’s) predictions on a comprehensive level, beyond subpopulations, by
framing perception as weighted combined ranks for features with exact semantic
descriptions. The benefits include the ability to capture diverse subpopulations using
a granularity parameter that may be tuned. GAMs employ K-medoids clustering
technique to assemble similar local feature significances into clusters and determine
a pair-wise rank distance matrix among features. As a global attribution, the medoid
of the respective cluster describes the outline found in each cluster. As a result, this
method can be used to investigate features in distinct subpopulations of samples.
Gradient-based saliency maps [26] are mapping approach that displays the abso-
lute value of the dominant predicted class’s gradient (concerning the input features) as
a normalized heatmap. The pixels with the highest activation levels are emphasized,
and they are parallel to the most significant locations. The method’s description is the
1 Explainable Artificial Intelligence: Concepts and Current … 9
capability for an individual to sight which features in the image are being considered
in the categorization of the result. When propagating nonlinear layers, conversely, the
absolute value signifies that the gradient of neurons with negative input is inhibited.
Deep attribute maps are proposed by [27] as an approach for presenting the
generalization ability of gradient-based algorithms. The suggested layout also shows
comparisons between various saliencey-based explanatory techniques. The gradient
of the output is combined with the relevant input to give a heatmap that explains
a model’s prediction. The colors red and blue denote positive and negative depic-
tions of the final choice, respectively. Explanations are affected by noisy gradients
and input variability. Deep attribute is unable to justify what is the reason that two
models are providing similar or dissimilar results.
In [28] concept activation vectors (CAVs) are the strategy for explaining the global
internal states of NN’s neurons to the NN’s high-level latent properties. As a result,
CAVs define the notch at which these conceptual attributes refer to a series of human
ideas that the user has chosen. Human bias is unavoidable, but by illuminating the
corresponding perception, any errors in the model’s decision-making process can be
addressed, such as if some features are erroneously considered important. Beyond
that, automatic concept-based explanations [29] extract CAVs without the need for
human intervention, eliminating human bias. Human intelligible perceptions are
split at numerous spatial dimensions from in-class visuals rather than being chosen.
Concept-based strategies, on the other hand, rely on the perceptions being unique to
the class, and the efficiency of description is impaired if the same perception is used
in various sessions.
5 Current Progression
AI started in 1950. After that, a new era of human intelligence started. In 1963, the
term Machine Learning was coined to train machines. By using ML automated era of
machines begins, machines can perform analysis tasks by themselves. ML increases
accuracy in analysis tasks by reducing human error. Both ML and AI are used in
different sectors. But in 1976, Deep learning (DL) came into existence, especially
for medical services. But AI, ML, and DL have problem that their inner working
can’t be explained. From there two terms as explained at beginning of this chapter
came into existence.
In 2015, the term XAI was coined to explain the inner working of different models
of AI. It uses different approaches explained above to make the model’s working
transparent. Now XAI is used in many sectors.
According to the current status of XAI, it is employed in a variety of industries like
finance, manufacturing, défense, stock, and healthcare. It is gaining popularity in the
field of health care due to its nature and characteristics. It is utilized by medical experts
in healthcare to gain greater precision in decision-making processes, as there is a need
for more accuracy in the medical area. Its various strategies combine with existing
artificial intelligence techniques to deliver more accurate results. It is mostly utilized
in medical image processing for further analysis. In the sector of image processing,
it helps by using saliency maps. For image classification, Deep attribute maps are
considered the best choice. These maps are important for determining which parts
of a picture are considered by the model for prediction and which are not to make an
accurate prediction. In the sector of Video threat detection, it is used to detect any
forgery in the video.
1 Explainable Artificial Intelligence: Concepts and Current … 11
6 XAI in Medicines
The combination of XAI and healthcare is a powerful digital technology. The emer-
gence of AI-based diagnosis tools has sparked serious doubts regarding the accuracy
of AI-based outcomes. In the medical and healthcare sectors of the virtual sector,
trust is critical to AI’s success. If-else diagnostic models are organically explainable
since they are made up of feature value sets and deliver a score based on a feature
value of a specific case of disease detection. External symptomatic disease diag-
nosis is highly suited for if-else-based explainable medical diagnosis systems. If the
person’s symptoms list has a certain number of matching criteria using If-else-based
feature values, it can be determined if the person had a particular disease or not.
For instance, if the patient has a prior medical history. For example, if the patient
has a history of respiratory infection and cough, the odds of developing asthma are
increased. To cover a wide range of XAI, an explanation that is independent of the
AI model is required. Model agnostic are XAI techniques. A model agnostic method
is LIME. LIME is a framework for quantifying the weights of all aspects that go into
reaching a decision or making prediction. Other model-independent XAI approaches
exist, such as SHAPLEY [31]. Deep learning is a valuable tool for precise medical
diagnosis, but its black-box approach to prediction and conclusion limits it to a few
crucial areas of medicine. Some applications of XAI in medicine are shown in Fig. 5.
There are 7 pillars of XAI in medicine which are discussed below [32]. These are
important in the implantation of the XAI models and will help to make any model
explainable in medicine.
• Transparency: Understandable by any user.
• Domain Sense: The explanation should make sense in the context of the
application and to the system’s user.
• Consistency: Across all models, the explanation should be consistent.
• Parsimony: It’s best if the explanation is as straightforward as possible.
12 K. Kangra and J. Singh
XAI in
Medicine
Virtual R&D in
Monitoring Drug
Surgery
7 Applications of XAI
In theory, the scope of XAI should be as broad as that of AI. Natural language
processing (NLP), health care, engineering, and military are just a few of the impor-
tant areas. NLP and engineering are used in banking, finance, digitization, and
automation [33].
• Data Security: The European Union and its data protection regulatory agency have
a “right to explanation” clause, that helps XAI algorithms to come to a decision
without breaching data security.
• Medical: Based on a patient’s medical history, XAI can provide a diagnosis. The
use of AI/ML techniques in the healthcare image processing area makes it quicker
for doctors to spot cancer tumors and other lung diseases in patients.
• Défense: As a result of computerized artillery and monitoring systems, XAI is
becoming more important in defensive activity. During battle mode training and
real conflict strategy, XAI also gives excellent second-hand help.
1 Explainable Artificial Intelligence: Concepts and Current … 13
• Banking: One of the most important financial industries with the largest impact
on people’s lives is banking. Every day, a large number of fraudulent transactions
and conns by cheaters take place. Well-trained XAI models can help with both
the investigation and decrease false positives in fraudulent transactions.
8 Benefits of XAI
The objective of XAI is to describe how certain outcomes or suggestions are made. It
explains why AI performs as it does and fosters trust among people and AI models.
The main advantages of XAI are [34]:
• Improved explainability and transparency: Businesses can have a deeper under-
standing of advanced AI models and they act in various ways under certain circum-
stances. Humans can utilize an explanation interface to comprehend how these
AI models arrive at particular results, even if it’s a black-box model.
• Faster adoption: As organizations expand their understanding of AI models, they
will be able to entrust them with more crucial judgments.
• Improved debugging: When a system behaves strangely, XAI can be used to
diagnose the problem and assist developers in troubleshooting.
• Allowing for audits to meet regulatory obligations.
9 Challenges in XAI
10 Future of XAI
The global XAI market is expected to grow at a CAGR of 18.4% from 2022 to 2030,
from USD 4.4 billion in 2021 to USD 21.0 billion in 2030. Explainable AI is an
artificial intelligence method or technique in which the solution can be evaluated and
understood by humans. It differs from standard ML techniques, in which researchers
frequently fail to comprehend why the system has reached a particular conclusion.
Healthcare, retail, media and entertainment, aerospace, and défense are among the
areas that are embracing XAI. For example, in the retail industry, XAI is critical for
anticipating new fashion trends with reason, allowing retailers to stock and display
the latest products in their stores. Furthermore, in the e-commerce industry, the
notion of recommendation engines is an example of XAI, in which the system offers
things to the consumer based on his or her search history. Additionally, recommen-
dation engines promote accessories/additional products in addition to the one being
purchased. This is accomplished through the use of explainable AI, which allows the
system to recommend things that are beneficial to the customer [36].
Future investigations of medical XAI applications should include medical special-
ists in the design and development stages. Interdisciplinary collaboration is required
for a good medical XAI application. Medical experts, in particular, should provide
their prior medical expertise, as well as suggestions and feedback, to help enhance
the layout of AI algorithms. Medical XAI apps should be able to support doctors in
drawing explainable clinical inferences, according to AI experts and data scientists.
As a result, we expect that medical XAI models will gain acceptance in the medical
field.
11 Conclusion
XAI is, without a dispute, an important collaborative study in the AI system. The
ethical and quality-of-life considerations of utilizing AI in its existing situation in
real-world circumstances are problematic. This chapter delivers insight into XAI,
which has recently been identified as a critical necessity for ML methodologies to
be embraced in real-world applications. This chapter has focused on this topic by
highlighting the search for better interpretable ML algorithms, as well as discussing
1 Explainable Artificial Intelligence: Concepts and Current … 15
the principles that drive model explainability. In this chapter, various XAI techniques
such as LIME, SHAP, DEEPSHAP, etc. are covered. Every technique has its own
merits and demerits. In Shap, Shap values are calculated, based on those values
researchers will come up with a decision. If there is an error in the calculation of the
Shap value that will affect the decision of the researcher. This chapter also discussed
different applications and their need in the medical field and their future perspective
in every field especially concentrating on the medical field. By looking at the whole
scenario XAI is a powerful technique of AI. It can be concluded that this can act as
a game changer in the healthcare sector due to its transparent nature. Physicians can
easily make accurate decisions without the intervention of any technical expert.
References
1. D.M. West, The future of work: robots, AI, and automation. Futur. Work Robot. AI Autom.
1–205 (2018)
2. A. Barredo Arrieta et al., Explainable artificial intelligence (XAI): concepts, taxonomies, oppor-
tunities and challenges toward responsible AI. Inf. Fusion 58, 82–115 (2020). https://doi.org/
10.1016/j.inffus.2019.12.012
3. L.F. Huang, Artificial Intelligence, vol. 4 (2010)
4. Explainable AI—India | IBM. https://www.ibm.com/in-en/watson/explainable-ai. Accessed 12
June 2022
5. G. Montavon, W. Samek, K.R. Müller, Methods for interpreting and understanding deep neural
networks. Digit. Signal Process. A Rev. J. 73, 1–15 (2018). https://doi.org/10.1016/j.dsp.2017.
10.011
6. S.T. Mueller, R.R. Hoffman, W. Clancey, A. Emrey, G. Klein, Explanation in human-AI
systems: a literature meta-review. Def Adv Res Proj Agency 204 (2019)
7. G. Quellec, H. Al Hajj, M. Lamard, P.H. Conze, P. Massin, B. Cochener, ExplAIn: explanatory
artificial intelligence for diabetic retinopathy diagnosis. Med. Image Anal. 72(2018) (2021).
https://doi.org/10.1016/j.media.2021.102118
8. H. Wu, W. Chen, S. Xu, B. Xu, Counterfactual supporting facts extraction for explainable
medical record based diagnosis with graph network 1942–1955 (2021). https://doi.org/10.
18653/v1/2021.naacl-main.156
9. R. Kavya, J. Christopher, S. Panda, Y.B. Lazarus, Machine learning and XAI approaches for
allergy diagnosis. Biomed. Signal Process. Control 69, 102681 (2021). https://doi.org/10.1016/
j.bspc.2021.102681
10. N. Amoroso et al., A roadmap towards breast cancer therapies supported by explainable artificial
intelligence. Appl. Sci. (2021). https://doi.org/10.3390/app11114881
11. C. Dindorf et al., Classification and automated interpretation of spinal posture data using
a pathology-independent classifier and explainable artificial intelligence (XAI). Sensors
(Switzerland) 21(6323) (2021). https://doi.org/10.3390/s21186323
12. S. El-Sappagh, J.M. Alonso, S.M.R. Islam, A.M. Sultan, K. Sup, A multilayer multimodal
detection and prediction model based on explainable artificial intelligence for Alzheimer’s
disease (2021)
13. J. Peng et al., An explainable artificial intelligence framework for the deterioration risk predic-
tion of hepatitis patients. J. Med. Syst. 45(5) (2021). https://doi.org/10.1007/s10916-021-017
36-5
14. S. Sarp, M. Kuzlu, E. Wilson, U. Cal, O. Guler, The enlightening role of explainable artificial
intelligence in chronic wound classification. Electronics 1406 (2021). https://doi.org/10.3390/
electronics10121406
16 K. Kangra and J. Singh
15. W. Tan et al., The use of explainable artificial intelligence to explore types of fenestral otoscle-
rosis misdiagnosed when using temporal bone high-resolution computed tomography. Ann.
Transl. Med. 9(12), 969–969 (2021). https://doi.org/10.21037/atm-21-1171
16. D. Castelvecchi, The black box 2 0 |. Nature 538(7623), 20–23 (2016). [Online]. Available:
http://www.nature.com/news/can-we-open-the-black-box-of-ai-1.20731
17. L.H. Gilpin, D. Bau, B.Z. Yuan, A. Bajwa, M. Specter, L. Kagal, Explaining explanations: an
overview of interpretability (2019). [Online]. Available: https://paperswithcode.com/methods/
category/interpretability
18. P.J. Phillips et al., NISTIR 8312 four principles of explainable artificial intelligence (2021)
19. F.K. Dosilovic, M. Brcic, N. Hlupic, Explainable artificial intelligence: a survey, in 2018 41st
International Convention on Information and Communication Technology, Electronics and
Microelectronics MIPRO 2018—Proceedings, pp. 210–215 (2018). https://doi.org/10.23919/
MIPRO.2018.8400040
20. S.M. Lundberg, S.I. Lee, A unified approach to interpreting model predictions. Adv. Neural
Inf. Process. Syst. 2017(Section 2), 4766–4775 (2017)
21. W. Samek, T. Wiegand, K.-R. Muller, Explainable artificial intelligence: understanding,
visualizing and interpreting deep learning models (2017)
22. B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, A. Torralba, Learning deep features for discrim-
inative localization, in Proceedings IEEE Computer Social Conference on Computer Vision
and Pattern Recognition, vol. 2016, pp. 2921–2929 (2016). https://doi.org/10.1109/CVPR.201
6.319
23. R.R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, D. Batra, Grad-CAM: visual
explanations from deep networks via gradient-based localization, in Proceedings of the IEEE
International Conference on Computer Vision, vol. 2017, pp. 618–626 (2017). https://doi.org/
10.1109/ICCV.2017.74
24. A. Chattopadhay, A. Sarkar, P. Howlader, V.N. Balasubramanian, Grad-CAM++: generalized
gradient-based visual explanations for deep convolutional networks, in Proceedings—2018
IEEE Winter Conference on Applications of Computer Vision, WACV 2018, vol. 2018, pp. 839–
847 (2018). https://doi.org/10.1109/WACV.2018.00097
25. M. Ibrahim, M. Louie, C. Modarres, J. Paisley, Global explanations of neural networks 279–287
(2019). https://doi.org/10.1145/3306618.3314230
26. K. Simonyan, A. Vedaldi, A. Zisserman, Deep inside convolutional networks: visualising
image classification models and saliency maps, in 2nd International Conference on Learning
Representations ICLR 2014—Workshop Track Proceedings, pp. 1–8 (2014)
27. M. Ancona, E. Ceolin, C. Öztireli, M. Gross, Towards better understanding of gradient-based
attribution methods for deep neural networks (2018)
28. B. Kim et al., Interpretability beyond feature attribution: quantitative testing with concept
activation vectors (TCAV), in 35th International Conference on Machine Learning ICML 2018,
vol. 6, pp. 4186–4195 (2018)
29. A. Ghorbani, J. Zou, J. Wexler, B. Kim, Towards automatic concept-based explanations, in
NeurIPS (2019)
30. J. Dieber, S. Kirrane, Why model why? Assessing the strengths and limitations of LIME (iii)
(2020). [Online]. Available: http://arxiv.org/abs/2012.00093
31. P. Gohel, P. Singh, M. Mohanty, Explainable AI: current status and future directions, 1–16
(2021). [Online]. Available: http://arxiv.org/abs/2107.07045
32. C.E. Muhammad Aurangzeb Ahhmad, A. Teredesai, V. Kumar, K. Inc., Explainable-AI-in-
healthcare (2021)
33. XAI (Explainable AI) & top 5 use cases—GPU ON CLOUD. https://gpuoncloud.com/xai-exp
lainable-ai-top-5-use-cases/. Accessed 12 June 2022
34. Explainable AI (XAI) in 2022: Guide to enterprise-ready AI. https://research.aimultiple.com/
xai/. Accessed 12 June 2022
1 Explainable Artificial Intelligence: Concepts and Current … 17
35. Explainability of AI: The challenges and possible workarounds | by Rohitha Elsa Philip |
Medium. https://medium.com/@rohithaelsa/explainability-of-ai-the-challenges-and-possible-
workarounds-14d8389d2515. Accessed 12 June 2022
36. S.T. Mueller, R.R. Hoffman, W. Clancey, A. Emrey, G. Klein, Explanation in human-
AI systems: a literature meta-review (2021). https://www.nextmsc.com/report/explainable-ai-
market. Accessed 12 June 2022
Chapter 2
Explainable Artificial Intelligence (XAI):
Understanding and Future Perspectives
Megha Gupta
Abstract Currently, we are seeing a rapid and broad acceptance of artificial intel-
ligence (AI) in our everyday lives, which is contributing to the acceleration of the
transition towards a more algorithmic society. The availability of enormous datasets,
as well as recent advancements in deep learning methods, are allowing AI systems to
perform at or even beyond the level of human performance on an expanding variety
of challenging tasks. However, because of their layered non-linear structure, these
strong models have traditionally been regarded as “black boxes,” since they provide
no information about how they arrive at their predictions and hence cannot be used
trustily to generate predictions. They are often opaque, but they are capable of making
accurate predictions that cannot be explained in any other way. All of these choices
are having an increasing amount of effect on the lives of individuals. The distrust of
totally non-human, autonomous artificial intelligence systems is established as the
cornerstone of the movement. The root of distrust lies in a lack of knowledge as to
why intelligent systems make certain decisions in certain situations. As a result, this
problem has sparked a fresh discussion over explainable artificial intelligence (XAI).
Several lines of research work have since picked up on the definition, understanding,
and implementation of explainability, including expert systems, machine learning,
recommender systems, and approaches to neural-symbolic learning and reasoning,
all of which have occurred primarily during different periods in the history of artificial
intelligence. This chapter serves as an introduction for academics and practitioners
who are interested in learning about essential parts of the new and quickly increasing
field of research on XAI. The first section, titled “Introduction,” provides an overall
summary of the Explainable Artificial Intelligence. Section 2 describes the need of
trust and transparency in AI, which is what led to the development of the idea of XAI.
Section 3 discusses the many approaches that contribute to the functioning of XAI.
The principal applications domain of Explainable Artificial Intelligence is the topic
M. Gupta (B)
Department of Computer Science, MSCW, University of Delhi, New Delhi, India
e-mail: [email protected]
Abbreviations
AI Artificial intelligence
ANN Artificial neural network
CAM Class activation mapping
DeConvNet Deconvolution neural network
HCI Human computer interaction
IG Integrated gradients
LIME Local interpretable model-agnostic explanations
LRP Layer-wise relevance backpropagation
ML Machine Learning
SHAP SHapley Additive exPlanations
SR Salient relevance
VQA Visual question answering
XAI Explainable artificial intelligence
1 Introduction
End users may better understand, trust, and control the next generation of AI systems
with the help of the XAI program’s new or updated ML approaches, which are
designed to develop explainable models. XAI is aimed for the end user who relies on
AI systems to make decisions or suggestions, or take action, and needs to understand
the system’s reasoning. If, for example, a big data analytics system recommends that
an intelligence analyst follow up on a given behavior, the analyst has to know why.
It’s also important for an operator to grasp the system’s decision-making process if
he or she wants to deploy an autonomous vehicle for future missions.
AI that is capable of producing explanations for how it works is dubbed XAI
(explainable Artificial Intelligence). The difference between present AI process and
the contribution of XAI to it could be easily understood through Figs. 1 and 2.
The purpose of explanations is to aid users in maintaining and effectively using the
system model, as well as to assist users in debugging the model to avoid and correct
wrong conclusions. Explanations, on the other hand, can be used for instructional
purposes as well as help individuals discover and grasp new concepts in an application
2 Explainable Artificial Intelligence (XAI): Understanding … 21
Explainable Output to
User
Model
area. It’s important to keep in mind that explanations play an important role in
persuading users that the system’s decisions are best for them.
The goal of an explainable artificial intelligence (XAI) system is to make its
behavior more understandable to humans by offering explanations for its actions.
There are several broad ideas that may be used to the development of more effective
and human comprehensible artificial intelligence systems. When asked to describe
its capabilities and understandings, the XAI system should be proficient to describe
what it has done so far and what it plans to accomplish in future; it should also be
able to divulge the relevant information on which it is acting. While explainability in
artificial intelligence has lately garnered a great deal of attention, the beginnings of
this line of research may be traced back many decades to a time when AI systems were
primarily created as (knowledge-based) expert systems and other similar systems.
The authors of works [1–3] provide a summary of the increased activity in XAI
across a variety of fields and industries. The authors of [1] highlight how important
it is to support the output of a model, e.g., in precision medicine, where specialists
need significantly more information from the model to support their diagnosis than
a simple binary prediction. The author of study [2] stressed that there is a trade-
off between the performance of a model and its transparency, and that this is a
very important consideration. It is a common misconception to believe that if one
just concentrates on performance, the resulting systems would become opaquer.
22 M. Gupta
Most of today’s applications now use AI black box systems. Transparency and
explainability are not essential if the overall performance of the system is good
enough with the machine learning model deployed. When these technologies fail,
the repercussions aren’t anything out of the ordinary, such as the mobile phone camera
not being able to identify a person or the translation service producing grammatical
2 Explainable Artificial Intelligence (XAI): Understanding … 23
errors. Thus, the openness and trust requirements for these AI systems are rather
low. The scenario is substantially different in critical applications. There may be a
problem with the lack of transparency of ML approaches in this case. It may not be
an option to rely on a data-driven system whose rationale is incomprehensible if a
single bad decision can result in threat to life and health of humans e.g., in case of
autonomous driving or in medical field or it may lead to huge financial losses e.g.,
in algorithmic trading etc. Because of this lack of transparency, machine learning is
being used cautiously in industries like healthcare, e-commerce, entertainment etc.
To build trust, an AI system must be able to explain its reasoning behind its actions.
This is crucial, not just in medical or safety-sensitive applications.
When compared to an incomprehensible intelligent system, one of the most
urgent goals of a XAI system is to help end users comprehend how the intelligent
system functions. Users’ mental models of the underlying intelligent algorithms are
improved by machine learning explanations, which provide understandable trans-
parency for the complicated intelligent algorithmic structures. Transparency in a
XAI system may also enhance user experience by making model outputs easier to
comprehend and interact with.
When utilized correctly, XAI may identify and dissuade hostile cases [9] while
also increasing transparency and fairness. For the sake of assessing the accuracy of
production projections and fending off competitors, openness is essential. Using an
adversarial example to deceive a classifier into thinking a phoney picture is in fact
genuine might have a negative impact on the accuracy of the classifier. The reliability
of AI algorithms to minimize threats [10] and give transparency in terms of model
comprehension, textual or visual reporting should be of primary relevance as we
increasingly depend on autonomous algorithms to assist our everyday lives.
In order to build confidence in the intelligent algorithm, XAI system provides
explanations to the end users. System dependability and accuracy may be assessed
and calibrated using a XAI system. Consequently, users’ faith in the algorithm leads
to their dependence on the system. Recommendation systems, autonomous systems,
and crucial decision-making systems are all examples of how XAI intends to increase
user trust via its open architecture.
The information and accessible answers to circumstances, as well as the trust we
establish, are the primary determinants of our social lives, choices, and judgments.
Explanations and rationales are preferable to blind faith in one’s own wisdom when
it comes to making decisions that aren’t ideal. In order to build trust with end-
users, including specialists, developers, legislators, and others, it is critical to explain
why a certain choice was taken. As we move towards a networked AI-driven socio-
economic ecosystem, stakeholders and government agencies need to be able to trust
the predictions of classifiers.
24 M. Gupta
3 Methods of Explainable AI
The explainable model’s core algorithmic principle may be divided into many subcat-
egories dependent on how it was implemented as shown in Fig. 3. XAI approaches
may be divided into two categories: those that concentrate on changes or adjust-
ments in input data, and those that focus on the model’s architecture and parameters.
Both perturbation-based and backpropagation-based approaches may be used. This
section summarizes the most important approaches to explainable AI.
3.1 Perturbation-Based
XAI Methods
LIME CAM
SHAP SR Maps
Prediction Axiomatic
Difference Analysis Attribution Maps
Randomized Input
Sampling PatternNet
Explanation
Nowadays industries are adopting the environment in which the demand for IoT
devices XAI, on the other hand, may have a big impact on a wide variety of industries
that depend on AI systems as shown in Fig. 4. In this section, we look at some of the
possible areas in which explainable models is proving to be more useful.
(1) Transportation: Automated cars have the potential to reduce traffic fatalities
and increase mobility, but they also present issues in terms of AI decision-
making that must be explained. Object classification is critical in autonomous
Transportation Healthcare
Defence
Finance
XAI
Academic
Research Construction
cars because they must make split-second choices based on what they see. When
a self-driving automobile exhibits unexpected behavior due to a coding error.
The results might be disastrous. This isn’t science fiction; a self-driving Uber
in Arizona just killed a lady. In this case, it was the first reported death of a
completely automated car. Unidentified individuals said that the car’s software
detected an item in front of it but handled it as if it were a plastic bag or a piece of
tumbleweed blown by the wind as such, it is only via an understandable structure
that such an incident may become clearer and prevent it from occurring. It is
possible that the XAI might be used in transportation. There is still a long way
to go before self-driving car behavior can be explained.
(2) Healthcare: Human life depends on the medical diagnostic model. How can
we have faith in a black-box model’s instructions to treat a patient? About
two decades ago, a computer model called an artificial neural network (ANN)
was developed to help doctors decide whether or not to admit patients with
pneumonia to a hospital or treat them as outpatients. Neural networks were
shown to be significantly more accurate than traditional statistical approaches
in the early stages of research. In the end, it was discovered that the neural net
had incorrectly predicted that patients with asthma who had pneumonia would
have a reduced mortality rate and hence should not be hospitalized. Although
this may seem paradoxical, the training data showed that asthma patients with
pneumonia were admitted to the ICU, intensively treated, and survived [27]. It
was subsequently concluded that the AI system should be abandoned since it
was too risky to deploy in clinical practice. In order to avoid this critical issue,
we must read the model. Preliminary effort has been done recently to make
AI-based healthcare systems understandable [27, 28]. According to the rising
quantity of these papers, there is an increasing interest and difficulty for utilizing
XAI techniques in health care.
(3) Law and Justice: AI has the ability to improve recidivism risk assessments
and lower the costs of both crime and jail in the criminal justice system. It is,
nevertheless, important to ensure that the criminal decision model is fair, honest,
and non-discriminatory when it is used to estimate the likelihood of reoffending
in court. Mr. Loomis was sentenced to jail because of the use of proprietary
risk assessment software in Loomis v. Wisconsin [29]. Gender and ethnicity
are taken into consideration while using “Correctional Offender Management
Profiling for Alternative Sanctions: COMPAS,” the software at issue in this
case. It was unclear to the judge how the algorithms were employed and what
the causal audit procedure was. Despite the importance of being able to explain
the process by which a judgement is reached in the legal system, relatively few
attempts have been undertaken so far to do so [30–32].
(4) Finance: The use of artificial intelligence (AI) techniques in the financial
services industry may lead to advancements in wealth management, investment
advising, and client service. However, there are concerns about data security and
equitable lending with these new technologies. Loan providers have to adhere to
strict regulations in the financial business. Using AI-based algorithms in credit
scores and models presents a big problem since it is more difficult to offer
2 Explainable Artificial Intelligence (XAI): Understanding … 29
borrowers with the necessary “reason code”—the explanation of why they were
refused credit. Because of an opaque ML algorithm’s output, denial is much
more difficult. At Equifax and Experian, exciting research efforts are underway
to develop automated explanation codes and make AI credit score conclusions
more understandable and auditor friendly [33].
(5) Defense: Researchers in the military were the ones who started the present, well-
known XAI project [34], and the topic’s increased awareness is primarily due to
the call for study and the solicitation of DAPRA Projects. The AI explainability
dilemma is not exclusive to AI in the military. A paper from the MIT Technology
Review examines the limitations of depending on autonomous systems for mili-
tary operations, [35] Knight. This may lead to ethical and legal quandaries that
are reminiscent to those encountered in the healthcare industry. With the DAPRA
Ambitious XAI programme and a few other research efforts, the academic AI
research community is strongly represented in this application space. XAI’s
requirement is confirmed by the work done in each of the mentioned fields.
These kinds of studies, however, are still in their infancy, and much more study
has to be done. Other potential uses for XAI include cybersecurity, educa-
tion, entertainment, governance, and image identification, to name just a few.
Automated decision-making can cause harm in many different areas of life,
including employment, insurance, and social benefits, as well as housing and
differential pricing of goods and services. Future of Privacy Forum [36] illus-
trates the various areas where automated decision-making can do harm as well
as the areas where providing automated explanation can turn them into trustful
processes.
where this knowledge is critical, such as finding groupings of brain areas that work
together to solve a certain task (brain networks) rather than merely pinpointing the
most significant single voxels in each region.
In addition, explanations are too simplistic. In a heatmap, you can see which pixels
are most essential without having to think about more abstract ideas like what is in the
picture or what the scene looks like. To comprehend the model’s behavior, humans
must interpret the explanations. One of the trickiest parts of this process is interpreting
what you’ve seen. Model behavior should be explained on a more abstract, human
intelligible level by meta-explanations that collect information from these low-level
heatmaps. Low-level explanations have been gathered and the semantics of brain
representations have been quantified. Meta-explanations are a promising field of
study for future researchers.
Primary research areas in which enabling technologies and approaches for XAI
might possibly be found are:
(1) Data science: Algorithms based on AI/ML are data hoarders, always searching
for new sources of information to feed their insatiable appetite for data. This
data is crucial to the backward route that aims to create better explanations
and justifications. As a result, data science becomes a critical component in the
process of explaining things.
(2) AI/ML: To create an explanation, we believe that utilizing AI/ML as a
computational process to explain AI/ML is an intriguing work path to follow.
(3) Nature of Human Conduct: For the purpose of creating artificial explanations,
it is worthwhile to first simulate how people explain their actions and behaviors
to one another. As a result, using human science ideas may result in new and
better models that can be explained.
(4) Human Computer Interaction (HCI): The way a person interacts with a computer
may influence how well he or she understands and trusts the system. It is possible
to apply approaches from the discipline of Human–Computer Interaction (HCI)
to assist and build systems that are more transparent.
Research on the use of explanations in the human–machine interface is a promising
avenue for future development. Research on human elements in explainable AI
has already begun (e.g., [8]). Prerequisites for effective human–machine interac-
tion include constructing explanations with the correct user emphasis, which means
asking the right questions in a certain manner. In spite of this, additional research is
needed to optimize explanations for best human use. There is a lack of a formal and
widely agreed-upon definition of what constitutes an explanation in the notion of
explainable AI. The development of mathematically sound explanation techniques
was a first step in this direction. According to [37], for example, the explanation issue
is approached using Taylor decomposition as a theoretical framework. It’s another
potential avenue toward the objective of building a general theory of explainable AI
using axiomatic techniques [38].
In the end, the utilization of explanations that go beyond imagery is an
unsolved problem. To increase the model’s performance or minimize its complexity,
2 Explainable Artificial Intelligence (XAI): Understanding … 31
future work should demonstrate how explanations may be included into a wider
optimization process.
6 Conclusion
References
1. E. Tjoa, C. Guan, A survey on explainable artificial intelligence (XAI): towards medical XAI.
arXiv:1907.07374 (2019)
2. F.K. Dosilovic, M. Brcic, N. Hlupic, Explainable artificial intelligence: a survey, in 41st
International Convention on Information and Communication Technology, Electronics and
Microelectronics (MIPRO), pp. 210–215 (2018)
3. R. Guidotti, A. Monreale, S. Ruggieri, F. Turini, F. Giannotti, D. Pedreschi, A survey of methods
for explaining black box models. ACM Comput. Surv. 51(5), 1–42
4. R. Guidotti, A. Monreale, S. Ruggieri, F. Turini, F. Giannotti, D. Pedreschi, A survey of methods
for explaining black box models. ACM Comput. Surv. 51(5), 1–42 (2018)
5. U. Bhatt, A. Xiang, S. Sharma, A. Weller, A. Taly, Y. Jia, J. Ghosh, R. Puri, J.M.F. Moura, P.
Eckersley, Explainable machine learning in deployment, in Proceedings of the 2020 Conference
on Fairness, Accountability, and Transparency (Association for Computing Machinery, New
York, NY, 2020), pp. 648–657. Retrieved from https://doi.org/10.1145/3351095.3375624
6. M. Ribera, A. Lapedriza, Can we do better explanations? A proposal of user-centered explain-
able AI, in Joint Proceedings of the ACM IUI 2019 Workshops Colocated with the 24th ACM
Conference on Intelligent User Interfaces (ACM IUI 2019), vol. 2327 (2019). CEUR-WS.org
32 M. Gupta
7. A.B. Arrieta, N.D. Rodriguez, J.D. Ser, A. Bennetot, S. Tabik, A. Barbado, S. Garcia, S. Gil-
Lopez, D. Molina, R. Benjamins, R. Chatila, F. Herrera, Explainable artificial intelligence
(XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion
58, 82–115 (2020). Retrieved from https://doi.org/10.1016/j.inffus.2019.12.012
8. I. Lage, E. Chen, J. He, M. Narayanan, B. Kim, S. Gershman, F. Doshi-Velez, An evaluation
of the human-interpretability of explanation. arXiv preprint arXiv:1902.00006 (2019)
9. I. Goodfellow, J. Shlens, C. Szegedy, Explaining and harnessing adversarial examples, in
Proceedings of the 3rd International Conference on Learning Representations, ICLR (2015)
10. A. Kurakin, I. Goodfellow, S. Bengio, Adversarial machine learning at scale, in Proceedings
of the 5th International Conference on Learning Representations, ICLR (2016)
11. D.M. Zeiler, R. Fergus, Visualizing and understanding convolutional networks, ed. by D. Fleet,
T. Pajdla, B. Schiele, T. Tuytelaars. Computer Vision—ECCV 2014 (Springer International
Publishing, Cham, 2014), pp. 818–833
12. T.M. Ribeiro, S. Singh, C. Guestrin, Why should I trust you? in Proceedings of the 22nd ACM
SIGKDD International Conference on Knowledge Discovery and Data Mining (New York,
USA, 2016), pp. 1135–1144
13. M.S. Lundberg, I.S. Lee, A unified approach to interpreting model predictions, in Advances in
Neural Information Processing Systems, pp. 4765–4774 (2017)
14. M.L. Zintgraf, S.T. Cohen, T. Adel, M. Welling, Visualizing deep neural network decisions
prediction difference analysis, in Proceedings of the 5th International Conference on Learning
Representations, ICLR (2017)
15. V. Petsiuk, A. Das, K. Saenko, RISE: randomized input sampling for explanation of black-box
models, in British Machine Vision Conference BMVC (2018)
16. C. Burns, J. Thomason, W. Tansey, Interpreting black box models via hypothesis testing. arXiv
preprint arXiv:1904.00045 (2019)
17. K. Simonyan, A. Vedaldi, A. Zisserman, Deep inside convolutional networks: visualising image
classification models and saliency maps, in Proceedings of the 2nd International Conference
on Learning Representations, ICLR—Workshop Track Proceedings (2013)
18. B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, A. Torralba, Learning deep features for discrim-
inative localization, in 2016 IEEE Conference on Computer Vision and Pattern Recognition
(CVPR) (IEEE, 2016), pp. 2921–2929
19. D. Batra, Grad-CAM: visual explanations from deep networks via gradient-based localization,
in Proceedings of the IEEE International Conference on Computer Vision (2017)
20. A. Chattopadhay, A. Sarkar, P. Howlader, V.N. Balasubramanian, Grad-CAM++: generalized
gradient-based visual explanations for deep convolutional networks, in 2018 IEEE Winter
Conference on Applications of Computer Vision (WACV) (IEEE, 2018), pp. 839–847
21. H. Li, Y. Tian, K. Mueller, X. Chen, Beyond saliency: understanding convolutional neural
networks from saliency prediction on layer-wise relevance propagation. Image Vis. Comput.
83–84, 70–86 (2019)
22. M. Sundararajan, A. Taly, Q. Yan, Axiomatic attribution for deep networks, in 34th
International Conference on Machine Learning, ICML 2017 (2017)
23. M. Ancona, E. Ceolini, C. Oztireli, M. Gross, Towards better understanding of gradient-based
attribution methods for deep neural networks, in 6th International Conference on Learning
Representations, ICLR 2018—Conference Track Proceedings (2018)
24. A. Shrikumar, P. Greenside, A. Kundaje, Learning important features through propagating acti-
vation differences, in Proceedings of the 34th International Conference on Machine Learning,
vol. 70(ICML17), p. 31453153 (2017). JMLR.org
25. J.T. Springenberg, A. Dosovitskiy, T. Brox, M. Riedmiller, Striving for simplicity: the all
convolutional net, in 3rd International Conference on Learning Representations, ICLR 2015—
Workshop Track Proceedings (2015)
26. P.-J. Kindermans, K.T. Schutt, M. Alber, K.-R. Muller, D. Erhan, B. Kim, S. Dahne,
Learning how to explain neural networks: patternnet and patternattribution, in 6th Interna-
tional Conference on Learning Representations, ICLR 2018—Conference Track Proceedings
(2018)
2 Explainable Artificial Intelligence (XAI): Understanding … 33
27. R. Caruana, Y. Lou, J. Gehrke, P. Koch, M. Sturm, N. Elhadad, Intelligible models for
healthcare: Predicting pneumonia risk and hospital 30-day readmission, in Proceedings of
the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining,
pp. 1721–1730 (2015)
28. A. Holzinger, C. Biemann, C.S. Pattichis, D.B. Kell, What do we need to build explainable AI
systems for the medical domain? (2017). [Online]. Available: https://arxiv.org/abs/1712.09923
29. J. Lightbourne, Damned lies & criminal sentencing using evidence based tools. 15 Duke Law
Technol. Rev. Tech. Rep. 327–343 (2017). https://scholarship.law.duke.edu/dltr/vol15/iss1/16.
Accessed 6 June 2018
30. S. Tan, R. Caruana, G. Hooker, Y. Lou, Detecting bias in black-box models using transparent
model distillation (2018). [Online]. Available: https://arxiv.org/abs/1710.06169
31. C. Howell, A framework for addressing fairness in consequential machine learning, in
Proceedings of the FAT Conference Tutorials, pp. 1–2 (2018)
32. R. Berk, J. Bleich, Statistical procedures for forecasting criminal behavior: a comparative
assessment. Criminol. Public Policy 12(3), 513–544 (2013)
33. Equifax, Equifax launches NeuroDecision technology (2018). [Online]. Available: https://inv
estor.equifax.com/news-and-events/news/2018/03-26-2018-143044126. Accessed 6 June 2018
34. D. Gunning, Explainable artificial intelligence (XAI). Defense Advanced Research Projects
Agency (DARPA). Accessed 6 June 2018. [Online]
35. W. Knight, The U.S. military wants its autonomous machines to explain themselves, MIT
Technology Review (2017). [Online]. Available: https://www.technologyreview.com/s/603
795/theus-military-wants-its-autonomous-machines-to-explain-themselves. Accessed 6 June
2018
36. Future of Privacy Forum, Unfairness by algorithm: distilling the harms of automated decision-
making (2017). [Online]. Available: https://fpf.org/wp-content/uploads/2017/12/FPF-Automa
tedDecision-Making-Harms-and-Mitigation-Charts.pdf. Accessed 6 June 2018
37. G. Montavon, S. Lapuschkin, A. Binder, W. Samek, K.R. Muller, Explaining nonlinear
classification decisions with deep Taylor decomposition. Pattern Recogn. 65, 211–222 (2017)
38. M. Sundararajan, A. Taly, Q. Yan, Axiomatic attribution for deep networks, in International
Conference on Machine Learning (ICML), pp. 3319–3328 (2017)
Chapter 3
Explainable Artificial Intelligence (XAI):
Conception, Visualization
and Assessment Approaches Towards
Amenable XAI
Abstract For last decade, due to the accessibility of huge databases and recent
advancements in deep learning methodology, machine learning systems have arrived
at or transcended tremendous performance in a spacious variety of tasks. One can
see this speedy development in speech analysis, image recognition, sentiment anal-
ysis, strategic game planning and many more, for e.g. in medical field, it’s used
for diagnosing different diseases, like breast cancer etc., based on their symptoms.
But many state-of- the-art models is facing lack of transparency and interpretability
which is a major hindrance in many applications, e.g. finance and healthcare where
visualization, interpretation and explanation for model’s decision is an obligation for
trust. This is an implicit problem of the current techniques carried by sub-symbolism
(e.g. Deep Neural Networks) that were not shown in the last hype of AI (specifically,
rule based models and expert systems). Models underlying this problem come within
the so-called Explainable AI (XAI) field, which is extensively acknowledged as a
racial feature for the practical deployment of AI models. As a result, explainable
artificial intelligence (XAI) has turned into scientific interest in last recent years. So,
this chapter epitomizes contemporary developments in Explainable AI that describes
explainability in Machine Learning, constituting a fiction definition of explainable
Machine Learning that envelopes such prior conceptual propositions with a consider-
able focus on the audience for which the explainability is needed. Except of this defi-
nition, this chapter starts a confabulation on its various techniques that are essentials
for analysing interpretability and explainability of Artificial Intelligence, and also
gives a comparison between two medical experiments, that are based on predicting
heart disease using disparate Explainable Artificial Intelligence techniques, which
can give a lead for researchers as well as practitioners or newcomers in the field of
Artificial Intelligence for selecting suitable methods with Explainable AI to grasp
T. Nizam (B)
Jamia Hamdard, New Delhi, India
e-mail: [email protected]
S. Zafar
Jamia Hamdard, New Delhi, India
e-mail: [email protected]
the advances of AI in their action sectors, without any previous bias for its dearth of
interpretability.
1 Introduction
In the last decemvir, there were many tasks that were initially thought to be computa-
tionally impassable, but since 2012, artificial intelligence (AI) and machine learning
(ML) systems have achieved tremendous human performance in many fields.
Because of considerable hardware improvements combined with new optimiza-
tion algorithms these leadings in the AI field were achieved. At the same time, there is
too much growth in high quality open-source libraries also which allows developers
as well as researchers to quickly code and test models which also result in great work
in AI field [3]. With the upgrading in image classification, speech recognition, clas-
sical (board) games, object detection and many more have edged to their procreation
to the real-world applications outside research labs, mostly in the area of supervised
learning.
As the time passed, one can see advances in application demanding areas, such as
bioinformatics, government, medicine, churn prediction and content recommenda-
tion but in these applications also crucial trust-related problems are raised. In exten-
sions to additions in aforementioned areas, future applications will also combine
interpretable science, reliable ML and cognitive assistance [4]. Due to these changes,
many ethical challenges will occur that society will have to accommodate fast in
order to drive the developments to the positive directions and outcomes. More unfair
wealth distribution might be done when automation will substantially change the job
market. Generation of fake content and content recommendation teamed with other
technologies will significantly impact the social dynamics [5]. Those algorithms
will prescribe many things and that will alter human lives in ways, so humans will
desire to trust them in order to obtain those prescriptions. Instead of this, to sustain
trust [6], i.e. reliability, explanatory justifiability, unbiasedness (fairness), usability,
privacy etc, systems must amuse many criteria (assurances). Since humans are social
creatures addicted to life in human communities such affirmations are assumed due
to bias towards human decision making. Due to novel status of Artificial Intelligence
in our lives as well as being of the human formulation, it generates more disbeliefs.
At starting, AI interpretation was easy and the previous years have endorsed
the boost of opaque decision systems such as Deep Neural Networks (DNNs). The
experimental success of Deep Learning models such as DNNs branches from a
combination of adequate learning algorithms and their vast parametric space. DNNs
are considered as complex black-box models due to latter space which is consist
of hundreds of layers and millions of parameters. Transparency is reverse of black-
box-ness which means understanding or searching the technique by which model
3 Explainable Artificial Intelligence (XAI): Conception … 37
works. As black-box Machine Learning (ML) models are widely being used to give
essentials predictions in crucial contexts, the appeal for transparency is surging from
the different stakeholders in AI [5, 6]. If we use or create any decision which are not
legitimate, justifiable or that simply do not grant gathering accurate explanations of
their behaviour, then it might be dangerous for us [7]. All explanations that support
the output of a model are important, e.g., experts need many more information from
the model than a simple binary prediction for encouraging their diagnosis in precision
medicine. In general, people are restrained to follow techniques that are not precisely
tractable, trustworthy and interpretable, which results demanding for ethical AI [7,
8]. The systems will be progressively opaque if one only thinks about solely on its
performance. This is correct in the sense that there is a trade-off between model’s
transparency and its performance. However, if one works on the understanding of
a system then it can lead to the correction of its dearth. While developing a ML
model, its implementation can be improved by considering its interpretability as an
additional design driver. This interpretability basically helps in impartial decision-
making and can act as a guarantee that only useful variables produce the output
[9].
So, for removing the limitations of effectiveness of the recent generation of AI
systems, explainable AI (XAI) gives a suite of ML techniques that:
• Generates more explainable models while keeping a high level of learning
achievement.
• Empowers humans to appropriately trust, understand and properly balance the
imminent generation of artificially intelligent partners [10].
Previous surveys and overviews of interpretability in machine learning are given
in [2, 3, 14, 15, 17]. In this chapter we survey the overtures in the explainability and
interpretability of machine learning models using the supervised learning paradigm.
This chapter is composed as follows: in Sect. 2 we deal with the definitions and
preliminaries. In Sect. 3 we classify the techniques for interpretability and explain-
ability. Section 4 offers discussion and comparison of two research done using AI
techniques which also discuss the methodology that can be used in XAI for accurate
result with proper explanation. And at last, in Sect. 5, we conclude the chapter which
also discuss current state of research ideas.
Before processing with our literature survey, it is good to first build a common point
of understanding on which the term explainability stands for in the context of AI
and, more specifically, ML. This is important aim of this section, i.e. to pause at the
different definitions that have been used in regards to this concept.
38 T. Nizam and S. Zafar
Terminology description:
In some literature interpretability and explainability are interchangeable misused
which impedes the formation of common grounds. There are important differences
between these concepts. Expressing any procedure or action taken by a model with
the purpose of detailing or clarifying its internal functions is considered as explain-
ability [10, 11]. Explainability is an active characteristic of a model. By contrast,
interpretability denotes to a passive characteristic of a model implying to the level
at which a given model makes sense for a human watcher. This characteristic is also
considered as transparency.
To compile the most commonly used terminologies, in this section we simplify
the differences and sameness among terms often used in the XAI communities and
ethical AI.
• Intelligibility (or Understandability): It represents the characteristic of a model
to make a human grasp its functions—how the model works—without any
commitment for explaining its algorithmic means using which the model operates
data internally or one can say without discussing its internal structure [12].
• Comprehensibility: The capability of a learning algorithm to express any model’s
learned knowledge in a human knowable fashion is known as comprehensibility.
It is specially used for ML models [13].
• Interpretability: The aligning of abstract concept into a domain that humans
can make sense of is called interpretability. In other words. It is expressed as the
capability to describe or to give the meaning in knowable terms to a human.
• Explainability: An interface between a decision maker and humans is known as
explainability. It is considered as a notion for explanation. In other words, it works
as definite proxy of the decision maker as well as conceivable to humans [13, 14].
• Transparency: If a model is understandable by itself, then it is considered to be
transparent.
In all the above definitions, understandability appears as the most important
concept in XAI. Both interpretability and transparency are firmly tied to this concept,
while transparency attributes the characteristics of a model to be understandable for
people by its own whereas understandability measures the level to which a human
can understand a decision taken by a model. The potential of the audience to grasp
the knowledge contained in the model is considered as Comprehensibility.
None of the aforesaid definitions is restrictive or specific enough to facilitate
formalization. They implicitly build upon user’s preferences, expertise and other
circumstantial variables.
3 Explainable Artificial Intelligence (XAI): Conception … 39
The best description of a normal model is the model itself; it perfectly depicts itself
and is simple to understand [21]. This method is narrowed to the model families
with lower flexibility or complexity, such as decision trees rules and linear models.
On the other hand, there are some model families such as support vector machines
(SVM), artificial neural networks (ANN), random forests and boosted tree, which
are considered opaque and their complexity block users from imaging the logic
behind predictions. The latter are most often taken as black-boxes and they are
handled with in a post-hoc manner. This arrangement between performance and
transparency is imaginary depicted in Fig. 1. The large part of task is done for
classification works. Different conditions can be implied on models in order to grow
their interpretability. Model sparsity, monotonicity and model’s size are some of
the properties that can be used for increasing interpretability. Even the selection of
model family (representation) can be treated a constraint on models that influences
the interpretability. Transparent models are both explainable and interpretable.
40 T. Nizam and S. Zafar
Fig. 1
Performance-transparency
trade-off [4]
Predictive performance increases the use of complex, opaque models. This is done
because of increased availability of data and hardware improvements. However,
explainability and interpretability are topics that have to be properly addressed.
Post-hoc explainability aims models that are not easily interpretable by design
by resorting to distinct means to improve their interpretability, such as visual expla-
nations, text explanations, explanations by example, feature relevance explanations
techniques etc. Each of these methods envelops one of the simplest ways humans
describe processes and systems by themselves. Furthermore, actual techniques, or
actual bunch of methods are defined to ease the future work of any researcher that
aims to search for a specific method that suits its knowledge. Not only this, the classi-
fication also involves the type of data in which the methods have been applied. There
is one point to be noted that many methods might be justifiable for many different
types of data, although the grouping only deals with the type used by the authors
that proposed such method [18–20]. Overall, post-hoc explainability techniques are
categorized first by the aim of the author (explanation technique e.g. Explanation by
simplification), then, by the technique used (actual technique e.g. sensitivity analysis)
and finally by the category of data in which it was practiced (e.g. images).
I. Text explanations accord with the problem of carrying explainability for a model
by means of learning to build text explanations that help explaining the results
from the model [26]. Text explanations also involve every technique generating
symbols that denote the working of the model.
II. Visual explanation techniques for post-hoc explainability intents at visualizing
the model’s action. Many of the visualization techniques existing in the literature
come along with dimensionality reduction methods that permit for a human
interpretable normal visualization.
III. Local explanations handle explainability by separating the solution space and
providing explanations to less complex solution subspaces that are admissible
for the whole model. These explanations can be composed by means of methods
with the differentiating attributes that these only describe part of the whole
system’s working.
IV. Explanations by example deals with the extraction of data examples that
express the result produced by a certain model, enabling to get an improved
understanding of the model itself. Similarly, to how people react when trying
to describe a given process, explanations by example are actually centred
in obtaining representative examples that catch the inner correlations and
relationships found by the model being analysed.
V. Feature relevance explanation methods for post-hoc explainability defines the
inner working of a model by computing an important score for its managed
variables. These scores compute the sensitivity (affection) a feature has upon
the output of the model. A comparison of the scores among distinct variables
42 T. Nizam and S. Zafar
discloses the effect provided by the model to every such variables when gener-
ating its output. Feature relevance methods can be considered to be an indirect
technique to explain a model [3].
4 Empirical Analysis
The growth in healthcare infrastructure of a nation plays a vital role in finding the
health-related condition of life of its public. The growing trends in artificial intel-
ligence has aided the service of computational resources to boost the healthcare
infrastructure globally [1–4]. One such situation that has captured a lot of attention
from research scientists is heart attack [5–8], controlling to the fact that 33% of
deaths annually is because of it. However, the decision-making process in this field
still relies upon the guesswork of the cardiovascular expert or upon the expertise.
This makes it deeply ambiguous to satisfy the quality of the decisions [9]. Therefore,
there is an urgency for a system to assist the process of cardiovascular risk judgement.
The growth in the number of artificial intelligence (AI) fanatics have helped in the
process by providing remarkable funding for AI-based research in medical decision
making. Nevertheless, the most important concern in the technique has endured the
explainability of the results produced by AI-based systems. Specially, in the case of
medical diagnosis, it is too much important for the decisions to have a conceivable
justification. There are many artificial intelligence methods that can help healthcare
decision support systems such as neural networks, genetic fuzzy systems and genetic
programming etc.
There are three important aspects of all AI systems need to be taken into consid-
eration: Reliability which refers to the accuracy of the results, Explainability which
considers for giving relevant explanations of the results and Understandability that
aims on the usability. With this perspective, in last decades, computer technolo-
gies with machine learning techniques have made medical aid software as a support
system for prior diagnosis of heart disease. Identification of any heart related diseases
at first stage can avoid the death risk. Different Machine Learning methods are used
in medical data to figure out the pattern of data and deciding prediction from them.
Healthcare data are normally enormous in volumes and complicated in structure.
Machine Learning algorithms are able to handle the massive data and mine them to
find the meaningful information. Machine Learning algorithms read from past data
and make predictions on real time data. This kind of Machine Learning framework
for coronary illness expectation can boost cardiologists in deciding quicker actions
so more patients can get medicines within a smaller timeframe, thus saving broad
number of lives.
3 Explainable Artificial Intelligence (XAI): Conception … 43
Machine Learning is a branch of Artificial research [2] and has become a very trendy
facet of data science. The Machine Learning algorithms are designed to execute a
huge number of tasks such as classification, prediction, decision making etc. To learn
the Machine Learning algorithms, training data is needed. After the learning stage,
a model is generated which is taken as an output of Machine Learning algorithm.
After this, that model is tested and validated on a group of unseen real time test
dataset. The final accuracy of the model is then compared with the real value, which
advocates the overall correctness of predicted result.
There are many popular methods that are used with Machine Learning Techniques.
Some of them which are used in Chaps. 1 and 2 are as follows:
With the help of Machine Learning Algorithm, Decision Tree Decision Tree algo-
rithm is used to create the Classification models. Decision Tree classification algo-
rithm based on the tree-like structure. This is a type of supervised learning, where
it is already known what to achieve. Both the numerical and categorical data can be
applied on Decision tree algorithm. Decision tree has a root node, its branches and
leaf nodes. The evaluation of Data is done on the basis of accessing path from the
44 T. Nizam and S. Zafar
root to a leaf node. In [chapter index] they used a dataset—CHDD, which contains
total of 283 tuples that were assessed down the decision tree. For the heart disease
prediction, they possibly came to a positive or negative assessment. After that, these
were compared to the real parameters to check for the false negatives/false positives
which displays the accuracy, sensitivity and specificity of the model.
Artificial Neural Networks are nearly clumsy electronic models based on the neural
structure of the brain.
The brain actually learns from experience. It is a proof given by nature that some
problems that are above the limit of current computers are actually handled by small
energy efficient packages. By using this brain modelling one uses a less technical
method to design machine solutions. These biologically encouraged techniques of
computing are considered to be the next vital advancement in the computer field.
Computers have problems recognizing even easy patterns. But now, enhancement in
biological research provides an initial understanding of the natural thinking mecha-
nism. This procedure of saving information as patterns, making use of those patterns,
and then solving problems encloses a new field in computing, which is called—Arti-
ficial Neural Network (ANN). ANN are those types of computers whose architecture
is created after the brain. They typically made of hundreds of normal processing units
which are wired together in a typical communication network. Each node or unit is a
simplified model of real neuron that sends off a new fires or signals, if it receives an
adequately strong Input signal from another node/nodes to which it is attached. ANN
3 Explainable Artificial Intelligence (XAI): Conception … 45
The concept fuzzy denotes to things that are not understandable or are clear. In the
real world there are many situations where we face a situation when we can’t evaluate
whether the state is true or false, their fuzzy logic gives very important flexibility for
reasoning. In this way, we can think about the uncertainties and inaccuracies of any
situation.
In the Boolean system truth value, 1 denotes the absolute truth value and 0 shows
the absolute false value. But in the fuzzy logic, there is no concept for the absolute
truth and absolute false value. But in fuzzy system, it has an intermediate value that
represents which thing is partially true and partially false (Fig. 3).
A genetic algorithm (GA) is a type of heuristic search algorithm which is used to solve
optimization and search problems. This algorithm is also a subset of evolutionary
algorithms, which are used in computation. Genetic algorithms apply the idea of
genetics and natural selection for giving solutions to problems. These algorithms
are taken as having better intelligence than random search algorithms because they
consider historical data to perform search to the best performing region within the
solution space. GAs is also built on the characteristics of chromosomes and their
genetic structure. There is fitness function which helps in giving the features of all
individuals within the population. The higher the function, the better the solution.
techniques for training FIS membership function parameters. In Chap. 2, the system
is trained for 1000 epochs. This training process stops whenever the defined epoch
number is reached. In a novel technique and in next step, GA is used for optimizing the
membership function parameters. As a result of bringing together these three algo-
rithms (the neural network, the fuzzy logic, and the genetic algorithm), for predicting
heart attack a new classification algorithm is generated. GA is set to minimize the
cost function (CF) which is defined as root mean squared error (RMSE) between the
expected (desired) and the predicted heart attack level.
⎡
l n
l1 Σ
CF = RMSE = √ (D Oi − P Oi )2 a n (1)
n i=1
Specificity = NN NN + NY (2)
Sensitivity = YY YY + YN (3)
Precision = YY YY + NY (4)
Accuracy = YY + NN YY + YN + NN + NY (5)
48 T. Nizam and S. Zafar
In Chap. 1, researches have used an online dataset, i.e. UCI Repository, which
have The Cleveland Heart Disease Dataset. This dataset contains 1025 instances in
total, which have been used in this research work. Each data is considered based on
14 attributes. This dataset has missing values too, which is manipulated through data
pre-processing using some statistical techniques. The Class Value 1, is represented
as “tested positive for the disease”, and Class value 0 represents “tested negative for
the disease”. There is a division in this dataset into certain percentage, such as 20%
data has been taken as testing dataset and rest 80% data is taken as training dataset.
Different attributes of the original dataset have some missing values also which can
give an imprecise result and may degrade the accuracy of model also. To overcome
this problem, by using a method, “mean of column”, missing values are replaced.
This method basically replaces 0 with either mean values of neighbourhood values
or taking the average of neighbourhood values. Then accordingly it converts the
0 value with the newly calculated value. After that the values of the dataset were
changed from Numeric to Nominal so that it can be dataset compatible with the ML
Techniques used.
The model is made in Weka which is a Data Mining Tool. The Waikato Environ-
ment for Knowledge Analysis (WEKA) is an open-source machine learning software
which developed by Waikato University, New Zealand. The software easily executes
different standard data mining tasks such as clustering, data pre-processing, regres-
sion, classification feature selection and visualization. It provides an easy environ-
ment to load data in the form of URLs, files or databases. It examines and conceives
the true positive, confusion matrix, recall, precision and false negative etc. in a
convenient way.
Four accuracy measures have been taken for comparison of the four models, they
are as follows:
(1) Precision or Positive Predictive Value: It is the average probability of relevant
retrieval.
(3) Accuracy: The accuracy of a classifier is given as the percentage of total correct
predictions divided by the total number of instances.
⎡ ⏋
Accuracy = Number of True Positives + True Negatives
/[Total Instances] (8)
(4) ROC (Receiver Operating Characteristic) Area: ROC depicts the performance
trade-off between the true positive rate (TPR) and false positive rate (FPR) of a
3 Explainable Artificial Intelligence (XAI): Conception … 49
classification model.
Classifier is generally taken as “GOOD” when ROC value is less than 0.80, “FAIR”
when ROC value is 0.77. The ROC value which is very close to 1 is considered as
best model with high accuracy.
At the end of experiment, researchers found that Random Forest and SVM
performed very well in comparison to Decision tree and Gaussian Naive Bayes.
Model developed with SVM produces 98% accuracy which is approximately 13%
greater than Decision tree and 8% greater than the Naïve Bayes. Similarly, model
build with Random Forest generates the best prediction result with 99% accuracy,
which is itself more accurate than our second best SVM model for heart disease
prediction. Unfortunately, decision tree is not suitable for this dataset.
5 Conclusion
By presenting this research, we have tried to analyse the different machine learning
techniques and to predict if someone in particular, given different individual
indications and attributes, will get coronary illness or not.
In Chap. 1, researchers used Cleveland dataset for heart diseases. This dataset
contains 1025 instances and used percent split to divide the data into two categories
which are testing and training datasets. They have also considered 14 attributes and
implemented four different algorithms, i.e. SVM, Random-forest, Decision Trees and
Naïve Bayes, to check the accuracy. By the end of this research, they have discovered
that maximum accuracy level is achieved by using Random Forest in dataset whereas
Decision Tree is playing out the least with an accuracy level.
In Chap. 2, a novel Adaptive Neural Fuzzy Inference System (ANFIS)-Genetic
Algorithm (GA) algorithm for predicting heart attack is proposed. By using the
explainablity of fuzzy rules, the optimization power of GA and training capability of
neural network, the algorithm was executed on the Cleveland dataset. After that, the
trained Fuzzy Inference System (FIS) obtained from ANFIS was given into GA to
optimize the membership function parameters. As a result, the Root mean Squared
Error (RMSE) between the predicted ones and the expected results obtained from
the algorithm reduced from 0.82754 to 0.75372. The efficiency of the proposed algo-
rithm was analysed by evaluation functions such as specificity, sensitivity, accuracy,
precision and RMSE.
50 T. Nizam and S. Zafar
The above results showed that the performance of the ANFIS-GA algorithm
in predicting heart attack was much satisfactory because this algorithm provides
explainable result. Providing explainable results are very useful for clinicians and
patients. This algorithm was designed in such a way that it gives matrices corre-
sponding to the predicted patients in different classes of having heart attack. Detecting
the attributes with highest significance and representing the satisfactory performance
of the algorithm on the new model may be taken as a good criterion for giving
explainable predictions based on the relevance of the features.
References
1. V. Sharma, S. Yadav, M. Gupta, Heart disease prediction using machine learning techniques, in
2020 2nd International Conference on Advances in Computing, Communication Control and
Networking (ICACCCN), pp. 177–181 (2020). https://doi.org/10.1109/ICACCCN51052.2020.
9362842
2. M. Aghamohammadi, M. Madan, J. Hong, I. Watson, Predicting heart attack through explain-
able artificial intelligence, pp. 633–645 (2019). https://doi.org/10.1007/978-3-030-22741-
8_45
3. A. Barredo Arrieta, N. Diaz Rodriguez, J. Del Ser, A. Bennetot, S. Tabik, A. Barbado González,
S. Garcia, S. Gil-Lopez, D. Molina, V. Benjamins, R. Chatila, F. Herrera, Explainable artificial
intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI.
Inf. Fusion 58 (2019). https://doi.org/10.1016/j.inffus.2019.12.012
4. F. Došilović, M. Brcic, N. Hlupic, Explainable artificial intelligence: a survey (2018). https://
doi.org/10.23919/MIPRO.2018.8400040
5. S. Pouriyeh, S. Vahid, G. Sannino, G. De Pietro, H. Arabnia, J. Gutierrez, A comprehensive
investigation and comparison of machine learning techniques in the domain of heart disease, in
2017 IEEE Symposium on Computers and Communications (ISCC) (Heraklion, 2017), pp. 204–
207. https://doi.org/10.1109/ISCC.2017.8024530
6. C. Raju, E. Philipsy, S. Chacko, L. Padma Suresh, S. Deepa Rajan, A survey on predicting heart
disease using data mining techniques, in 2018 Conference on Emerging Devices and Smart
Systems (ICEDSS) (Tiruchengode, 2018), pp. 253–255. https://doi.org/10.1109/ICEDSS.2018.
8544333
7. Y.-X. Wang, Q.H. Sun, T.-Y. Chien, P.-C. Huang, Using data mining and machine learning
techniques for system design space exploration and automatized optimization, in Proceedings of
the 2017 IEEE International Conference on Applied System Innovation, vol. 15, pp. 1079–1082
(2017)
8. S. Chattopadhyay, A neuro-fuzzy approach for the diagnosis of depression. Appl. Comput.
Inform. 13, 10–18 (2017)
9. A.M. Sagir, S. Sathasivam, A novel adaptive neuro fuzzy inference system based classification
model for heart disease prediction. Pertanika J. Sci. Technol. 25(1) (2017)
10. D. Richard, D. Mala, GI-ANFIS approach for envisage heart attack disease using data mining
techniques. Int. J. Innov. Res. Adv. Eng. (IJIRAE) 5 (2018)
11. D. Bonanno, K. Nock, L. Smith, P. Elmore, F. Petry, An approach to explainable deep learning
using fuzzy inference. Next-Generation Analyst V. (2017)
12. B.W. Israelsen, ‘I can assure you ... that it’s going to be all right’—a definition, case for, and
survey of algorithmic assurances in human-autonomy trust relationships. ArXiv:170800495
Cs Stat, August 2017
13. T.T. Nguyen, P.-M. Hui, F.M. Harper, L. Terveen, J.A. Konstan, Exploring the filter bubble:
the effect of using recommender systems on content diversity, in Proceedings of the 23rd
International Conference on World Wide Web (New York, NY, USA, 2014), pp. 677–686
3 Explainable Artificial Intelligence (XAI): Conception … 51
14. K. Crawford, Opinion I Artificial Intelligence’s White Guy Problem,” The New York Times,
25-Jun-2016.
15. B. Goodman, S. Flaxman, European Union regulations on algorithmic decision-making and a
‘right to explanation’. ArXiv 1606.08813 Cs Stat, Jun 2016
16. S. Khedkar, V. Subramanian, G. Shinde, P. Gandhi, Explainable AI in healthcare. SSRN
Electron. J. (2019)
17. E. Tjoa, C. Guan, A survey on explainable artificial intelligence (XAI): toward medical XAI.
IEEE Trans Neural Networks Learn Syst 32(11), 4793–4813 (2021). https://doi.org/10.1109/
TNNLS.2020.3027314
18. M.T. Ribeiro, C. Guestrin, Why should I trust you? Explaining the predictions of any classifier,
pp. 1135–1144 (2016)
19. D. Wang, Q. Yang, A. Abdul, B.Y. Lim, Designing theory-driven user-centric explainable AI,
in Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems—CHI
’19, pp. 1–15 (2019)
20. G. Marcus, Deep learning: a critical appraisal. ArXiv: 1801.00631 Cs Stat, Jan 2018
21. S. Lundberg, S.-I. Lee, A unified approach to interpreting model predictions. ArXiv:1705.07874
Cs Stat, May 2017
22. A. Cano, A. Zafra, S. Ventura, An interpretable classification rule mining algorithm. Inf. Sci.
240, 1–20 (2013)
23. J. Wang, R. Fujimaki, Y. Motohashi, Trading interpretability for accuracy: oblique treed sparse
additive models, in Proceedings of the 21th ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining (New York, NY, USA, 2015), pp. 1245–1254
24. T. Van Gestel, B. Baesens, P. Van Dijcke, J. Suykens, J. Garcia, T. Alderweireld, Linear and
non-linear credit scoring by combining logistic regression and support vector machines. J.
Credit Risk I (2005)
25. R. Piltaver, M. Lustrek, M. Gams, Multi-objective learning of accurate and comprehensible
classifiers—a case study (2014)
26. A. Bennetot, J.L. Laurent, R. Chatila, N. Diaz-Rodriguez, Towards explainable neural symbolic
visual reasoning, in NeSY Workshop IJCAI 2019 (Macau, China, 2019)
27. https://en.wikipedia.org/wiki/Bayes27_theorem
28. https://en.wikipedia.org/wiki/Naive_Bayes_classifier
29. https://towardsdatascience.com/understanding-random-forest-58381e0602d2
30. Images of neural network—Bing images
31. Fuzzy Logic | Introduction—GeeksforGeeks
Chapter 4
Explainable AI (XAI): A Survey
of Current and Future Opportunities
1 Introduction
The first thing which comes to the mind of most people when they hear the word
intelligence is usually robots. The reason behind this is that high-quality movies
4 Explainable AI (XAI): A Survey of Current and Future … 55
often feature machines that behave like humans and that cause destruction in the
world. But nothing could be further from the truth. Artificial intelligence wishes that
human intelligence could be described in a way that makes it become easier for a
computer to solve complex problems and perform heavy calculations or tasks. Artifi-
cial intelligence aims to imitate processes in humans. When it comes to defining clear
processes, such as learning, thinking, and perception, researchers and developers in
the field make rapid progress unexpectedly. Some people think that designers will
soon be able to create better programs than people can read or understand right now.
Others, however, continue to hold onto this view because all psychological processes
involve quantitative judgment influenced by human knowledge. The terms used to
describe artificial intelligence in the past become obsolete as technology advances.
For example, machines that perform simple math or visualization of text are no longer
considered a combination of artificial intelligence because we now consider these
skills as part of any computer [3].
With the advancements of AI in various fields, the major problem that comes out
is the lack of interpretability and trust. These factors are the major drawbacks that
make its applicability difficult in healthcare and finance where the need for a rational
model to make decisions is important for trust. AI has achieved increasing momentum
in its application in many areas to deal with increased complexity, scalability and
automation, which are also penetrating digital networks today. The rapid increase in
complexity and sophistication of AI-powered systems has evolved to such an extent
that people do not understand the complex mechanisms by which AI systems work
or how they make certain decisions—a particular problem when AI-based systems
calculate outputs that are unexpected or seemingly unpredictable. This is especially
true for opaque decision-making systems, such as those using deep neural networks
(DNNs), which are considered complex black-box models as shown in Fig. 1.
Artificial intelligence has received criticism from both the scientific community and
the general public since its inception. Another theory that emerges is that machines
will evolve to the point where people will not be able to keep up with them and will
move on its own, re-establishing itself in terms of power. Another is that technology
has the power to be armed and can attack people’s privacy. Some controversies have
focused on artificial intelligence and whether robots and other intelligent machines
should be granted the same human rights. Self-driving cars have caused controversy
because their cars are usually built considering the small number of accidents and
injuries. These vehicles will determine which route would cause the least damage if
given the opportunity to choose between colliding with one person at a time.
There are concerns that people may be forcibly fired from their jobs as more and
more businesses try to create certain jobs by using smart devices. Self-driving cars
may eliminate the need for taxis and car-sharing services, and manufacturers may
quickly switch workers into machines, further avoiding the need for human skills.
This concept will affect the life of the people who drive cars. The advancements in
the field of AI-based machines are done to reduce human efforts but at the same time,
it tends to reduce the job roles for humans. In the upcoming years, it is the tendency
that many decisions will be made by intelligent machines, and the human jobs role
in taking decisions would decrease.
2 Applications of AI
fields without looking at its computer-related features. Some of these are listed below
[4], Fig. 2.
• Speech recognition: Computer speech recognition, automatic speech recognition
(ASR), or text-to-speech recognition skills are the skills that use a natural language
processing system (NLP) to convert a person’s speech into a text format. To
make voice search, many devices integrate such algorithms into their systems.
For example, Siri. Artificial Intelligence has played a very important role when it
comes to disability not only physically or partially or for those who are sometimes
unconscious or even learning. To get to the world where you are and where you are
known, the blind can just talk and Artificial Intelligence will do the necessities.
The illiterate get the edge that they do not need to learn to read or write as they
can speak their questions or instruct the software to perform certain tasks such as
calling someone and Artificial Intelligence can answer questions and make calls.
• Customer service: In today’s world, human agents are being replaced by tangible
agents when it comes to customer care. They respond well to frequently asked
questions (FAQs) for common questions, offer personal advice, suggest user sizes,
best-selling products, and change the way of thinking about customer engagement
across all websites and forums. Example Messages for bots, Slack, and Facebook
Messenger. While visiting e-commerce sites, we are confronted with a host of
visually impaired agents who ask, “How can I help you?”. These message bots
narrow the gap between Queries and Solutions and strengthen the sales process.
They save our time as we do not need to scroll down to read the description or
get out of the app to ask our questions via email.
58 M. Kumari et al.
• Computer view: This Artificial Intelligence feature helps computers and systems
to extract important information from digital images, videos, and other visual
inputs, and based on this audio and visual input, it can take action. It differs from
image recognition functions in its ability to provide recommendations. Computer
vision is developed using convolutional neural networks and it is used for tagging
images on social media, radiology image in health care, and self-driven vehicles
within the vehicles industry. It has made it easy to contact our friends on social
media and forward a photo or post to them.
• Automotive industry: The automotive industry has benefited greatly from this
new use of Artificial Intelligence. It has made it easier for beginners to drive cars
with this AI feature.
• Recommendation engines: Artificial Intelligence algorithms also assist compa-
nies in finding data trends using previous data management which can help in
the development of highly productive cross-marketing action plans. This can also
help in making additional relevant suggestions to customers while undergoing
the payment process for online shopping. It empowered businesses and consumer
goods as AI-assisted them in selecting what to recommend to the consumer so that
opportunities to purchase goods and services could be increased. For example, if
we search for a particular type of shirt on any e-commerce website, then there is a
high chance that we will encounter the same type in our standard online drives that
force us to think about it, often, which affects our way of thinking and organizing.
• Automatic stock trading: Automated stock trading is designed to improve the
portfolio of stocks. Artificial Intelligence shapes its future as these commercially-
driven AI platforms are trading billions of dollars a day without human interven-
tion. Robo Advisors analyze a lot of data, facts, problems, threats, etc. with high
accuracy and reduce risk. Only Artificial intelligence has made the industry faster
with rocket fuel. Many companies have introduced programs based on the end of
Artificial Intelligence after considering their growing demand and high success.
• Language Learning: Artificial intelligence has become an integral part of manu-
facturing, language learning, e-commerce, fintech, education, skills development,
medicine, electricity, and the automotive industry in our daily lives. Intelligence
startup in learning a foreign language has opened the way for personal education.
Integration of Artificial Intelligence into learning.
• Medicine: Artificial intelligence can be used in almost every part of medicine. It
helps in collecting patients’ data with great accuracy and minimal possibilities of
error. When physicians take the readings themselves, being a human tendency to
ignore some lesser important facts and figures, they make mistakes. Even getting
the medicine prescriptions written using Artificial Intelligence is highly appreci-
ated as many patients face adverse consequences including cases of death due to
doctors’ poor handwriting making it difficult for the pharmacists to figure out the
correct medicines and doses. Researchers in the field of AI has devised various
kind of tools to solve various problems using probability and other mathematical
operations. Most of the operations are done by various logic developed. The oper-
ations or the algorithms to develop artificial intelligence help in the innovation of
many new things. The researchers or the developers develop after examining a lot
4 Explainable AI (XAI): A Survey of Current and Future … 59
In recent decades large amounts of datasets, high computational powers, and machine
learning systems have developed for high-performance tasks. Artificial intelligence is
an approach toward the development of technology-based on computational learning
and intelligence. AI has the potential of analysing and solving the complex prob-
lems of the medical industry. These days AI is playing an important in the medical
field as the medical industry faces challenges in analyzing and application of a lot
of knowledge to solve various complex problems. The advancement of AI in the
field of medicine is to help clinicians to formulate the diagnosis, make complex
decisions, and in the prediction of the results. It is developed to assist healthcare
workers to perform their tasks easily and accurately. In the case of AI development
for the medical industry, the role of Explainable Artificial Intelligence becomes
more important to deal with complex predictions and therapeutic decisions. In the
medical field, the algorithms of the complex calculations need to be checked to make
trustworthy and for this, the complex calculations of the AI algorithms need to be
developed in such a way that they are easy to understand. It is very important to
deal with complex calculations. Artificial neural network (ANN) is one of the most
popular AI techniques used in the field of medicine. These computational tools are
the tools that are by biological systems. These systems consist of various computer
machines connected known as neurons that are capable of data representation and
processing simultaneously. Their various abilities to retain historical data or exam-
ples or analyzing capability and many others make them very attractive and important
analytical tools. But the development of AI tools needs precision and to make them
reliable the algorithms or the codes need to be easily explainable to others to make
them reliable [5, 6].
The use of AI is increasing day by day in the field of neuroscience. Different
Artificial intelligence models help in the interpretation of multimodal datasets that
60 M. Kumari et al.
have the potential to provide unbiased insights into brain function and detection of
disorders of the brain. The ability of AI to make precise classifications and predictions
is there but, in some cases, it is unable to explain how inputs and the results are
relatable to each other. One big challenge for humans is the inability to keep a check
and modulate neural activities in real-time. To deal with the complex problems in
the field of neuroscience, advanced AI solutions need to be developed so that the
resulting models are able to are capable of interpreting, modulation, and sensing from
a huge database. New approaches to deal with these complexities are developed using
machine learning. Explainable AI has developed a new way with techniques in which
AI and ML complex algorithms are combined to develop easily explainable solutions.
Data-driven and theory-driven models both have a contrast. The data-driven models
are more accurate in terms of prediction while the theory-driven models are more
explainable. In general, the models with high performance are less explainable, while
the explainable models are generally less accurate. These are the general issues that
the ML community face. These problems can be solved by using models that are
simple and easier to explain or by building appropriate models.
Explainable AI is the form of AI in which the results are developed in such a way that
they are easy to understand by people other than the developers of that algorithm. If
humans have to accept any algorithm for a particular task, they need to have trust in
it, and to build trust the algorithms developed should be developed in such a way that
they are easily understandable and transparent. To make an algorithm trustworthy
transparency becomes a very important aspect for which the algorithm needs to be
easy to interpret or explain. For the purpose of extending transparency and trustwor-
thiness, the system was developed which was known as Truth Maintenance System.
This system tracks the reasoning and helps in the justification of the conclusions.
Latest AI approaches like deep learning and machine learning are very complex and
non-transparent in nature. To cope with this opaque nature and make the algorithms
easy to interpret various types of explainable models are developed like LRP or
Layer wise relevance propagation. In addition to this, work has also been done to
make glass-box models that are much more transparent for the purpose of inspection.
XAI models are developed having a primary focus on making it more understandable
to the people who practice AI rather than making it easier for the end-users. XAI
acts as a core for human-centric machine learning developments. There are various
types of machine learning (ML) models like rule lists, decision trees, simple naïve
Bayes Classifiers, etc. With the increase in complexity the understandability and the
explainability of the algorithms decrease as shown in Fig. 3.
Explainable AI (XAI) or Interpretive AI or Explainable device gaining knowledge
of (XML) is an AI wherein the outcomes of an algorithm are easy to understand with
the aid of people as shown in Fig. 3. It contradicts the belief of a “black box” in the
machine getting to know, which even its designers can’t provide an explanation for
4 Explainable AI (XAI): A Survey of Current and Future … 61
why the AI got here to a positive end [2]. XAI guarantees to assist users’ paintings
more efficaciously by way of improving the intellectual models of users of effective
AI structures and doing away with their defective thoughts.
XAI may be an enforcement of a right of definition XAI is crucial even supposing
there aren’t any prison or regulatory requirements—as an instance, XAI can improve
the person enjoy of a service or product by assisting cease customers believe that AI
is appearing nicely. In this way, the cause of XAI is to give an explanation for what
has been completed, what has been carried out, what is going to occur subsequently
and to reveal the records on which the moves are primarily based. Those elements
make it feasible to (i) affirm existing knowledge (ii) question existing understanding
and (iii) generate new thoughts.
Algorithms that are used in AI can be divided into white machine getting-to-
know algorithms and black container (ML). White box fashions ML fashions offer
clean consequences for professionals within the subject. Black container models,
alternatively, are extraordinarily tough to outline and aren’t understood even by way
of specialists inside the subject. The procedures that extract model parameters from
the education information and generate the labels within the take a look at data
can be explained and stimulated. Translation method a possibility to recognize the
machine gaining knowledge of the model and lay the basis for choice making in a
way that humans can recognize. A definition of a idea that is considered crucial,
but coherent which means does no longer but exist. it is advised that the definition
in ML may be thought of as “a summation of the factors of an interpretative space
committed to a specific model for choice making”. while algorithms meet those
requirements, they offer the basis that are for validating and to track selections and
for that reason validating themselves, developing algorithms, and discovering new
information. Every now and then it is also feasible to attain the end result with high
accuracy with the white field ML interpretation set of rules itself. this is especially
vital in areas along with remedy, law, protection, and finance and in these cases it’s
important to understand choices and construct agree with algorithms [7, 8].
AI structures increase behaviours to fulfil mathematically defined goal structures
chosen by way of gadget designers because the command to “boom the accuracy
of trying out how nicely a film assessment is within the test database”. The AI can
research well-known policies which can be useful in a check set, including “updates
containing the word ‘horrible’ may be terrible”. but, “opinions with ‘Daniel Day-
Lewis’ are generally exact”; these guidelines may not be famous in the event that they
62 M. Kumari et al.
appear to be they cannot do matters normally outside of the train set, or if humans
view the regulation as “dishonest” or “wrong”. The rules in XAI may be researched
to get a concept of how the system has the potential to generate actual-international
information for the future without a check set. AI is undergoing a lot of developments
these days. Most of the time, every company has plans to install AI, use it actively,
or rename its old tech engines with AI-enabled technology. As more companies use
AI and has advanced statistics in their business process and automated decisions, it
demands a specific reason that how the decision-making models grow bigger and
bigger. XAI is based on advanced methodology and strong thinking that depends
on how the results are derived for the proposed algorithms so that AI technology
can be clinically validated. In addition, XAI systems are more focused to build AI
mechanism or algorithms that are easy to understand for AI users. Some researchers
have encouraged the use of natural-language mechanical learning models, rather than
the use of post-hoc definitions, in which a second model is developed to describe the
first. This is because post-hoc models increase complexity in the decision-making
process and in part because it is often unclear how a post-hoc definition can faithfully
replicate a completely different model calculation. The main aim of Explainable
Artificial intelligence is to make the AI models explainable.
• Interpretability
To enable interpretability transparency becomes a very important property that was
traditionally the first step to protecting the rights in institutions. But with the new
advancements in the field of AI, the models are becoming much more complex that
are hard to get meaning, and hence the interpretation may vary for the required mech-
anism. With the increase in the interpretability of ML models, it becomes easier to
predict the reason behind the decisions. Interpretability is the concept of the advance-
ments in AI models that makes it easier for humans to understand the reasoning behind
the models. The more the interpretability of the model, the easier is the ability to
understand the model, and the more it becomes trustworthy for its application.
• Explainability
It is the extent that involves the development of the models in that form that is
easy for the users or the researchers to understand. In the case of explainability,
fidelity is one of the most important terms. The models with low fidelity lack the
explaining ability of the ML models. Explainable artificial intelligence (sometimes
known as shorthand “XAI”) refers to the power of the algorithm or model owners
to understand how AI has achieved its findings by making AI technology as bright
as possible [9]. AI is capable of making decisions automatically. Like in case of
hiring, it is important for the decision-makers of the organization to understand how
the decisions are made by the AI models. Many organizations wish to use AI-based
mechanism but are uncomfortable allowing the model or AI to take decisions that have
a significant impact because they do not yet trust the model. The description helps in
this because it gives the details on how the models take decisions and hence helps to
figure out whether the decisions are favourable or not. Machine learning is growing
4 Explainable AI (XAI): A Survey of Current and Future … 63
The initial developments in the field of AI were easily interpretable. It was found as a
research topic at the workshop of Dartmouth College, USA. At that workshop, many
people predicted that machines could be developed that would have intelligence like
that of humans, and to make this vision a reality, they were funded with millions of
dollars. Later on, the AI researchers underestimated the complications and difficulty
in the process of the development of this project. In 1974, due to the criticisms
from James Lighthill, pressure from the British Government as well as Congress,
the U.S., the funding for this was stopped. Seven years later, after getting inspired
by the Japanese Government, Billions of dollars were provided to them. By the late
1980s, the funding was withdrawn again. The interest in the field of AI increased
and funding improved in the early decades of the twenty-first century. During this
time period, machine learning concepts and algorithms were developed and were
successfully applied in various industries. In the early stage of the development of
AI, its applications were restricted to only specific tasks due to many reasons such
as the low computational powers of the processors. AI requires a lot of information
in the form of databases and in 70’s it was very difficult to build a large database. In
the 1980s, an attempt was made to deal with the common-sense knowledge problem.
A massive database was created that contained mundane facts remembered by an
average person. Douglas Lenat, who led this project, stated that no other shortcut is
possible, the only way is to teach the concepts of the machine one by one. Now AI
has achieved many of its old goals and now is used in many industries. Some of the
success of AI has been able to due to an increase in computer power. Knowledge
engineering helps AI to deal with real-world problems more efficiently. With the
increase in the use of Artificial intelligence in various fields, the complexities of the
algorithms have also increased with the increase in the use of the concepts like deep
learning and machine learning. With advancements, the need for the explainability
of the complex algorithms was felt and many researchers highlighted the importance
of explainability for maintaining transparency and increasing the trust factor [10].
Many mathematical structures are used to discover the ML mechanism. Math-
ematical knowledge generally helps in understanding the resulting pattern. When
these patterns are easy to understand it becomes easier to improve those patterns
to deal with the deep complexities. Integrating is one of the most commonly used
methodologies.
64 M. Kumari et al.
6 Challenges
To deal with big and complex problems, we need high-dimensional and large datasets.
To monitor brain functioning there is a need for development in the acquisition
system. There is a need to achieve more accuracy in the solutions developed using
the latest Explainable models. Next, there is a need for development in different
models as most of the models operate with single-level and multi-level modelling
is a hard task if classical methods are used. To manage these problems accurately,
a need for XAI models is felt. For this, we need to use the tools that can help us to
understand the transformation of the inputs and the resulting output that happen in
the brain network. Explainability is gaining great interest in radiology. In the medical
field, interpretability includes many factors that are sometimes not considered in other
fields, that include the responsibility and risk factor. It is a difficult task for scientists
to accept the validity and to rely on the developed solutions, without knowing how
they are derived. Users need to know the probability when a result can be wrong
and without knowing the working model this can’t happen. These days many Deep
learning libraries have started including their XAI libraries. The assessment criteria
like usability and reliability help the Machine learning community to keep a check
on how the algorithms are used and how they can be improved. AI and Machine
learning concepts have a great potential to transform the way of dealing with complex
problems in the future. It has the ability to create a great impact and solve the
complexities present in various fields. In many of its applications outside medicine.
Explainable AI allows the users to understand and interpret how the results are
derived. The demand for XAI in medical field is increased due to the requirement
of fair and ethical results. Giving an explanation to the practitioners allow them to
make the life- saving decisions more easily. Clinical Decision Support Systems are
in great need of Explainable models as these systems help medical practitioners to
make complex decisions. The developers are more interested in local explanations.
The increase in the complexities and the human need to solve the insight problems
has given an opportunity for AI to emerge and solve complex problems. To tackle the
real-world problems, the algorithms developed by AI should be easily understandable
[11, 12].
A sound research approach can handle the small issue of detecting study bias.
However, the intrinsic black-box structure, while improving predictive performance,
might cause problems in high-stakes decision-making if adequate justifications for
the choices made aren’t provided. It’s crucial to understand when a model will work
and when it won’t, as well as the reasons behind any specific predictions it makes and
the degree of its limitations. A common concern among potential users of artificial
intelligence is that it is often unclear how technology can reach their conclusions. If
AI algorithms are locked in a “black box” that prevents people from knowing how
the outcomes were achieved, it might be difficult to trust the developed algorithms
because human experts sometimes are unable to explain the findings of AI. Being
able to define AI can help organizations establish greater trust with customers, clients,
and investment partners. One of the most important benefits of being able to define
4 Explainable AI (XAI): A Survey of Current and Future … 65
artificial intelligence is that it can help technology owners determine whether human
bias has affected the model. This can be especially important when AI is called to
make life or death decisions, such as a hospital setting where medical professionals
may be required to explain to patients and their families why certain decisions have
been made. In order to move beyond interpretability and investigate the domain
of XAI that especially deals with deep learning models, various advancements are
undergoing day by day. According to the essay the earliest and most popular XAI
technique was developed by Ribeiro et al. in their paper Explaining the Predictions of
Any Classifier, and it is known as Local Interpretable Model-agnostic Explanations
(LIME). The use of LIME on photos to identify a husky as a wolf sheds light on
the importance of explainability in preventing poor learning. The study shows how a
model came to understand that the presence of snow in the image was a key criterion
for identifying the animal as a wolf. These explanations make it simple for us to spot
incorrect learning, which helps us avoid bias in both data and models [13–15].
7 Applications of XAI
In the healthcare sector AI can play a crucial role but the risk of using an unreliable AI
system is much greater. The choices made by AI models that help doctors categorize
serious diseases based on structured criteria or unstructured data such as medical
imaging naturally have a broad impact. If a system makes a prediction and if the
doctor needs to spend time to check whether the AI system’s decision is accurate and
reliable then the model would be of no use. In that case, an AI system that predicts
with explains the reasons for its conclusion will be much more valuable. To save a life
in the healthcare sector the reason for taking a decision while undergoing treatment
is very important.
In the medical field, when a patient suffering from a disease undergoes diagnosis,
then, in that case, explainable AI can play a vital role to explain the process of
diagnosis. It can help the doctors to make a treatment plan, take the correct decisions
in complex situations, and also to explain their diagnosis process to the patients. This
can help in the development of trust between the doctors and patients. For example,
XAI can help in making decisions is the diagnosis of pneumonia. AI can also help
in the diagnosis of cancer in the healthcare sector where medical imaging is used.
Currently, many attempts are made to discover how explainable AI models can be
developed so that they could be applied to neuroscience. With explainability, these
models also need to have the capability to discover a data-driven manner [16].
Banking, financial services and insurance (BFSI)—BFSI has benefited greatly
from XAI’s potential to transform the sector. Credit risk assessment based on arti-
ficial intelligence is widely used in the banking and insurance industries. In indus-
trialized countries, multimodal premium estimation is becoming more popular with
pay-as-you-drive and pay-how-you-drive mechanisms in which machine learning
algorithms are used in their decision-making. Industry can implement AI solutions
more effectively if XAI systems are able to deliver superior results and clear expla-
nations while gaining adequate trust and meeting regulatory standards. With larger
financial stakes, actions such as loan denials, increases or decreases in health insur-
ance or auto insurance costs, or incorrect investment recommendations, XAI can
be seen as an invaluable tool in explaining these options to the parties who will be
affected.
• Automobiles
The future of the automotive industry is autonomous driving, which is evolving.
Self-driving or autonomous vehicles are exciting as long as they don’t make any
mistakes. In this risky use of AI, one mistake can lead to the loss of one or more
lives. Explainability is essential to know about the capabilities and limitations of AI
models prior to implementation. It is important to understand the limitations of driver
assistance systems (or autopilot) when consumers are using them so that the causes of
the limitations can be identified, clarified and addressed as soon as possible. Parking
assist and voice assistants are one of the attractive features that largely represent a
comparatively low-risk model choice. However, there are alternative situations such
as brake assist or self-driving vehicles.
• Judiciary
AI technologies are increasingly being used to make decisions in the legal systems of
Western countries. As with parole based on the likelihood of repeat offending, biases
in AI applications have far-reaching effects, and justice in them is fundamental as
they affect human rights and freedoms. XAI is undoubtedly a rapidly expanding field.
4 Explainable AI (XAI): A Survey of Current and Future … 67
The major difference between AI and XAI models are given in Table 1.
The main difference between the XAI and AI is that the XAI deals with the
problems with the practical aspects and it develops solutions that are easily under-
standable. The explainable methods used in Explainable AI are highly influenced by
the factors that a human needs for practical application. In recent studies or develop-
ments, it been has suggested that the pursuit of definition in AI strategies should be
taken as the secondary goal in pursuit of excellence in AI and that the promotion of
special XAI development may limit the overall performance of AI. XAI models are
based on advanced methodology and strong thinking from evidence-based medicine.
XAI systems are more focused on making AI systems understandable and more trust-
worthy or transparent to AI. Some researchers have encouraged the use of natural-
language mechanical learning models, rather than the use of post-hoc definitions, in
which a second model is developed to describe the first. This is because post-hoc
models increase complexity in the decision-making process and in part because it is
often unclear how a post-hoc definition can faithfully replicate a completely different
model calculation [19–22].
A typical AI application is comparable to explainable artificial intelligence, or
XAI, except that the processes and outcomes of a XAI algorithm may be described
so that people can understand them. Because of its complexity, artificial intelligence
makes judgments in real-time based on the knowledge it has learned from the data
it has been fed. We cannot completely optimise the AI application to be all that it
is capable of doing when we do not fully grasp how AI is making these judgments.
This is due to the fact that the bulk of AI using ML functions in a “black box,” or a
space that is unable to offer any obvious insights into how it generates judgments.
Transparency and explainability are not strictly required because many AI/ML appli-
cations are somewhat benign decision engines that are integrated with online retail
recommender systems. The stakes are substantially greater for other, riskier decision-
making processes including medical diagnosis in the healthcare sector, financial
investment decisions, and safety–critical systems in autonomous vehicles. As a result,
in order to be seen as trustworthy, dependable, and consistent, the AI employed in
those systems must be explicable, transparent, and intelligible. Applications for AI
and ML are widely used in business today to automate decision-making and to get
analytical insights [23–25].
An explainable AI model would allow a brand to enhance revenue by identifying
the real drivers of sales, whereas data models may be taught to anticipate sales based
on variable data. AI is a simulation of human intelligence processes by a computer
system. AI includes some special features such as natural speech recognition and
machine vision. AI is accelerating around the world in a very fast speed. AI requires
a set of specialized software and hardware for writing and training the machine to
get the desired work from the machine. Some of popular programing languages used
for AI are Python, R and Java. In short AI work by ingesting a large amount of data
(labelled training data) and analysing that data for correlation and patterns and using
that analysis for predicting the future states. AI focuses mainly on three skills i.e.
self-correction, reasoning and learning. artificial Intelligence is getting used across
many industries to offer the whole lot from personalization, automation, monetary
decisioning, suggestions, and healthcare. For AI to be depended on and ordinary,
people need to be able to recognize how AI works and why it comes to make the
decisions it makes. XAI represents the evolution of AI, and offers possibilities for
industries to create AI packages that are relied on, transparent, impartial, and justified
[26–29].
Sometimes while learning, AI models learn unwanted tricks that do a great job
while training by satisfying the pre-programmed goals. For example, in the year
2017, a program given an image recognition function learned to “cheat” by looking
at the copyright tag present with horse pictures, in case of learning to say that the
horse was actually photographed. In another 2017 edition, a surveyed AI given the
task of capturing objects in the physical world learned to cheat as the trick was to be
placed in between object and viewer in such a way that it seemed to be catching a
false object. Explainable AI is defined as AI programs that describe post-prediction
thinking. Explainable AI is a term that comes under artificial intelligence that is called
as “interpreting.” Interpreting allows us to understand what the model is learning,
what other information it provides, and the reasons for its judgments because of the
real-world story we are trying to fix. If model metrics are inadequate, interpretation
is required. Model interpretation enables us to predict how the model will perform in
4 Explainable AI (XAI): A Survey of Current and Future … 69
different test environments by comparing it with the training environment. With the
help of explainable AI, you can provide transparency about how AI systems make
decisions and make the developed models trustworthy. XAI has its use case in various
fields such as the military, manufacturing industry, and healthcare sector. Explainable
AI is a new domain in the field of research and there are many practical challenges to
explainable models that are developed till now. Another challenge to clarity may be
due to the accuracy of the model performance, as descriptive performance systems
tend to have lower performance compared to non-descriptive models or black-box
models. Explainable AI is an important part of the future of AI because descriptive
artificial intelligence models explain thinking through their decisions. This provides a
growing level of understanding between humans and machines, which can help build
trust in AI systems. Explainable AI improves the transparency, integrity, integrity, and
accountability of AI systems. Explainable AI Systems can be useful when trying to
understand the thinking behind a particular prediction or decision made by machine
learning models [30–32].
Without explaining the inner workings of an AI model and without knowing how
decisions are made, there is a risk factor to relying on the developed algorithms if
the user is unaware of what factors are taken into consideration and how the inputs
are processed. XAI gives the necessary clarity and helps to build transparency to
enable greater confidence in AI-based solutions. Thus, XAI is recognized as a key
feature for building artificial models and making them capable of making trustworthy
decision-making models and developing them in such a way that takes care of the
fundamentals of AI models as well as the fundamentals of the AI users. Trust is one
of the major factors for users to adopt AI-based models and solutions that incorporate
the decisions they make. However, there are some problems in the development of
explainable methods or models. It is an important factor that is achieving the easiness
of algorithm interpretability. Transparency influences the high-performance nature
of complex but opaque models [32, 33].
In this chapter, the current and future opportunities of XAI are surveyed. It is
concluded that XAI offers an opportunity where outcomes of AI algorithms can
be easily understood with the aid of people. With the development of complex algo-
rithms, the need for the development of explainable AI algorithms is also increasing
day by day. The need for explainability, to make a model trustworthy and transparent,
is increasing. But the questions that arise are whether the development of this tech-
nology will affect the job roles of the people? will the employment decrease? will
the problem of complexity of the algorithms be reduced, if yes then to what extent?
XAI can be used in various AI fields such as neuroscience, healthcare, BFSI,
automobiles, judiciary, etc. However, it has various challenges like the requirement
of high-dimensional and large datasets but the deep learning libraries can allow
XAI libraries to develop algorithms for real-world applications which can be easily
70 M. Kumari et al.
understood. Thus, in the future XAI will play an attractive role to build trustworthy
AI models.
References
1 Introduction
Edge computing has the potential to decentralize using cloud applications simulta-
neously increasing bandwidth and lowering latency. These benefits are realized by
transferring application-specific calculations across the cloud, data-generating equip-
ment, and elements of the wireless and landline network infrastructure network edges.
However, existing Machine learning and artificial intelligence (AI) (ML) method-
ologies presuppose that calculations [1] are carried out in a robust computing envi-
ronment, also including datacenters with plenty of processing and storage for data
capabilities]. The recent Dagstuhl Seminar—“213,421” gathered community input
from a varied variety of professionals to provide light on the rapidly expanding
subject of Edge AI, which combines edge computing and AI/ML. This article is the
product of their efforts, and it examines demands for implementing AI techniques as
in context of edge computing are both technological and societal.
The phrase ‘Edge,’ as modern development and research progresses, remains a
nebulous term. Across communities and researchers, a shared understanding of what
the edge is, where it lives, and who supplies it is lacking. Similarity (latency and
topology), improved capacity of a network (effectively feasible data transfer rate),
lower processing strength, smaller size, and greater device heterogeneity are some of
its characteristics when compared to the cloud. When compared to end devices (the
last step), it offers more computing and storage resources. It is a fictitious entity that
allows you to outsource computation and storage without having to go to the cloud.
Present-day AI and ML techniques, which necessitate sophisticated computa-
tional structures, are becoming a growing source of friction [2] an improved request
supplied the data centre with adequate processing and storage for data capabilities.
Sending the required raw data to cloud, on the other hand other hand, places a strain
on the network in terms of bandwidth and throughput. Meanwhile, most businesses
are hesitant to share (possibly sensitive) data [3] with commercial cloud providers.
Edge AI, a rapidly emerging domain, addresses this tension. On the other end, appli-
cation of edge AI can be used in industry 5.0 like smart home, smart agriculture,
smart vehicles, smart education, smart critical industries, smart e-commerce, smart
transport and smart healthcare etc.
Edge AI has gradually made its entry into regular service areas that is linked
automobiles, instantaneous gaming, healthcare, and intelligent manufacturing, as
shown in Fig. 1.
Edge environments, from an infrastructure standpoint, create a distinct layer for AI
and provide possibilities for current technology like federated learning or embedded
AI, which aim to reduce RAM use on specific devices, increase maintaining data
privacy on a local device, and reduce communication needs between distributed
entities. These properties form the basis the usage cases depicted in Fig. 1. Our
Contribution is mentioned below:
• Discussed the use case if Edge AI in current era.
• 5G technology in industry 4.0.
• Elaborated the usage of cloud computing in next generation.
5 Recent Challenges on Edge AI with Its Application … 75
The transition from networks of the fifth generation (5G) to the sixth-generation (6G)
period shaping networking’s future is established. The advancements on this road
include network connectivity, including its speed, coverage, and durability, as well
as computer quality and latency. To realize the full future networks’ possibilities
with such a complex structure, several technologies must coexist and collaborate
both in terms of hardware and software. The establishment factors are generated
for edge computing and communications, for example or the use of Technologies
with dynamic network orchestration that are self-learning examples of this. Security
assurance techniques, secret communication, computation and learning are all key
viewpoints in holistic system trustworthiness. We go over the obstacles and opportu-
nities [4, 5] in this part. The following literature covers the efficiency of edge AI. V,
Mazzia proposed the structure micro perspective with training dataset. There is also
some literature which focuses on the privacy [4, 6–8] and security [9–11] issues.
76 K. Joshi et al.
Using the widespread use of smart phones or other portable electronic gadgets, impor-
tant compute, interaction, and sensor abilities have becoming useful in addressing
networking difficulties. Acquisition of local data, as used in federated learning,
decreasing reducing communication costs through device-to-device caching and
working together in the execution [5] of computationally complex tasks are just
a few examples. In the future, mobile devices will be able to select whether, the how,
when, and how much of such a task to assign a cloud or edge server. Furthermore,
individuals can assume the role of a computer worker, organize resource pools, and
split jobs according to their preferences in order to maximize their effectiveness and
usefulness.
While people or machines operated by people may contribute considerably Edge
AI, such participation comes with a number of drawbacks. Multi-agent systems can
be used to represent such problems. As a result, numerous mathematical methods
such as game theory and control can be used to handle problems. Furthermore, when
there is uncertainty and a lack of knowledge, ML and AI play a significant role. The
special qualities of people should be considered in all of the preceding procedures.
Humans, in particular, rely concerning heuristics and unjustified influences to make
decisions, while taking considerations include societal peer pressure and standards.
It’s also critical to think about reciprocal trust and respect each entity’s welfare while
interacting with self-interested entities. Finally, employing using people as either a
data source such as through GPS or body sensors raises privacy worries about are
inextricably linked to ethical and legal issues.
The future demands of the driving force behind societies produces using examples we
shift from present networks and systems for 5G and beyond. As a result, developing
novel technology to address societal use-cases develop critical. In a limited sense, it
contradicts present-day usage cases like as well as dependability, which are driven
by technology advancements instead of derived from society itself. Current vertical
trends include resource-efficient manufacturing, the production and distribution of
green energy, organic farming, and retail logistics optimization, to name a few.
To meet their diverse demands, heterogeneous entities are anticipated to create and
transfer a vast infrastructure for edge computing. Despite sharing certain common
objectives, such actors frequently have competing interests; for example, more advan-
tage for one attribute may diminish the benefit for the other (here the utility can
correspond to major sustainability, improved environmental factors). The finding
of a Pareto-optimal and [12] the most reliable answer to this issue is extremely
difficult due to the fact that different utility measurements are frequently at odds.
In fact, the problem is exacerbated since there are multiple decision-makers rather
5 Recent Challenges on Edge AI with Its Application … 77
than a single centralized figure. This is due to asymmetry of knowledge and varied
types/preferences, as well as self-interest. Artificial intelligence (AI) can help solve
this challenge by allowing using distributed systems communicate, acquire knowl-
edge, make choices, effectively turning Smart systems are interconnected chunks of
edge intelligence.
The debate over the future of 5G has recently been dominated by hypothetical use
situations, as well as overarching objective of developing real-time integration of
communication, AI, and edge computing services that adapt to applications’ changing
78 K. Joshi et al.
needs. These predicted solutions, which range from ordinary living smart transporta-
tion and medical benefits, are important, but they must be evaluated in the situation
with technology and new economic models.
There are only a few business cases, but we were able to find three instances
where AI technology is being used for whole new tasks and models are yet defined
for them:
• Interpreting the results of joint sensing and communication capabilities in
future networks. Future higher frequency Communications provides some degree
of “radar-like” environment recognition. This is an example of significant change
Regarding the actions or operations of present networks, and the security issues
involved are substantial.
• Optimal link-level communication details discovery through ML. Theoret-
ically, which is doable. The practicality within the technologies still is debat-
able, albeit with the required education expenses are covered by the optimization
benefits accrued.
• Interoperability and collaborative use of data and a technologies. The majority
of modern systems operate within single organizations However, we pose the
following query: “What are our options promote more interoperable technolo-
gies.” The need to identify as well as discuss pertinent security, privacy, and
ethical issues, as well as the instruments to address them, is shared by diverse
stakeholders the min trust worthy manner.
The collaboration between future edge nodes and clouds shapes the future cloud
computing perspective. Latency, bandwidth, data localization, scaling, ease of use,
security, and fault tolerance are the key reasons for moving workloads from the
cloud to the edge. This collaboration encompasses not just technical obstacles, but
also business and stakeholder issues. The possibility conflict between the edges of the
cloud must work closely together to deliver the greatest assistance available clients, is
an important non-technical obstacle. We’ll go over the technology problems in detail
from a cloud perspective in the following sections. These issues, in our perspective,
are: Resource management, energy constraints and efficiency, security, trust, and
privacy and sporadic connectivity are the other four topics.
Even though the number edge servers could supply a significant number of resources,
required proximity [7] number edge servers severely limit the quantity of resources
available when in contrast to a cloud setting. Owing to hardware restrictions and task
execution lag, the quantity of different ML jobs that may be done at the same time is
5 Recent Challenges on Edge AI with Its Application … 79
limited. While execution and instantaneous loading of the associated models could
improve the amount of ML tasks that the edge can perform, this might lengthen the
latency of jobs it needs to be loaded. Thus, for tasks that require low latency, proactive
resource reservation may be required. Because resource reservations can substantially
reduce how many ML tasks there are supported, the need for resource reservations
should also be decided based on the possibility of failing and the repercussions of
such error. The administration of these resources in crucial situations can be difficult,
especially when resources are scarce. The issues are discussed in Fig. 2.
Importantly, edge AI operations are frequently embedded in broader settings,
such as a control workflow. As a result, various services provided via cutting-edge
resources must engage with one another. Not accounting for the notion data items
could need must be passed forward to services result in these services not being
ready to receive either data or not having the computational capacity to handle data
items. As a result, there may be a stream of data items. This is particularly true when
stream processing is used, such as when fresh data items must be processed and then
transmitted to various destination services based on the categorization. As a result,
it could be fascinating to watch and analyze how jobs are completed based [8] based
on overall priority level in light of the various loads at the edge.
Edge nodes are projected to be less cost and energy efficient compared to cloud data
centers in general. Economies of scale, therefore may act in opposition to edge data
80 K. Joshi et al.
centers by enabling greater cooling, for example. Renewable energy can be deployed
at both cloud data centers and edge nodes, however Unknown where these units are
produce greater energy efficiency. Large data centers, a single hand, offer greater
chances Using renewable energy, enhancing their persistence and environmental
but compatibility they also take in substantially more energy than edge nodes. The
opposite is true for edge nodes may be able to they generate their own power more
easily compared to cloud data centers because of modest geographic spread and
scale, despite their restricted Amount of renewable energy units available.
Furthermore, data must be transported over The Internet uses energy; hence it
increases the likelihood that edge nodes may outperform cloud data centers that
use less energy: Edge nodes can process data, for example, by detecting anoma-
lous instances and transmitting only pertinent information, by dynamically adapting
monitoring rates. Therefore, the amount of processing that can be done at the edge
determines how much energy can be saved. Another option is to shift part of the
training process to the edge; using techniques like federated education (see Sect. 4.3).
Only learnt parameters (such as gradients) must be communicated in this situation,
compared to raw data, which many various magnitudes bigger.
Another possible the benefit of edge nodes over general cloud server racks is the
more specifically constructed hardware, which is increasing common during market
(e.g., hardware accelerators) and makes better use of limited processing and graphical
capacities.
Edge AI, as seen in Fig. 3, provides prospects for increased trust, security, and
privacy while also posing new obstacles. Because some threats may be discovered
early and countermeasures taken, security may be enhanced. However, the number
of potential targets and attack channels is substantially greater, resulting in a bigger
assault potential.
Because of its dispersed structure and potential liability issues, Edge AI may
have a problem with trust in the system. Open implementation and specifications,
on the other hand, may help to boost edge server confidence. It may also be required
to distinguish between devices within the user’s control and resources offered by
providers (e.g., Azure, AWS, GCP). The implementation of a reputation system that
analyses and controls trust, as seen in other fields, is one viable option for building
trust. However, it will take time for new players to establish a reputation.
Another feature the data of Edge AI is what needed to complete the task of ML
is retained on servers at the edges. As a result, it’s widely assumed that edge servers
offer more privacy and trust than the cloud. As opposed to popular belief, Edge
cannot ensure privacy; rather, it creates additional obstacles for ensuring privacy.
That is, because edge node trust is more difficult to maintain and administer than
cloud provider trust, data privacy and security are harmed, even while the data’s
location prevents some assaults.
Generally, we discuss utilizing edge servers with weak internet connections, i.e.,
edge servers in areas in which internet connection are bad. Edge intelligence has a
lot of potential in those places as well, and it can be a driving force for digitalization
in non-urban areas. Water and air pollution monitoring, as well as natural hazard
prediction (wildfires, floods, volcanic eruptions, and so on) are examples of such
systems. For example, in water pollution monitoring, edge resources could be used
to identify chemicals in the water.
Scenarios involving environmental monitoring where, streaming data must be
analyzed in almost real-time an absence of a solid network access in those places
could be the key driving reason. Monitoring systems are unable to continuously
transmit raw data processing on the cloud due to intermittent connectivity and energy
constraints, as opposed to less data-heavy control signals such as the result of machine
learning still possible to communicate algorithms despite disruptions and restricted
bandwidth. Furthermore, the participants who rely on the results Examples of ML
algorithms frequently located near the sources of data. By avoiding transferring
cloud storage for data, an entire processing stage may avoid, such as automatically
initiating countermeasures. In the prior a good example of monitoring water contam-
ination edge hence, services instruct opening or closing pumps or valves based on
the circumstance.
4 Evolving AI and ML
This paragraph discusses edge nodes from an AI perspective, and difficulties that it
raises. Edge servers, cloud servers, and mobile devices, Fig. 4 depicts the various
82 K. Joshi et al.
AI complexity levels and data availability: While mobile devices have the most data
at their disposal, their processing capabilities are often constrained. As a result,
preprocessed information can sent to community edge servers for tasks that mobile
devices can’t handle. The edge server offers extra capabilities and can do more
complex tasks than the mobile devices; however data access [9] is limited owing to
preprocessing. The inability to finished will be forwarded to a cloud server, capable
of execute even most sophisticated models, but only a fraction of the data is received.
Tasks that are unable to be performed will be forwarded a connection to the cloud
server, which execute most sophisticated models, but only is given a portion of the
edge server’s information.
The following are three challenges that have been highlighted and should be
addressed by the community: (1) the accessibility of accelerators for applications
using AI; (2) identifying a trade-off between precision and resource use; and (3)
Using Federated edge server learning is the three issues.
used less elastically, which might lead to let hardware idle. It is anticipated that the
huge hardware accelerators would be put in the cloud due to business considerations.
To cooperate with them restricted resources, one current trend is to break down
smaller the model and AI algorithm parts that could be processed these tiny hardware
accelerators are used. The quality of service and the user experience of apps running
too close to the edge improved by combining the interference of these small hardware
accelerators. In general, huge models that require a lot of resources will need to be
moved to both the ground and the cloud.
Offloading and multi-tier architectures are also possible with machine learning.
Unlike classical workloads, which have intrinsically contradictory criteria for
expanding to higher tiers and nodes while holding the workload’s state consistent
across nodes, ML, particularly neural network training, has a distinct state distri-
bution synchronization models. Additionally, in contrast to conventional distributed
systems, overall consistency criteria easily possible reduced, resulting in a trade-off
of network bandwidth and replica count for relatively minor accuracy losses. As a
result, an essential question is what precision and delay constraints using the appli-
cation, as well as how to check [10] the model’s quality during its whole life-cycle
on the edge.
Similar to before stated, edge servers face greater resource limitations than clouds
of servers. Due to their complexity, certain models may only be able to operate in
the cloud.
The common goal today’s AI research frequently aims to attain the greatest
feasible precision or to maximize the incentive mechanism. However, in the case
of Edge AI, the quality of the outputs is [11] not the sole criterion for success; other
factors such as energy usage and memory must also be addressed. As a result, models
must be flexible in order to change for the current state of memory and resources
at hand. There must be a way to adjust the dimensions of individual ML models
for this purpose. Miniaturization techniques such as quantization and approximation
can be used to achieve this. In general, these strategies provide a [13] comparison
of performance measures and model correctness. TinyML and Tensor-Lite are two
examples of Edge AI minimization. The evaluation of adjusting the ML model in
light of present accessible resources is an important research topic. Furthermore,
we address the prospect of offloading specific jobs to the cloud: if the precision of
the scaled-down model implemented near the edges is inadequate, offloading cloud-
based task may be necessary. The clouds can then offer very accurate results and
deliver them the edge once more [14].
84 K. Joshi et al.
the combining their impacts technological enabling factors and non-tech require-
ments such user purchasing power, socioeconomic transformation behavior, as well
as commercial interests are reflected in this Edge AI roadmap.
Despite its potential and potential, when deployed on a large scale, Edge AI may
encounter significant difficulties, such as energy optimization, trust worthiness,
ethical, security, and privacy concerns.
The power consumption of Edge AI needs to be improved as a crucial compo-
nent of sustainability. On Edge AI embedding infrastructures (such as roadside units
and micro base stations) to continuously enable sophisticated autonomous driving
and services for Extended Reality (XR) in the upcoming years. Through the pipe
line Edge AI has the option to exchange accuracy with efficiency of medical image
data, [20] data capture, transport, calculation, and storage reduced less time and less
power is used. For instance, it is possible to transport and selectively process noisy
signals from many sensors in order to conserve energy. Instead of being perfect and
totally accurate, a group of applications would indeed be pleased with “acceptable”
accuracy correct results. The optimization design can further increase energy effi-
ciency by adding this different dimension of accuracy. Concerning trust worthiness,
Edge AI gains from being so close to the end devices. But because of the scattered
deployment with profound insights in to personal circumstances, perceived safety,
86 K. Joshi et al.
and trust worthiness for Edge AI services have been causing stakeholders to express
their worries (e.g., end users, public sectors, ISP).
To succeed trust worthy Edge AI requires essential building components, such as
systems for verification and validation that guarantee transparency and explain ability,
particularly while training and deployment of Edge AI in distributed, unregulated
settings. The trust worthiness of Edge AI is a first step toward establishing appropriate
regulatory and governance frame work, which Edge AI’s potential can be built.
When we talk about the security Regarding Edge AI’s privacy, there are two things
to think about, i.e.,
(1) Use of edge AI for privacy and security;
(2) Due to the deployment of Edge AI, there are data security and privacy risks, as
shown in Fig. 3.
Edge AI has the potential to be a key tool for ensuring network security and privacy.
Edge AI uses edge computing to process AI rather than using the cloud processes
the local info that is closer to that same user. That eliminates the need of transferring
eliminates data transfer between both the user and the cloud. The possibility potential
is fixed during the data transfer phase of the backhaul. Furthermore, local storage
and processing of data [21] can enhance privacy. Processing’s primary objective is
to move the interface of the AI workflow to the gadget while keeping data limited
to the device. The potential for a decentralized AI algorithm, which would remove
one single point of failure as in cloud AI system, is another security advantage of
edge AI. Edge AI might in some ways be is a son due to increased intelligence, to
reduce the impact of the attack on the immediate environment but also mitigate it
there already. There are growing concerns about privacy and safety a rising Edge AI
is used in this. Decentralization of Edge AI increases that number of access points
that AI system hackers and exposes previously undiscovered attack vectors. Edge AI
devices could lack the very same level of protection as the cloud, making them easy
entry points for an assault the AI system. Additionally, edge devices are reachable
physically, embedded Impact of AI on edge devices on capturing edge devices ever
than a non-AI edge device. For example, seizing control of Edge Ai technologies
might threaten or seize control of nearly all the local work services provided by a
certain place.
future cloud, and AI/ML are needed to capture the most recent technology advance-
ments for the envisioned roadmap. In terms of future advancement, AI applications
may be more elaborated for better environment and in this section, various issues has
been discussed on industry 4.0 and industry 5.0 with the usage of 5G technology in
future so here we expect the perspectives the ideas presented in this paper may offer
the community an alternative perspective and, in the long term, help to generate the
global adoption of Edge AI in future global world.
References
1. Y.L. Lee, P.K. Tsung, M. Wu, Techology trend of edge AI. in 2018 International Symposium
on VLSI Design, Automation and Test (VLSI-DAT) (IEEE, 2018, April), pp. 1–2
2. S. Greengard, AI on edge. Commun. ACM 63(9), 18–20 (2020)
3. X. Wang, Y. Han, C. Wang, Q. Zhao, X. Chen, M. Chen, In-edge ai: Intelligentizing mobile edge
computing, caching and communication by federated learning. IEEE Network 33(5), 156–165
(2019)
4. E. Li, L. Zeng, Z. Zhou, X. Chen, Edge AI: on-demand accelerating deep neural network
inference via edge computing. IEEE Trans. Wireless Commun. 19(1), 447–457 (2019)
5. T. Rausch, W. Hummer, V. Muthusamy, A. Rashed, S. Dustdar, Towards a serverless platform
for edge {AI}. in 2nd USENIX Workshop on Hot Topics in Edge Computing (HotEdge 19,
2019)
6. A.Y. Ding, E. Peltonen, T. Meuser, A. Aral, C. Becker, S. Dustdar, L. Wolf, Roadmap for edge
AI: a Dagstuhl perspective. ACM SIGCOMM Comput. Commun. Rev. 52(1), 28–33 (2022)
7. S. Yang, Q. Liu, X. Wan, J. Jiang, L. Ai, Edge-rich MoS2 nanosheets anchored on layered Ti3C2
MXene for highly efficient and rapid catalytic reduction of 4-nitrophenol and methylene blue.
J. Alloy. Compd. 891, 161900 (2022)
8. G.K. Agarwal, M. Magnusson, A. Johanson, Edge AI driven technology advancements paving
way towards new capabilities. Int. J. Innov. Technol. Manag. 18(01), 2040005 (2021)
9. M.K. Mudunuru, X. Chen, S. Karra, G. Hammond, P. Jiang, K.C. Solander, T.D. Scheibe,
EdgeAI: How to use AI to collect reliable and relevant watershed data (No. AI4ESP-1095).
Artificial Intelligence for Earth System Predictability (AI4ESP) Collaboration (United States,
2021)
10. C. Mwase, Y. Jin, T. Westerlund, H. Tenhunen, Z. Zou, Communication-efficient distributed
AI strategies for the IoT edge. Future Gener. Comput. Syst. (2022)
11. O. Vermesan, J. Bacquet, (eds.) (2019). Next generation Internet of Things: Distributed
intelligence at the edge and human machine-to-machine cooperation (River Publishers)
12. V. Mazzia, A. Khaliq, F. Salvetti, M. Chiaberge, Real-time apple detection system using
embedded systems with hardware accelerators: an edge AI application. IEEE Access 8,
9102–9114 (2020)
13. H. Yu, D. Yu, C. Wang, Y. Hu, Y. Li, Edge intelligence-driven digital twin of CNC system:
Architecture and deployment. Rob. Comput. Integr. Manufac. 79, 102418 (2023)
14. E. De Giovanni, F. Forooghifar, G. Surrel, T. Teijeiro, M. Peon, A. Aminifar, D. Atienza Alonso,
Intelligent edge biomedical sensors in the internet of things (IoT) era. in Emerging Computing:
From Devices to Systems (Springer, Singapore, 2023), pp. 407–433
15. Y. Singh, Z.A. Sheikh, On efficient and secure multi-access edge computing for internet of
things. in Proceedings of Third International Conference on Computing, Communications,
and Cyber-Security (Springer, Singapore, 2023), pp. 299–310
16. T.T.H. Kim, L. Lu, Y. Chen, ReRAM-based processing-in-memory (PIM). in Processing-in-
Memory for AI (Springer, Cham, 2023), pp. 93–120
88 K. Joshi et al.
17. A.K. Sharma, P. Singh, P. Vats, D. Jain, Deep learning and machine intelligence for opera-
tional management of strategic planning. in Proceedings of Third International Conference on
Computing, Communications, and Cyber-Security (Springer, Singapore, 2023), pp. 475–485
18. A. Bourbah, B. Meliani, Z. Madini, Y. Zouine, The next-generation 6g: trends, applications,
technologies, challenges, and use cases. in Proceedings of Seventh International Congress on
Information and Communication Technology (Springer, Singapore, 2023), pp. 761–770
19. L. Zanella, Y. Wang, N. Dall’Asen, A. Ancilotto, F. Paissan, E. Ricci, M. Pistore, Responsible AI
at the edge: towards privacy-preserving smart cities. in Ital-IA 2022 Convegno del Laboratorio
nazionale CINI-AIIS (2022, March)
20. M. Kirola, M. Memoria, A. Dumka, K. Joshi, A comprehensive review study on: optimized
data mining, machine learning and deep learning techniques for breast cancer prediction in big
data context. Biomed. Pharmacol. J. 15(1), 13–25 (2022)
21. M. Diwakar, A. Tripathi, K. Joshi, M. Memoria, P. Singh, Latest trends on heart disease
prediction using machine learning and image fusion. Mater. Today: Proc. 37, 3213–3218 (2021)
Chapter 6
Explainable Artificial Intelligence
in Health Care: How XAI Improves User
Trust in High-Risk Decisions
S. Praveen (B)
Integral University Lucknow, Lucknow, India
e-mail: [email protected]
K. Joshi
Uttaranchal Institute of Technology, Uttaranchal University, Dehradun, India
1 Introduction
Machine
Learning
XAI
Deep Learning
2 Explainable AI Principles
they left was because they got bad service? In its explanation, the system will say
this. Gives people the power to act. XAI lets people who are affected by certain
decisions, like granting a mortgage, challenge and possibly change the outcome
[6].
4 Explainable AI Example
XAI can give a detailed explanation at the model level of why a certain decision was
made. This explanation is in the form of a set of clear rules. Every person over the
age of 40 who has less than $433 in savings per month and who requests for credit
with a repayment term exceeding 38 years will be rejected a loan, according to the
simplified application below. The same is true for candidates under the age of 35
who have a monthly savings of less than $657 (Fig. 3).
6 XAI in Finances
Finance is another area that has a lot of rules and needs to explain its decisions.
AI-powered solutions must be able to be checked, or they will have a hard time
getting into the market. AI can, among other things, help give out credit scores,
evaluate insurance claims, and optimize investment portfolios. But if the algorithms
give biased results, it could hurt their reputation and even lead to lawsuits. Not too
long ago, Apple got a lot of attention for its Apple Card, which was unfair to women
because it lowered their credit limits. Steve Wozniak, who helped start Apple, said
that this was true. He remembered that he and his wife don’t have separate bank
accounts or assets, but when they both applied for an Apple Card, his limit was
10 times greater than hers. New York’s Department of Financial Services examined
the firm after this incident. With AI that can explain its results, people can avoid
embarrassing situations like this. For example, XAI can help with the process of
giving out loans. If a client’s loan application is rejected, the system will be able to
explain why in detail and defend that decision. This lets people improve their credit
scores and apply again later [7].
Self-driving cars use a lot of data, which needs to be analyzed and made sense of
by AI. In the event of any problems, the system’s judgments must be accessible to
drivers, technicians, authorities, and insurance companies. Also, it’s very important
to know how vehicles will act in an emergency. Paul Appleby, a former CEO of
the data management software firm Kinetica, expressed his worry as follows: “If a
self-driving vehicle finds itself in a situation where an accident is likely, what steps
it should take? Put pedestrians at serious risk to safeguard the safety of the driver?
Avoid people on the street while putting the safety of the passengers at risk?” These
are hard questions to answer, and different people would have different ideas about
how to handle them. But it is important to give the algorithm rules it can follow
in these situations. This will help people decide if they are comfortable riding in a
car that can make certain decisions on its own. Also, the explanation given after an
incident will help developers make the algorithm better in the future [8].
IBM Watson IoT, demonstrates how explainable AI is used in their AI-based mainte-
nance product. In order to fix a piece of equipment, the system gives human workers
a number of possibilities. The confidence interval for each choice is shown as a
percentage. Therefore, while making a decision, the user may still draw on their
“tribal wisdom” and experience. As well, each recommendation can show the output
of the knowledge graph along with the information used during the training phase
[9].
and Google Scholar. In his article, they show how XAI can be used to diagnose
breast cancer through an experiment and show how XAI can be used in medical
applications. In the last part of the paper, a summary is given of the XAI methods
used in medical XAI applications, the problems that researchers have faced, and
where they think research should go in the future. The results of the survey show
that medical XAI is a promising area of research, and the goal of this study is to
give medical experts and AI scientists a place to start when making medical XAI
applications [14].
The majority of artificial intelligence (AI)-based diagnostic tools for healthcare work
as “black boxes,” which means that the findings do not provide an explanation of
why the computer believes the patient has a certain illness or ailment. Although
AI technologies are very strong, since physicians and regulators cannot indepen-
dently check their outcomes, adoption of these algorithms in healthcare has been
gradual. However, a brand-new kind of algorithm known as “explainable AI” (XAI)
is simple for people to comprehend. Therefore, it is probable that providers will
really apply the linked diagnoses since all indications point to XAI being quickly
accepted throughout the health care industry [11]. The black box feature of AI is
acceptable—and maybe even desirable—for many industries outside of healthcare
because it enables businesses to protect their valuable trade secrets. For instance,
deep learning in AI recognizes speech patterns to enable a user’s preferred voice
6 Explainable Artificial Intelligence in Health Care … 97
assistant to start their favorite movie. Deep learning algorithms identify connections
and patterns without ever requiring the operators to be aware of the data elements
that are most crucial to the outcome. For many uses of AI, there is no risk in relying
on it to continue to provide accurate results since the outcomes verify the algorithms.
However, the black box nature of AI makes it difficult for physicians and regulators
to trust it, maybe for good reason, in industries like health care where errors can
have fatal consequences. Doctors are largely taught to recognize outliers, or unusual
situations that don’t respond to conventional therapies. We cannot be certain that an
AI system would recognize such outliers or otherwise accurately diagnose patients,
for example, if it has not been trained correctly with the necessary data and we are
unable to comprehend how it makes its decisions.
The FDA, which now verifies AI algorithms by looking at what kind of data is
fed into the algorithms to make their judgments on the data, finds the black box
component of AI to be troublesome for the same reasons. A doctor stands between
the solution and the patient’s ultimate diagnosis or course of action, which is another
reason why many AI-related developments bypass the FDA [7].
For instance, the FDA continues to mandate that doctors be able to independently
verify the rationale behind the software’s recommendations in its most recent draught
guidance, which was published on September 28. This is done to prevent the software
from drawing increased scrutiny as a medical “device.” Because physicians can verify
the results of the algorithms, software is only loosely controlled. Think of a medical
imaging where clinicians may double-check worrisome lumps that the algorithm has
flagged. The difficulty for doctors with algorithms like deep learning, however, is
that they lack knowledge of the context in which a diagnosis was made.
Because of this, the XAI algorithms being created for use in the health care
industry can justify their findings in a way that people can comprehend. Many of the
XAI algorithms created so far are rather basic, such decision trees, and have very
specific applications. But as they advance, they will probably be the algorithms that
dominate the healthcare industry. Companies in the healthcare industry would be
advised to invest in its growth [11].
This case study focused on how XAI assists a physician in saving the life of a
patient with West Nile virus. Case study demonstrates the Key characteristics of XAI
utilizing a key medical case of an early identification of a West Nile virus infection
in a person. Without AI and XAI, the patient’s life may have been jeopardized. There
are four characters in this story: the patient, the West Nile virus, the physician, and
the AI + XAI software. The doctor learns that AI and XAI just saved a life and
comes to trust AI via XAI at the conclusion of the story. On the long road ahead,
this symbolizes one of the first stages of collaboration between people and machines
[15].
98 S. Praveen and K. Joshi
There are still several current concerns and obstacles at the junction of machine
learning and explanation [3].
• Starting with computers as opposed to humans. Should XAI systems person-
alize explanations for specific users? Should they consider the lack of expertise
among users? How can explanations be used to facilitate interactive and human-
in-the-loop learning, including allowing users to engage with explanations to offer
feedback and drive learning?
• Precision versus interpretability an important stream of XAI research on explana-
tion investigates interpretability’s strategies and constraints. Interpretability must
evaluate accuracy and fidelity tradeoffs and achieve a balance between precision,
interpretability, and tractability.
• Simplifying explanations by using abstractions. Patterns at the highest level
provide the foundation for defining large plans in large phases. It has long been
difficult to automate the discovery of abstractions, and today’s cutting-edge XAI
research is focused on understanding how abstractions are discovered and shared
in learning and explanation.
• Comparing the explanation of skills to that of choices. Reflecting on novel circum-
stances is an indication of expertise among highly skilled specialists. It is vital
to assist end users comprehend the competences of AI systems in terms of what
competencies a certain AI system has, how the competencies should be tested,
and if an AI system has blind spots, i.e., whether there are classes of solutions it
cannot discover.
13 Conclusion
This Chapter mainly focuses on the XAI versus AI Introduction. How XAI is a more
effective technology in the medical industry. For the most part, medical diagnostics
that make use of AI are “black boxes,” meaning that the findings don’t provide any
context for why the computer came to this conclusion about the patient. Although AI
technologies are exceptionally strong, their acceptance in health care has been gradual
due to the inability of physicians and regulators to validate their outcomes. However,
humans can readily comprehend XAI. All indications indicate to the quick adoption
of XAI throughout the health care industry, making it probable that providers will
use the accompanying diagnoses.
From the standpoint of human-centered research, study on competences and
knowledge might move XAI beyond the function of describing a specific XAI system
and assisting its users in determining acceptable trust. In the future, XAIs could play
significant societal roles. These tasks may involve learning and explaining, working
with other agents to link information, establishing cross-disciplinary insights and
common ground, cooperating in educating people and other agents, and relying on
6 Explainable Artificial Intelligence in Health Care … 99
References
Abstract The growth of information technology (IT) has resulted in physical de-
vices being connected to the internet and having the ability to recognize other devices.
Artificial Intelligence algorithms are processed on edge or the devices of users. Edge
Computing based on the same premise, stores, processes, and manages data directly
at Internet of Things (IoT) endpoints. Edge artificial intelligence uses the device’s
hardware to process data and performs machine learning and deep learning proce-
dures. In the model, you can troubleshoot and improve model performance while
also assisting others in understanding the behavior of your models. To make the area
more real, explainable edge devices come in a wide range of costs and capabilities. A
decade ago, we couldn’t imagine that explainable edge artificial intelligence would
be at today’s level. Now it is a part of industries and even devices for customer service.
The best example of explainable edge AI is virtual assistants such as Alexa, google
assistant. They learn from the user’s world and phrases and can store them directly
on the device. These are just a few examples later, and we have possible applica-
tions in future works on the explainable edge artificial intelligence. Edge Computing
Platform facilitates the development and elastic operation of apps and services. Its
benefits the AI assisting in overcoming the technical obstacles that AI-enabled apps
experience. Combination of edge and AI is buzzwords within the industry to deliver
the performance and reduce the cost compared to state of arts XAI applications.
Moreover edge artificial intelligence are reducing the latency, improving user expe-
rience, and reducing the necessary bandwidth, consequently reducing the costs of
internet services. It surfs this movement since the need for data processing on the
device themselves also represents the increasing use of artificial intelligence. Arti-
ficial intelligence edge processing focused on model which trained them in central
data center using historical datasets. Compression techniques of data that enables
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 101
A. E. Hassanien et al. (eds.), Explainable Edge AI: A Futuristic Computing Perspective,
Studies in Computational Intelligence 1072,
https://doi.org/10.1007/978-3-031-18292-1_7
102 A. Kumar et al.
squeezing large artificial intelligence models into small hardware form factors could
push some training to the edge over time.
1 Introduction
Explainable artificial intelligence (XAI) is a set of processes and methods that allows
human users to comprehend and trust the results and output created by machine
learning algorithms. Explainable AI is used to describe an AI model, its expected
impact and potential biases. It helps characterize model accuracy, fairness, trans-
parency and outcomes in AI-powered decision making. Explainable AI is crucial for
an organization in building trust and confidence when putting AI models into produc-
tion. AI explainability also helps an organization adopt a responsible approach to AI
development.
As AI becomes more advanced, humans are challenged to comprehend and retrace
how the algorithm came to a result. The whole calculation process is turned into
what is commonly referred to as a “black box” that is impossible to interpret. These
black box models are created directly from the data. And, not even the engineers or
data scientists who create the algorithm can understand or explain what exactly is
happening inside them or how the AI algorithm arrived at a specific result. There are
many advantages to understanding how an AI-enabled system has led to a specific
output. Explainability can help developers ensure that the system is working as
expected, it might be necessary to meet regulatory standards, or it might be important
in allowing those affected by a decision to challenge or change that outcome.
The explainable artificial intelligence (XAI) is a strong tool for addressing ethical
and legal concerns regarding AI systems by answering important how? and Why?
Questions. As a result, AI researchers have recognized XAI as a necessary component
of reliable AI, and explainability has recently gotten a lot of attention. Despite the
increased interest in XAI research and the requirement for explainability across a
wide range of areas, XAI still has a number of drawbacks. This blog post provides
an overview of the current state of XAI, as well as its advantages and disadvantages.
Interpretability aids in ensuring objectivity in decision-making by detecting and
correcting bias in the training dataset. It provides resilience by indicating potential
adversarial disturbances that modify the prediction. It also assure that only relevant
7 Role of Explainable Edge AI … 103
variables are used to infer the result, i.e., that the model reasoning is based on genuine
causality.
Explanations can take many different forms depending on the goal and situation.
Human-language and heat-map explanations of model activities are shown in Fig. 1.
The machine learning model presented below is designed for doctors to utilize and
can diagnose brain tumor using MRI. Based on the MRI on the far left, the original
report gives a “ground-truth” report from a doctor. The Generated report includes an
explanation of the model’s diagnosis as well as a heat-map of the tumor locations
that influenced the conclusion. The Generated report offers doctors with an easy-to-
understand and validated explanation of the model’s diagnosis.
Explainable AI is about making sense of black box models and AI is about making
sense of transparent AI models. XAI implements efficient techniques and methods
to ensure that every decision made during the process can be traced and explained
whereas in AI it often arrives at a result using the algorithms but the arrangements
of the systems do not fully understand how the algorithms reached to the results.
AI has achieved growing momentum in many fields to deal with the complexity.
Humans do not grasp the complex mechanisms by which AI systems work or how
they make certain decisions because AI-powered systems have progressed to the
point where humans do not understand the complex processes by which AI systems
work or how they make certain decisions. There’s a risk that an AI model won’t be
considered trustworthy or authentic if the core functionality and judgments it makes
aren’t explained. The role of domain knowledge interacting with an XAI system
in a realistic complex decision-making scenario, XAI brings the needed clarity and
openness to AI-based solutions, allowing for increased trust [1]. As a result, XAI is
recognized as a critical feature for the practical deployment of AI models in systems
and, more critically, for ensuring those AI users’ fundamental rights to AI decision-
making are respected. Our ultimate objective is to provide a comprehensive taxonomy
to newcomers in the field of XAI that can be used as a reference tool to encourage
future research advances, as well as to encourage experts and professionals from
other disciplines to embrace the benefits of AI in their respective activity sectors,
without prejudice due to its lack of interpretability.
104 A. Kumar et al.
with specific addresses, there would be an opportunity to benefit from the knowledge
gained by analyzing the history of patterns in many fields, including business, soci-
etal, health, political, and many others. According to a CISCO report cited by the
researcher, the IoT sensor commercial market is estimated to reach $22.48 billion
by 2023, with a predicted 29.3 billion linked devices accessible [4].
The rest of the chapter is organized as follow: Sect. 9.2 represents the concepts of
explainable artificial intelligence with benefits and challenges. Section 9.3 illustrates
the example of explainable artificial intelligence in real world with the future of XAI.
Finally, Sect. 9.4 discussed the future of XAI summarizes the work with conclusion
follow with references.
2 Concepts of XAI
XAI will develop a set of algorithms that will enable human users to com-prehend,
trust, and manage the next generation of artificially intelligent part-ners. This defini-
tion combines two notions (understanding and trust) that must be addressed before
proceeding. It describes a model’s ability to explain its function—how it works—to
a person without having to explain its internal structure or the computational means
by which the model processes data in-side. Trust can be defined as the belief that a
model will behave as planned, a specific issue although it should undoubtedly be a
feature of any explainable model [5], and this does not imply that every trustworthy
model is explainable on its own, nor does it indicate that trustworthiness is a property
of each explainable model a feature that is easily quantifiable.
The major principles in XAI deals with the Intelligence to interpret predictions from
machine learning models, resiliency, reliability, bias, and account-ability are some of
the other characteristics. In most cases, these phrases are defined as part of a larger set
of principles or pillars. The four principles of explainable artificial intelligence are
highly affected by the relationship of AI system with human recipient information.
Before proceeding to the principles of the XAI, the key term i.e. the output of an
AI system. [6] For an example grammar checking system, the output is grammatical
error and correction recommendation. For disease detection system, the output is
detected disease and suggested precautions. The XAI objective is to provide the
information with trust, confidence, informative and transparency, Fig. 3 provides the
106 A. Kumar et al.
• Explanation
• Meaningful
• Explanation accuracy
• Knowledge limits
2.1.1 Explanation
This principle obligates AI system to provide proof, support and logic for each output.
It does not require that the evidence be correct, informative, or understandable in and
of itself; it only implies that a system can provide an explanation. Currently, a corpus
of work is being done to build and validate explainable AI approaches. It imposes
no demands on those explanations in terms of quality. The ideas of Relevance and
Explanation Accuracy establish a framework for assessing explanations.
2.1.2 Meaningful
examples of wide groups. The goal of this principle may vary, e.g. what is mean-
ingful to a doctor may be different than what is meaningful for a researcher. So,
the confluence of the AI system’s explanation and a person’s existing knowledge,
experiences, and beliefs influences meaningfulness. The challenge in modeling the
interface between AI and humans is due to all of the factors that determine meaning-
fulness. Both computational and human variables must be taken into account when
developing systems that produce relevant explanations.
The Explanation and Meaningful principles only need a system to create explanations
that are meaningful to a user group when used together. This principle exposes
accuracy on a system’s explanations. In the same way as the Meaningful principle
allows for distinct explanation accuracy metrics for different groups and individuals,
this principle allows for different explanation accuracy metrics for different groups
and individuals. Some clients will require concise explanations that focus on the
important point(s) but lack the intricacies required to fully explain the algorithm’s
output generation process. However sometimes nuances data may be meaningful to
experts. Overall, if a system can create multiple types of explanations, it may be
deemed more explainable. Because of the various layers of explanation, the metrics
used to assess an explanation’s accuracy may not be universal or absolute.
This principle asserts that systems detect situations in which they were not designed
or approved. To operate in which their responses are not trustworthy. This method
protects answers by identifying and disclosing knowledge boundaries, ensuring that
no judgment is made when it is not necessary. Two ways, through which system can
reach its knowledge limits, initially the question can be outside of the domain of the
system. For an example, system has to classify the wild animals but the user may
input the image of parrot. The system may produce an answer indicating that it was
unable to locate any wild animals in the input image and hence could not provide an
answer; this is a response as well as an explanation. Another ways depending on an
internal trust level, a knowledge limit can be reached, and the confidence of the most
likely solution may be too low, for an e.g. reptiles classification system, the input
image of a lizard is too blurry to identify. In this situation the system may recognize
the image of reptiles, but it is in low quality, explanation is found the image of reptiles
but the quality of an image is too low to identify it.
108 A. Kumar et al.
The description of XAI principles explained in Sect. 9.2.1, this section explains the
working environment of XAI. It examines each aspect and its impact on the final
result. Users can use this analysis to create new situations and see how different input
values affect the output. Users may see the factors that have a positive or negative
impact.
Success in machine learning has led towards artificial intelligence applications,
continuity to deliver better to the end client. Developing an autonomous system to
learn and decide automatically. To maintain high level of accuracy and prediction
explainable model work efficiently. It enables human user to understand, trust and
efficiently manage the artificially intelligent partners. In the Fig. 4 new machine
learning system will be able to justify their actions, identify their advantages and
disadvantages, and communicate how they will act in the future. It is shown in
Fig. 4.
Number of advantage in XAI which put them in limelight and emphasizes the need
for faster large-scale adaptation around the world, the benefits of XAI can be found
across wide range of sectors and job activities. Some of the benefits of XAI are listed
below:
• Building user trust
• Reducing cost of mistakes
• Satisfying legal requirements
• Code confidence and compliance
7 Role of Explainable Edge AI … 109
Using the XAI in the system the user can understand how the output is generated with
relevant inputs. When the error occurs and the procedure to reduce the error. It also
provides the risk assessment report for the risky posed by the automated decision
making system, so the user trust the system.
AI model have shown most significance evidence of bias. It has proven a game-
changer for many users. User is seeing improved to increase productivity and
efficiency hence the legal requirements is satisfy using the XAI.
XAI provide the explanation knowledge with the accuracy tends to increase the
system’s confidence. Critical system such as banking and healthcare system re-quires
high code confidence from user for optimal solution. Increasing regulatory pressure
implies that companies should easily adapt and implement XAI in order to comply
with the government.
Using black box strategies to make an unexplained conclusion might result in legal,
ethical, and operational issues. Because black-box models are not verifiable or trace-
ability prior to deployment making assurances about their behavior difficult, explain-
ability is an issue that is both scientifically and socially relevant, and it is at the core
of several areas of contemporary machine learning and AI research. The following
are some of XAI’s challenges:
110 A. Kumar et al.
2.4.1 Bias
It’s a difficult challenge to make sure that the AI algorithm doesn’t learn biased or
unbiased world views as a result of gaps in the training data, model, or objective
function.
2.4.2 Safety
It’s tough to tell whether AI is trustworthy or not without looking at how it came
to its conclusions. Because of the generalization in statistical learning theory, which
essentially shows how organizations cover the gaps in unseen knowledge, this is a
challenge.
2.4.3 Fairness
This strategy can be applied to any number of algorithms or learning methods. The
model agnostic approach will aid businesses in approaching XAI’s internal workings
as a mysterious black box.
Model-Specific Technique
This method can be used for a small number of algorithms or a specific collection
of algorithms. The internal workings of XAI are treated as a white box in the model
agnostic approach.
XAI is a key research field. Our AI systems could become more reliable, legal,
efficient, fair, and robust with the help of explanations of their workings, which could
increase their acceptance and economic value. General opportunities and challenges
7 Role of Explainable Edge AI … 111
in XAI are analyzed potentially pros and cons from their perspective to overcome
the challenges in XAI [7]. Most of the organizations are working on their proprietary
XAI solutions. Real world implementation of XAI is discussed in this section.
The developing field of explainable AI (or XAI) can give banks more clarity on
their AI governance while assisting them in navigating concerns of transparency and
trust [11]. The goal of XAI is to improve the explainability, usability, and compre-
hension of AI models for human users without compromising performance or fore-
cast accuracy. A strong XAI programmed can benefit firms in a variety of other
ways as well. Depending on the questions being asked and the modeling methods
being employed, explainability tools can reveal various forms of information about
a model. To lead explainability initiatives in conjunction with their AI teams and
business departments, several banks are forming specific task forces. The accuracy
of a company’s total fraud detection is impacted by the insights that XAI provides
to teams by explaining why models are performing or not performing. Teams cannot
be expected to know how to increase fraud detection rates without this important
understanding. The two fundamental elements of XAI—feature importance (which
is technically connected to XAI) and model evaluation and performance—deliver on
these promises. Features look over a period of time specified data, such as payment
type, transaction amount, location, and so forth. Before deploying the model, teams
can test the model’s performance. Prior to relying on a machine learning model for
fraud detection and prevention, organizations should test the model’s performance
to determine its accuracy and false positive rate.
from their datasets and acting on them in real time, marketers can use XAI to alter
their entire marketing campaign if it is used properly. Customers can receive individ-
ualized messages at the right times throughout the consumer life cycle with the aid of
XAI. Additionally, XAI can assist marketers in identifying at-risk clients and deliv-
ering information to them that will encourage them to re-engage with the business.
When compared to humans, XAI can analyses tactical data more quickly and apply
machine learning to reach quick decisions depending on campaign and customer
context. Team members now have more time to concentrate on strategic projects
that will later guide XAI-enabled campaigns. With AI, marketers can use real-time
data to choose better media instead of waiting until the conclusion of a campaign
to make selections. Consumers react differently to communications across channels;
some may be moved by comedy, others by logic, and others by an emotional appeal.
XAI and machine learning can track which messages customers have interacted with
and build a more thorough user profile. From there, marketing departments can send
users more tailored communications based on their choices.
3.5 Summary
This chapter defines the briefly introduction of explainable artificial intelligence and
explain about the internet behavior efforts with real world examples of XAI. The
major points to recall are as follow:
• Explainable Artificial Intelligence is useful in building trust and confidence when
putting AI models into production.
• AI is differing from XAI in terms of black box decision.
• XAI approaches to interpret black-box testing.
• Behavioral efforts assist through XAI
• XAI deals with the Intelligence to interpret predictions from machine learning
models, resiliency, reliability, bias, and accountability are some of the other
characteristics.
• Users can use this analysis to create new situations and see how different input
values affect the output.
• In defense XAI to improve human decision-making in more effective to deal with
the real time problems.
• Autonomous vehicle is the rapidly expanding on road with effective features and
ease of access.
• XAI helps in fraudulent activities to detect the intruder and block their hacking
activities through blocking the card immediately.
• In marketing sector XAI achieves a tremendous success to deliver appropriate
messages to customer and establishing trust and keeping hold of existing clients.
7 Role of Explainable Edge AI … 115
4 Future of XAI
Since 2000, there have been 14 times as many active AI startups. Recent develop-
ments in deep learning have enabled AI to handle a variety of marketing and sales
choices as well as search engines, virtual assistants, and online translators. The level
of autonomous driving technology is quickly rising to that of fully automated driving.
Despite Elon Musk’s claim that “over a million robo taxis would be on the road for
sure by next year,” Modern GPUs are now effective enough to be utilized for activ-
ities other than visual rendering, such machine learning or crypto currency mining.
Data scientists found that these are repeated parallel jobs even if CPUs are typically
employed to do them. GPUs are therefore commonly employed in AI models to facil-
itate effective learning. A type of autonomous supervised learning is self-supervised
learning (also known as self-supervision). This method, unlike supervised learning,
handles the labeling process on its own and does not rely on humans to categories
the data.
5 Conclusion
The objective of this chapter has been contributed towards understanding the concepts
of explainable artificial intelligence to stores, processes, and manages data directly at
Internet of Things (IoT) endpoints. The internet behavior efforts promising to provide
reliable outputs that could be compared to experts’ interpretation, the explainable
artificial intelligence is growing rapidly in every sector of organization to complete
the task in easiest and user friendly. It will provide the steps through which end user
trust the system. The overall structure of this chapter provides an efficient knowledge
to the user to understand the concept of XAI.
References
1. Dikmen, T.L. Webb, B.P. Chang, Y. Benn, ‘The ostrich problem’: Motivated avoidance or
rejection of information about goal progress. Soc. Pers. Psychol. Compass, 7(11), pp. 794–807
(2013)
2. T.L. Webb, B.P. Chang, Y. Benn, ‘The ostrich problem’: Motivated avoidance or rejection of
information about goal progress. Soc. Pers. Psychol. Compass 7(11), 794–807 (2013)
3. L.A. de Souza Jr, R. Mendel, S. Strasser, A. Ebigbo, A. Probst, H. Messmann, J.P. Papa,
C. Palm, Convolutional neural networks for the evaluation of cancer in Barrett’s esophagus:
Explainable AI to lighten up the black-box. Comput. Biol. Med. 135, p. 104578
4. G. Nyman, J. Peltonen, M. Nelson, J. Karjalainen, M. Laine, T. Nyberg, H. Tuomisaari, On the
behavioral theory of the networked firm. In Big data and smart service systems, (pp. 179–192).
Academic press (2017)
116 A. Kumar et al.
1 Introduction
Artificial Intelligence (AI) has achieved tremendous growth in many activity sectors.
In the last few years, the existence of AI at the core of almost all developments can
be easily observed. The inclusion of learning, reasoning and adaptation capabilities
has helped AI enable systems to perform better than any human being [1]. Besides,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 117
A. E. Hassanien et al. (eds.), Explainable Edge AI: A Futuristic Computing Perspective,
Studies in Computational Intelligence 1072,
https://doi.org/10.1007/978-3-031-18292-1_8
118 S. Sinha and P. Vashisht
Explain to
Identify
Weakness
XAI
Explain to Explain to
Justify Monitor
Decision Performance
• An explanation is needed to justify the decision taken. This helps in building trust,
mainly when unexpected outcomes are observed. XAI systems will provide the
required information to explain the results.
With the tremendous growth in information sources, data fusion approaches are
being utilized to exploit them simultaneously for solving any learning task. Undoubt-
edly, using the fusion of heterogeneous data sources has drastically improved ML
node performance. Different data fusion techniques can also enrich the explainability
of ML models besides maintaining the privacy of training data used by ML models.
Data fusion can be done at a centralized or distributed level, where it can be either
knowledge fusion, data fusion or model fusion. Either of these is going to impact
the ML models and help provide results. But, the acceptability requires explanation,
and hence the fusion approaches need to be exploited to achieve explanation while
maintaining privacy.
The objective of this chapter is to
• Provide an overview of the essential concepts and terms used in XAI-related
research.
• Define diverse approaches to post-hoc explainability.
• Elaborate on different data fusion models and discuss the need for XAI for data
fusion.
• Highlight some of the application domains that demand XAI.
• Enumerate a series of challenges XAI, especially in the context of edge
intelligence.
120 S. Sinha and P. Vashisht
Data fusion is the combined examination of various interconnected datasets that offer
different perspectives on the same phenomenon. In general, more accurate inferences
can be drawn from the correlation and fusion of data from multiple sources than from
the examination of a single dataset. In the context of combining multiple data sets
for analysis, this section frames the primary Data Fusion Models and their levels.
Both in the research and business arenas, many data fusion frameworks have been
created. By choosing the best method for the given challenge, these frameworks have
been applied in several projects to assist in the creation of fusion systems. Conven-
tional data fusion models are discussed in this section, along with their existing capa-
bilities. It should be highlighted, however, that these structures can be redefined and
enhanced in order to use more modern techniques, such as switching from static to
dynamic processing, from human–machine teaming, from centralized to distributed
processing, and from the insignificant amount of data to enormous real-time data.
The Data Fusion Subpanel, later known as the Data Fusion Group, was established
in the middle of the 1980s by the Joint Directors of Laboratories. Data fusion
now includes sensor and information fusion because of the World Wide Web. The
JDL/DFIG announced a data fusion model that separated the various processes. The
model specifies the number of stages of fusion, each of which takes advantage of
recent advancements in AI to facilitate evaluation at a lower level (tier 0–3) and refine-
ment at mid-level (tier 4–6). However, contextual restrictions are based on mission,
objectives, and goals in system management (tier 6). The Data Fusion Information
Group (DFIG) model currently has seven tiers, and they are depicted in Fig. 2 and
are discussed as follows:
• Tier 0: Subject Evaluation, based on the relationship between the pixel and signal
levels, provides approximation and forecast the signal and object observable
states.
• Tier 1: Object Assessment, is constructed on data and their relationship and state
estimation, which offers valuation and prediction of entities and objects.
• Tier 2: Situation Assessment offers assessment and forecast of relationships
between objects and entities.
• Tier 3: Threat Refinement, estimates and predicts the impacts of participant’s
planned or anticipated activities.
• Tier 4: Process optimization is an aspect of resource management that offers
adaptive information collection and processing to give an assessment of sensing
targets and impact forecasting for planned or estimated actions by participants.
8 Explainable Data Fusion on Edge … 121
The model is useful for general understanding, but it does not direct a developer
in determining the methods that should be utilized [8]—as a result, it is ineffective
for creating an architecture for a practical system. In addition to broadening the func-
tional model and connecting the taxonomy to fields outside of the original military
focus, Steinberg, Bowman, and White proposed expanding and revising the JDL
model [9]. These changes included integrating a data fusion tree architecture model
for system description, design, and development.
of the detected phenomena are required to complete this task. These models might
draw from the experimental analysis.
Level 2: The feature extraction and feature fusion processes make up level 2. These
procedures are used to draw conclusions about the data on a symbolic level. While
maximizing information given, they seek to minimize the data content. Lists of
estimates with associated probabilities (and beliefs) are the result of this level.
Level 3: The association of object and events are drawn at this level. The infor-
mation that has been obtained, the libraries and databases that are available, and
the interactions between people are used to assemble potential courses of action.
The waterfall approach has the same flaws as the JDL model since they are iden-
tical. The waterfall model’s main drawback is the absence of any feedback data flow,
despite being more accurate than other models in its analysis of the fusion process.
Although not widely utilized abroad, the waterfall approach has been used in the
defence data fusion sector in Great Britain [10].
In 1999, Bedworth and O’Brien introduced the Omnibus model which is shown in
Fig. 4 [10]. This model combines the Boyd loop, Dasarathy, and Waterfall models.
The Boyd control loop is described by the authors as an iterative process with four
elements (observe, orient, decide, and act) acting in a close loop. The Dasarathy
model, in contrast, is made up of the three fundamental stages of data fusion: data,
feature, and decision.
The model was developed after examining the advantages and disadvantages of
previous models, and it incorporates the majority of the strong points of other method-
ologies. Although the model offers a much more fine-grained organization of the
8 Explainable Data Fusion on Edge … 123
processing layers, it exhibits a cyclic nature similar to the Boyd loop. The concept
is designed to be applied recursively at two different abstraction levels numerous
times within the same application. The model is first used to describe and structure
the entire system. Second, the system’s individual subtasks are modelled using the
same structures. In the Omnibus model, highly sophisticated hierarchical separa-
tion of the sensor fusion jobs does not permit a horizontal division into tasks that
reflect distributed sensing and data processing. Decomposition into modules that may
be independently implemented, independently tested, and independently reused for
other purposes is therefore not supported by the paradigm.
Data fusion is a complex idea with apparent benefits but also many difficulties,
especially with new paradigms with artificial intelligence integrated applications.
The development of artificial intelligence (AI) has led to a meteoric rise in AI-based
services. AI technology, such as machine learning (ML) and deep learning (DL),
achieves state-of-the-art performance in a variety of sectors. Additionally, present
applications frequently use centralized data management, which necessitates user
uploading of data to the central data centre. But there is an enormous amount of
data generated and gathered by billions of mobile users and Internet of Things (IoT)
devices that are dispersed at the network edge. Such a large amount of data requires
enormous bandwidth resources to upload to the distributed environment such as the
cloud, which may also cause users to experience intolerable latency [11]. With the
development of cloud computing, edge computing [12–16] brings cloud services
closer to end users. Edge computing provides platforms for computer resources
that are typically found at the edge of networks, such as networking, storage, and
processing. Edge servers, which can include IoT gateways, routers, and microdata
centres in mobile network base stations, on automobiles, among other locations, are
the devices that provide services for end devices. Edge intelligence is the result of
combining edge computing and AI to address the significant issues facing AI-based
applications.
124 S. Sinha and P. Vashisht
The discussion of edge intelligence is provided in the next part, along with exam-
ining several fundamental unresolved problems apart from potential theoretical and
technical directions.
3 Edge Intelligence
In edge intelligence, there are certain outstanding problems that remain unsolved. It is
essential to recognize, examine, and seek out fresh theoretical and technical answers
to these problems. In this article, we examine several significant edge intelligence
challenges and potential solutions. Scarcity of data at the edge, data consistency
on edge devices, poor adaptability of statically trained models, privacy and security
concerns, and incentive mechanisms are some of these difficulties [20].
learning may therefore be an option when there is a lack of target training data,
and there are some similarities between the source and target domains.
• A model can be made more reliable through data augmentation by having more
data during the training process [23]. For instance, expanding the number of
photos while maintaining the semantic meaning of the labels through flipping,
rotating, scaling, translating, cropping, etc. The network would become invariant
to these deformations and perform better with unseen data through training on
enhanced data.
Applications that rely on edge intelligence, such as voice recognition, activity recog-
nition, emotion recognition, etc., typically gather information from a significant
number of sensors that are dispersed throughout the edge network. The informa-
tion collected might not, however, be reliable. Different sensing contexts and sensor
heterogeneity are two variables that contribute to this issue. The environment (such
as the street and library) and its weather (such as rainy, and windy) introduce back-
ground noise to the sensor data that is collected, which may have an effect on the
model’s accuracy. The unexpected difference in the acquired data from sensors may
also be due to the heterogeneity of those sensors (such as hardware and software).
Various sensors, for instance, have various sensitivities, sampling speeds, and sensing
efficiency. Even the sensor data that was gathered from the same source may differ
between sensors. As a result, the model training would vary depending on how the
data were different, for example, the parameters of the features [24]. Existing sensing
applications still struggle with this variation.
If the model is trained centrally, this issue might be readily resolved. The ability
to learn the invariant characteristics to the variations is ensured by the centralized
huge training set. However, edge intelligence does not cover this ground. Future
research into this issue should concentrate on finding ways to counteract the varia-
tion’s detrimental impact on model accuracy. To achieve this, representation learning
and data augmentation are two potential research approaches that may be taken into
consideration. In order to make the model more resistant to noise, data augmentation
could be used to supplement the data during model training. For instance, in speech
recognition software for mobile devices, several types of background noise can be
added to mask environmental variance.
The performance of models is strongly influenced by data representation. In order
to develop models with more useful characteristics, representation learning focuses
on learning how to represent data [25], which could also be used to disguise hardware
disparities. The model’s performance would be much enhanced if we could “trans-
late” the representations between two sensors that are using the same data source
for this issue. As a result, representation learning offers a promising way to lessen
the effects of inconsistent data. Future efforts could be made in this regard, such as
developing processing pipelines and data transformations that are more efficient.
8 Explainable Data Fusion on Edge … 127
Heterogeneous edge devices and edge servers must cooperate to offer processing
power in order to realize edge intelligence. This technique involves sending locally
cached data and computation tasks (training or inference tasks) to unfamiliar
machines for additional processing. The data could include personally identifiable
information about users, such as tokens and images, which raises the possibility
of privacy breaches and attacks from hostile users. If the data is not encrypted,
malevolent people could easily get access to confidential information. A few initia-
tives [29–31] propose performing some local preparatory processing, which could
conceal sensitive data and cut down on the amount of data provided. The processed
data can still be used to extract personal information, albeit [32]. Additionally, by
adding malicious code, users might attack and take over a computerized device by
128 S. Sinha and P. Vashisht
controlling computer tasks being infected by a virus. It is the main obstacle that insuf-
ficient security and privacy protection measures methods to safeguard the security
and privacy of users attacked.
A viable solution is the credit system. The credit system employed by banks, which
verifies each user’s participation in the system and examines their credit information,
is comparable to this. The system would remove users with a history of poor credit.
As a result, everyone who uses computing equipment can trust them, and they are
all secure. Privacy could be protected via encryption, which is currently employed
in several works [20, 33]. However, before the training or inference tasks are carried
out, the encrypted data must be decrypted, which increases the amount of computing
required. Future work might focus more on homomorphic encryption to address the
issue [20]. A form of encryption known as homomorphic encryption enables direct
computation on ciphertexts and produces encrypted results. Once the data has been
decrypted, the outcome is identical to what would have been obtained by computing
the unencrypted data. Therefore, by using homomorphic encryption, the training or
inference process might be directly done on encrypted data.
The two most crucial phases of edge intelligence are model training and inference
and data collection. Securing the accuracy and usefulness of the information from
the gathered data during data gathering is difficult. The time, battery, and bandwidth
that data collectors use to perceive and gather data are all of their own resources.
It is unrealistic to believe that every data collector will be willing to contribute,
let alone for pre-processing data cleaning, feature extraction, and encryption, which
further uses resources. Each member must put aside personal interests in order to
cooperate for the purpose of collaborative model training/inference. For instance, one
master and multiple employees make up the architecture suggested in [34]. Workers
identify objects in a certain mobile visual domain and supply training instances for
masters via pipelines. Such an architecture is effective in private settings, such as at
home, where all of the equipment is driven to work together to develop a superior
intelligent model for their owner or master. It would not, however, function well
in public settings where the master initializes a job and assigns subtasks to new
players. In this situation, extra incentive issues develop that are not normally taken
into account in smart settings when all devices are not owned by a single master.
Participants must be motivated to collect data and complete tasks.
Future initiatives should consider reasonable incentive mechanisms. Participants
have a variety of missions, such as collecting data, processing it, and analyzing it, all
of which require different resources. Every participant wants to receive the greatest
potential prize. The operator, on the other hand, seeks to maximize model accuracy
at the lowest cost. The difficulties in creating the best incentive mechanism are
determining how to quantify the workloads of various tasks to match commensurate
rewards and how to jointly optimize these two competing objectives. Overcoming
these difficulties could be the focus of future work.
8 Explainable Data Fusion on Edge … 129
Although the deep learning tools available today can generate reliable results, they
are tough to comprehend due to their opaque nature. As the demand for trustable AI
worldwide is increasing, attempts are being made to integrate ethical standards during
design and implementation. Explainable AI, one of the components of trustable AI,
tries to dive deep into the rationale of the decision taken by the AI system. It is
considered a new research field that tries to make AI systems such that their behaviour
can be explained and the outputs are understandable to humans [35]. The primary aim
of XAI is to explore the logic involved in the decision-making process and outline
the merits and drawbacks of the system. This information can be utilized for further
improvements. Any XAI-powered AI system removes the confusion related to the
output of the system, making it a trustable system.
Several research activities around XAI have been carried out [37, 38] to date. Most
of this research highlights the goals that the explainable model possesses or should
possess, but the purpose behind achieving these goals is different from most of the
research carried out in XAI. A summary of some of the purposes of XAI is presented
in Table 2.
There are a few intriguing queries when the demand for XAI is generated. They start
with understanding the contextual aspects of XAI, i.e. what exactly is XAI, followed
by covering the identification of the purpose for which XAI has been demanded,
i.e. Why needed. And lastly, when one is conceived about exploring the inside of
the black-box models, they query how it can be done. The explainability of any AI
8 Explainable Data Fusion on Edge … 131
system can be considered not only after the black-box development has been done
rather, it can be done at all stages of development, namely, at the Pre-modelling stage,
at the model development stage known as Explanation modelling and also at Post-
modelling stage. At each stage, many ways of explainability exist, which depend
upon factors contributing to the AI model operations. Figure 5 shows the stages of
explainability of AI system [42].
• Pre-modelling Explainability—A suite of methods diverse in nature but having
a similar goal of understanding exists for pre-modelling explainability that works
mainly at the data level. These methods can be categorized as dataset description
standardization, exploratory data analysis, data set summarization and feature
engineering [43].
In real-time, the datasets usually come with incomplete documentation. This
makes the interpretation difficult. Standardization can make this interpretation
consistent, and smooth communication between users and creators regarding the
data can be done. Some specification standards are recommended as data state-
ments and data nutrition labels [42]. The exploratory data analysis aims to extract
a statistical summary that ranges from mean and standard deviation to the dimen-
sionality of the dataset. Dataset summarization provides a prediction on the given
dataset with the help of already used training data. As these training data are too
huge, posing storage challenges, a possible way is to store only a relevant subset
of the training data set. A good mechanism should be employed for selecting
the representative dataset. Explanation of any AI system can be as good as the
selection of features used to explain the predictions of model. Feature engineering
helps in identifying the relative importance of input features.
• Explainable Modelling—Explainable AI approaches can be applied at a local
or a global level which is either inherent or, Post-hoc. With the choice of specific
AI models that are inherently explainable, we can avoid the problem of explain-
ability due to the black box model. Linear models, decision trees etc., are some
Pre-Modelling Explainability
(Understand data used to develop model)
Explanable Modelling
(Develop models that are intrinsically explainable)
Post-Modelling Explainability
(Extract explanation from surrogate models to describe
pre-developed opaque models)
of the models that are inherently transparent and have low-performing power.
But, for high dimensional input, these simple, inherently explainable models also
do not guarantee explainability. Like, the linear regression model, when applied
to high dimensional data, may not be simulatable or, we can say, explainable.
It is customary to use an inherently simple or explainable model approach with
the black box method to obtain a high-performing explainable model. To better
understand, deep K-nearest neighbour is the amalgamation of K-nearest neigh-
bour inference applied on opaque training set learnt through deep hidden layers.
Self-explanation neural network and deep weighted averaging classifier (DWAC)
are a few more models obtained from the combination of the traditional transparent
approach and the black box model.
• Post-modelling Explainability—After the explainability, attention to data and
model next comes the explainability after the fact. The basic idea behind this
is to provide an explanation of the simplification or, Feature level explanation.
The post-hoc explainability techniques can be grouped based upon their target,
i.e. what they aim to explain about the model. The drivers are the cause for the
explanation, the explanation family that communicates with the user about the
driver and the target and finally, the estimator that deals with the overall computa-
tional process of obtaining the explanation. The local interpretable model-agnostic
explanation tool (LIME) [44] explains an instance of a complex deep learning
model, which is the model’s target. Input features serve as the driver, and the
importance score of each feature is communicated to the user that is computed
through the local evaluation of the model input termed as an estimator. The targets
are objects of an explainability method. They vary in their scope and complexity
and can be of inside or outside type. Inside targets are bothered about the details
internal to the model, and outside targets refer to the functionality or output of any
model. Drivers are a set of factors that impact the target, and explainability deals
with the causal relationship between the two. Training samples, input features and
choice of optimization algorithms can affect AI model development; hence, they
are the explanation drivers.
The post-hoc explainability methods aim at providing information content that
is easily understandable by a human. The impact of the driver on the target should
be communicated in an easy and understandable way, ensuring interpretability and
completeness. The importance score of each of the drivers, if–then based on decision
rules and decision trees, are some of the ways to explain the relationship between
the target and the drivers. In the fourth category, the estimators have a wide range,
and they vary in terms of their applicability and the underlying mechanism. The
estimation methods are either model specific, which works only for a specific method,
or model-agnostic, which can be used for any black box model. These models can
be used at the local level that offers interpretation for a particular instance or can
be applied to a global scenario where surrogate representation can be developed to
approximate any black box model.
Even though post-hoc explainability provides an understanding of the black-box
models, there are certain limitations to it [43]:
8 Explainable Data Fusion on Edge … 133
• These methods do not precisely match; instead only approximate the underlying
models while creating a surrogate white box model to explain the decisions made
by the system. But since it is only based on approximation, it cannot truly rely on
critical applications.
• These methods work for sample data collected randomly instead of the complete
data set. Due to randomness in the sampling process, the post-hoc explainability
suffers instability.
These shortcomings highlight the instability and low fidelity of post-hoc explain-
ability methods and simply points out that the post-hoc explainability methods are
helpful for model design and development but cannot be relied upon for regulatory
use.
is needed that the tools used here must be explainable so that the reasons behind
the decisions taken and selections made should be clear, making the tool/system
trustable.
• Military
The need for XAI was first felt in the military domain, and after a call for research
by DARPA [7] project, it widely spread in other domains too. AI has several
application areas in the military where it can showcase its potential. It can be used
in automated surveillance for tracing suspicious activities. Machine learning is
being applied for internal security and used to forecast and protect unauthorized
intruders. All these need to be a trustable and reliable system. They require XAI
to get better understandability as they all involve life and death.
• Legal
The law is the force to maintain decorum in society. All aspects of business, be
it merger, acquisition, selling or purchase, require a legal contract. As a powerful
abiding force in everyday’s life, the law is also influenced by advanced techno-
logical developments. In the legal domain, AI is leveraged for contract analytics,
criminal justice, litigation predictions, lawyer bots, etc. All these use ML power,
so complex models can be easily presented in a simple form. As the impact of all
these legal services critically influences human life, understanding their working
and decision-making is very much needed. XAI is essentially required to obtain
transparency, and recently, a few works have been initiated for the same [46, 47].
• Healthcare
The continuous growth in medicine and other healthcare monitoring aspects is
credited to the technological developments in recent years. The advanced devices
available for patient monitoring continuously capture data, and their analysis can
give much information regarding patient conditioning that is impossible even for
humans to explore. With this large amount of collected data, the information
provided by traditional AI techniques and even ML methods is sometimes insuf-
ficient. Recently, these enormous amounts of data are being used by deep learning
methods to identify the association between different types of patient data and
use them effectively. Even though the DNN-based AI techniques lead to a new
era of digital healthcare, the challenges still persist. With the number of parame-
ters approximating millions in number in the deep learning model, it is difficult
to exactly comprehend what the model portrays of the data, for example, inter-
preting the radiological images [48]. From the black box of DNN computations,
expressing intuitive interpretations that can clarify model uncertainty, support
model results, or provide new clinical insights is difficult. Explainability and
understandability are very important for the success of these models. Recently,
AI-based visualization methods have been developed to describe the AI models.
Some widely used models include occlusion maps [49], salience maps [50], and
class activation maps [51]. A few studies [52–54] have explored explainability in
medical AI in recent years.
The work in each of the domains supports the necessity for XAI. Such study is,
however, still in its early stages, and significant research efforts must yet be made
8 Explainable Data Fusion on Edge … 135
7 Conclusion
The conventional “Black Box” models that are traditionally being used in AI systems
are often marked for their lack of transparency and the way they hide potential biases.
XAI in fields like criminal justice, medical decisions and search engine outputs can
be helpful in removing potential biases while giving the proper explanation for the
decisions taken by the systems. Use of XAI leads to fair and ethical practices for
development of intelligent and responsible AI. They help in building trust for the
system. In future, XAI may find itself deep rooted in the field of medicine where it
provides the advantage of explanation that helps in easy understanding. The intent of
this chapter was to provide the brief about XAI that can help in better understanding
of the decisions made. In this chapter, the authors have discussed the emerging
discipline of edge intelligence. With the tremendous increase in the availability of data
and decreasing cost of computing devices, the rise of data-driven applications will
continue to proliferate. A Survey of available research highlights the long existence
of cloud and AI. With the increasing applicability of intelligence in real-time data-
centric applications, the cloud will be used to provide storage for AI training data,
and the edge will be utilized for inference and decision making. This chapter presents
a brief discussion on data fusion models and edge intelligence. With the increasing
performance of intelligent systems, the opacity of the system has increased. The
article takes a deep dive into the explainability of artificially intelligent systems.
The fundamentals of XAI are presented here in brief. At the same time, reviewing
the application domain of AI, the chapter details the opportunities in data-centric
applications and outlines the challenges and their solutions in data fusion and edge
intelligence.
References
1. D.M. West, The future of work: Robots, AI, and automation. Brookings Institution Press (2018)
2. A.B. Arrieta, et al. Explainable artificial intelligence (XAI): concepts, taxonomies, opportuni-
ties and challenges toward responsible AI. Inf. Fusion 58, 82–115 (2020)
3. G. Yang, Q. Ye, J. Xia, Unbox the black-box for the medical explainable AI via multi-modal
and multi-centre data fusion: A mini-review, two showcases and beyond. Inf. Fusion 77, 29–52
(2022)
4. D. Castelvecchi, Can we open the black box of AI? Nature News 538(7623), 20 (2016)
5. Preece, D. Harborne, D. Braines, R. Tomsett, S. Chakraborty, Stakeholders in Explainable AI,
(2018)
6. F.K. Došilović, M. Brčić, N. Hlupić, Explainable artificial intelligence: A survey, in 2018 41st
International Convention On Information And Communication Technology, Electronics And
Microelectronics (MIPRO) (IEEE, 2018)
136 S. Sinha and P. Vashisht
7. Gunning, David. “Explainable artificial intelligence (xai).“ Defense advanced research projects
agency (DARPA), nd Web 2.2 (2017): 1.
8. J. Llinas, D. L. Hall, An introduction to multi-sensor data fusion, in Proceedings of the
International Symposium on Circuits and Systems, vol 6537–540 (1998)
9. Elmenreich, Wilfried, A review on system architectures for sensor fusion applications, in
IFIP International Workshop on Software Technolgies for Embedded and Ubiquitous Systems
(Springer, Berlin, Heidelberg, 2007)
10. M. D. Bedworth and J. O’Brien, The omnibus model: A new architecture for data fusion?
In Proceedings of the 2nd International Conference on Information Fusion (FUSION’99),
(Helsinki, Finnland, 1999)
11. D. Xu et al., Edge intelligence: Empowering intelligence to the edge of network, in Proc.
IEEE 109(11), 1778–1837 (2021)
12. W. Shi, J. Cao, Q. Zhang, Y. Li, L. Xu, Edge computing: Vision and challenges. IEEE Internet
Things J. 3(5), 637–646 (2016)
13. T.X. Tran, A. Hajisami, P. Pandey, D. Pompili, Collaborative mobile edge computing in 5G
networks: New paradigms, scenarios, and challenges. IEEE Commun. Mag. 55(4), 54–61
(2017)
14. P.G. Lopez et al., Edge-centric computing: Vision and challenges. ACM SIGCOMM Comput.
Commun. Rev. 45(5), 37–42 (2015)
15. Y. C. Hu, M. Patel, D. Sabella, N. Sprecher, and V. Young, Mobile edge computing—A key
technology towards 5G, ETSI White Paper, vol 11, no 11 (2015), pp. 1–16
16. F. Bonomi, R. Milito, J. Zhu, S. Addepalli, Fog computing and its role in the Internet of Things,
in Proceedings of the first edition of the MCC workshop on Mobile cloud computing (MCC),
(2012), pp. 13–16
17. R, P. Kansakar, S. U. Khan, IFCIoT: Integrated fog cloud iot: a novel architectural paradigm
for the future internet of things. IEEE Consum. Electron. Mag. 6(3), 74–82 (2017)
18. Edge Intelligence in 2022—Deep Learning And Edge Computing—viso.ai
19. A. Munir et al., Artificial intelligence and data fusion at the edge. IEEE Aerosp. Electron. Syst.
Mag. 36(7), 62–78 (2021)
20. D. Xu et al., Edge intelligence: Architectures, challenges, and applications. arXiv:2003.12172.
(2020)
21. T. Xing, S. S. Sandha, B. Balaji, S. Chakraborty, and M. Srivastava, Enabling edge devices that
learn from each other: Cross modal training for activity recognition, in Proceedings of the 1st
International Workshop on Edge Systems, Analytics and Networking (ACM, 2018), pp. 37–42
22. F. Shahmohammadi, A. Hosseini, C. E. King, M. Sarrafzadeh, Smartwatch based activity recog-
nition using active learning, in Proceedings of the Second IEEE/ACM International Conference
on Connected Health: Applications, Systems and Engineering Technologies (IEEE Press, 2017),
pp. 321–329
23. Mathur, T. Zhang, S. Bhattacharya, P. Velickovic, L. Joffe, N. D. Lane, F. Kawsar, P. Lio,
Using deep data augmentation training’ to address software and hardware heterogeneities in
wearable and smartphone sensing devices, in 2018 17th ACM/IEEE International Conference
on Information Processing in Sensor Networks (IPSN) (IEEE, 2018), pp. 200–211
24. Mathur, A. Isopoussu, F. Kawsar, R. Smith, N. D. Lane, N. Berthouze, On robustness of cloud
speech apis: An early characterization, in Proceedings of the 2018 ACM International Joint
Conference and 2018 International Symposium on Pervasive and Ubiquitous Computing and
Wearable Computers (ACM, 2018), pp. 1409–1413
25. Y. Bengio, A. Courville, P. Vincent, Representation learning: A review and new perspectives.
IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013)
26. Z. Chen, B. Liu, Lifelong machine learning. Synth. Lect. Artif. Intell. Mach. Learn. 10(3),
1–145 (2016)
27. R. Vilalta, Y. Drissi, A perspective view and survey of meta learning. Artif. Intell. Rev. 18(2),
77–95 (2002)
8 Explainable Data Fusion on Edge … 137
28. Soller, J. Wiebe, A. Lesgold, A machine learning approach to assessing knowledge sharing
during collaborative learning activities, in Proceedings of the Conference on Computer Support
for Collaborative Learning: Foundations for a CSCL Community. International Society of the
Learning Sciences, 2002, pp. 128–137
29. H. Li, K. Ota, M. Dong, Learning IOT in edge: Deep learning for the internet of things with
edge computing. IEEE Netw. 32(1), 96–101 (2018)
30. A. E. Eshratifar and M. Pedram, “Energy and performance efficient computation offloading for
deep neural networks in a mobile cloud computing environment, in Proceedings of the 2018
on Great Lakes Symposium on VLSI. (ACM, 2018), pp. 111–116
31. Y. Kang, J. Hauswald, C. Gao, A. Rovinski, T. Mudge, J. Mars, and L. Tang, Neurosurgeon:
Collaborative intelligence between the cloud and mobile edge, in ACM SIGARCH Comput.
Archit. News 45(1), (ACM, 2017)
32. W. Yang, S. Wang, J. Hu, G. Zheng, J. Yang, and C. Valli, Securing deep learning based edge
finger-vein biometrics with binary decision diagram, IEEE Trans. Ind. Inf. (2019)
33. J. Konecnˇ y, B. McMahan, D. Ramage, Federated optimiza- ‘ tion: Distributed optimization
beyond the datacentre. arXiv:1511.03575, (2015)
34. Y. Huang, Y. Zhu, X. Fan, X. Ma, F. Wang, J. Liu, Z. Wang, Y. Cui, Task scheduling with
optimized transmission time in collaborative cloud-edge learning, in 2018 27th International
Conference on Computer Communication and Networks (ICCCN). IEEE, (2018), pp. 1–9
35. M. van Lent, W. Fisher, and M. Mancuso, An explainable articial intelligence system for small-
unit tactical behaviour, in Proceedings of the 16th International Conference on Innovative
Application of Artificial. Intelligence, (2004), pp. 900–907
36. A. Adadi, M. Berrada, Peeking inside the black-box: A survey on explainable artificial
intelligence (XAI). IEEE access 6, 52138–52160 (2018)
37. D. Doran, S. Schulz, T.R. Besold, What does explainable AI really mean? A new conceptual-
ization of perspectives. arXiv:1710.00794, (2017)
38. Gilpin, Leilani H., et al., Explaining explanations: An overview of interpretability of machine
learning, in 2018 IEEE 5th International Conference on data science and advanced analytics
(DSAA). IEEE, (2018)
39. P. Rani, C. Liu, N. Sarkar, E. Vanman, An empirical study of machine learning techniques for
affect recognition in human–robot interaction. Pattern Anal. Appl. 9(1), 58–69 (2006)
40. A. Chander, R. Srinivasan, S. Chelian, J. Wang, K. Uchino, Working with beliefs: AI trans-
parency in the enterprise, in Workshops of the ACM Conference on Intelligent User Interfaces,
(2018)
41. P. Langley, B. Meadows, M. Sridharan, D. Choi, Explainable agency for intelligent autonomous
systems, in AAAI Conference on Artificial Intelligence, 2017, pp. 4762–4763.
42. Online Source: https://towardsdatascience.com/the-how-of-explainable-ai-post-modelling-
explainability-8b4cbc7adf5f. Accessed on 12 jul 2022
43. Online Source: https://www.infoworld.com/article/3634602/explainable-ai-explained.html.
Accessed on 14 Jul 2022
44. M.T. Ribeiro, S. Singh, C. Guestrin, “Why should I trust you?” Explaining the predictions of any
classifier, in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining, 2016, pp. 1135–1144
45. Haspiel, Jacob, et al. “Explanations and expectations: Trust building in automated vehicles.
In Companion of the 2018 ACM/IEEE international conference on human-robot interaction.
2018.
46. C. Howell, A framework for addressing fairness in consequential machine learning, in Proc.
FAT Conf., Tuts., (2018), pp. 12
47. R. Berk, J. Bleich, Statistical procedures for forecasting criminal behavior: A comparative
assessment. Criminol. Public Policy 12(3), 513–544 (2013)
48. J.R. England, P.M. Cheng, Artificial intelligence for medical image analysis: a guide for authors
and reviewers. Am. J. Roentgenol. 212(3), 513–519 (2019)
49. M.D. Zeiler, R. Fergus, Visualizing and understanding convolutional networks, in European
Conference on Computer Vision, (Springer, 2014), pp. 818–833
138 S. Sinha and P. Vashisht
Abstract The amount of data generated, collected, and processed through computer
networks has increased exponentially in recent years. Network attacks have also
become an inherent concern in complex networks as a result of this increase of data.
The practise of assessing trust using attributes that influence trust is known as trust
evaluation. It is confronted with a number of serious challenges, including a shortage
of critical assessment data, a requirement for big data processing, a call for a simple
trust relationship expression, and the expectation of automation. Machine learning
(ML) has been applied to trust evaluation in order to overcome these issues and intel-
ligently and automatically evaluate trust. This research propose novel technique in
data fusion model with security and data optimization technique in edge computing.
Here the proposed data fusion is carried out using secure sequential discriminant
auto encoder in which the improvement of data accuracy, as well as for the maxi-
mizing of Edge-cloud based sensor networks lifespan. The fusion of edge cloud
data has been carried out using discriminant auto encoder which is integrated with
D. Prabakar
Department of CSE, SRM Institute of Science and Technology, Ramapuram Campus,
Chennai 600089, India
M. Sundarrajan
Department of CSE, SRM Institute of Science and Technology, Ramapuram Campus,
Chennai 600089, India
S. Prasath Alias Surendhar
Department of Biomedical Engineering, Aarupadai veedu Institute of Technology(AVIT),
Chennai, India
M. Ramachandran (B)
School of Computing, SASTRA Deemed University, Thanjavur, India
e-mail: [email protected]
D. Gupta
CSE Department, Maharaja Agrasen Institute of Technology, Delhi, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 139
A. E. Hassanien et al. (eds.), Explainable Edge AI: A Futuristic Computing Perspective,
Studies in Computational Intelligence 1072,
https://doi.org/10.1007/978-3-031-18292-1_9
140 D. Prabakar et al.
distributed edge cloud users, where the security of the network has been enhanced
using secure sequential fuzzy based trust model. The data optimization has been
established using Genetic swarm lightweight optimization algorithm. The experi-
mental analysis has been carried out based on data fusion as well as network security
model. The parametric analysis is carried out in terms of network security analysis,
throughput, Coverage fraction, Delay time, Energy consumption, Storage efficiency.
1 Introduction
Cloud computing (CC) has evolved into a viable alternative for providing an on-
demand platform for processing as well as sifting large amounts of data. Cloud
computing is being used in various fields, including but not limited to education,
finance, manufacturing, and healthcare [1]. As more methods rely on CC, novel-
methods that utilise benefits of CC while remaining lightweight and free of cloud’s
complexity are required to meet the properties of lightweight systems like IoT
devices. Although IoT devices can perform some essential activities like regulating,
actuating, and sensing, they cannot perform complex as well as sophisticated oper-
ations like controlling massive smart transportation methods or smart medical treat-
ments. Numerous IoT applications are time-critical, requiring fast decisions in order
to provide best possible performance. Examples include automotive networks as
well as aided healthcare systems [2]. Furthermore, CC is too sophisticated for these
devices to operate, and it lacks support for some of the most basic characteristics
of IoT methods, such as location awareness as well asbandwidth constraints. Fog
computing and edge computing are two novel technologies that can provide cloud
computing benefits while also addressing the unique peculiarities of IoT systems.
Although these new methods open the door to IoT system innovation as well as
growth, they also pose significant security and privacy concerns that may hinder IoT
system implementation and use [3].
With growth of IoT, a slew of smart applications are being developed that take
use of the ability to connect a variety of devices to internet. These apps will produce
a large volume of data, which will require to be analysed quickly in order to produce
useful as well as actionable data. EI refers to capacity to bring ML tasks from a remote
cloud closer to IoT devices, either partially or completely. Data fusion [4], also called
as data fusion or multi-sensor data fusion, refers to multisensor data resources. To
achieve goal of improving method stability, the system is able to obtain more precise
data. Data fusion is classified into three types based on the level of information
representation: datalevel fusion, featurelevel fusion, and decisionlevel fusion. The
datalevel fusion, which is the lowest level of fusion, immediately fuses collected
raw data, then extracts feature vector from fused data before being rated as well as
recognised.There is no risk of data loss as well as obtained result is most precise. In
9 Trust Model Based Data Fusion … 141
greenhouse WSNs, data fusion is a critical method for resolving low precision as well
as poor stability. It removes redundant dataas well as minimises data transmission
from various time and space multi-source data, meeting the goal of improving data
collecting accuracy as well as system stability.
Numerous activity areas that have adopted new data technologies have placed
Artificial Intelligence (AI) at the centre [5]. AI methods attains unprecedented levels
of performance while learning to solve gradually difficult computational tasks as
a result of these characteristics. AI-powered methods have lately advanced to point
where their design as well as implementation practically need no human intervention.
When such methods make decisions that affect people’s lives, there is a growing need
to comprehend how AI technologies deliver such data. While original AI methods
were simple to understand, opaque decision approaches such as DNNs (Deep Neural
Networks) have gained popularity in recent years. DL (Deep Learning) methods
like DNNs are empirically successful due to a combination of efficient learning
methods as well as their large parametric space [6]. However, a better understanding
of a method might lead to its shortcomings being corrected. Interpretability as an
additional method driver can enhance implementability of a machine learning model
for three reasons:
• Interpretability aids in ensuring objectivity in decision-making by detecting and
correcting bias in training dataset.
• Interpretability aids resilience by revealing potential adversarial disturbances that
could cause prediction to shift.
• Interpretability can serve as a guarantee that only relevant variables are utilized
to infer the result, i.e., that method reasoning is based on true causation.
All of this means that, in order to be regarded realistic, the system interpreta-
tion should provide either a knowledge of the model operations as well as predic-
tions, a visualisation of methods discrimination rules, or clues on what can disrupt
method. To prevent limiting effectiveness of today’s AI methods, eXplainable AI
(XAI) recommends developing a set of machine learning approaches. (1) build more
explainable models while preserving high learning performance (for example, predic-
tion accuracy), and (2) assist people in comprehending, suitably trusting, and effec-
tively managing the next generation of AI partners. XAI also analyses psychology
of explanation and draws on insights from Social Sciences [7].
Contribution of this chapter is as follows:
1. To propose novel technique in data fusion model with security and data
optimization technique in edge computing
2. To develop data fusion using secure sequential discriminant auto encoder in which
the improvement of data accuracy, as well as for the maximizing of Edge-cloud
based sensor networks lifespan
3. To enhance security using secure sequential fuzzy based trust model
4. To optimize the data transmission using Genetic swarm lightweight optimization
algorithm.
142 D. Prabakar et al.
2 Background
Statistical methods, hidden Markov methods, ANNs, and fuzzy logic are some of
solutions presented for IDS. Support vector machines (SVMs) were shown to be
highly accurate in creating IDS in recent studies [8]. However, its fundamental
drawback is that it necessitates a lengthy training period, limiting its use. Classic
data mining methods, such as association rule mining, are also used in IDS. For
example, to design and create IDS, [9] used rule-based approaches, where expert
knowledge is considered a rule set. As an analytical model, association rules were
utilised instead of human specialists. The extraction of a high number of association
rules, which improves model’s complexity, is a limitation of such methods. Feature
selection is one of the most important parts of a supervised classification model [10].
The computing time of the algorithms will be reduced by selecting essential features.
Many IDSs have been designed with feature selection because of the significant
aspects of network data [11]. In creating an IDS, work [12] classified main aspects
that are critical for real-world intrusion detection. In [13], the author used a feature
selection method to build a lightweight IDS. In big data framework for IDS, work
[14] created a PCA (principal component analysis) feature selection method. Work
[15] produced an IDS method that used a gain ratio as a feature selection strategy as
well as two classification methods, namely SVM as well as rule-based classification,
to identify class label as well as displayed greater accuracy levels for DoS attacks.
One disadvantage of this method is the high computational cost of using a separate
gain ratio and classification algorithm. They do, however, take a different approach
from the one taken in this SLR. In [16], the authors propose as well as assess a
framework based on distributed edge model that uses Docker method to create a
very lightweight as well as significant virtualization solution. Work [17] focuses on
relationship between DL as well as edge, whether it’s to utilise DL to optimise edge
or to execute DL algorithms on the edge. The research is separated into five sections:
edge DL applications, inference, edge computing and optimization. In [18], author
took a comprehensive look into edge intelligence from perspectives of edge caching,
training, inference and offloading [19] take a broad look at edge intelligence from
perspectives of edge caching, training, inference and offloading [20] created a multi-
agent method for boosting monitoring as well as security in cloud IoT environment.
Security is provided by intelligent agents sharing trust as well as certificates for
authentication.
3 System Model
This section discuss novel technique in data fusion model with security and data
optimization technique in edge computing. Here the proposed data fusion is carried
out using secure sequential discriminant auto encoder in which the improvement of
data accuracy, as well as for the maximizing of Edge-cloud based sensor networks
9 Trust Model Based Data Fusion … 143
lifespan. The fusion of edge cloud data has been carried out using discriminant auto
encoder which is integrated with distributed edge cloud users, where the security of
the network has been enhanced using secure sequential fuzzy based trust model. The
data optimization has been established using Genetic swarm lightweight optimization
algorithm.
Edge servers have far more resources and computing power than sensors and IoT
devices at the network’s edge. Because edge servers receive data from a variety of
edge of network devices with varying application needs, edge server performance
needs are similarly varied. Edge servers are created with high-performance fixed
evaluating servers with steady power sources as well as high bandwidth network
connections when AI/ML processing with severe latency constraints is required. Edge
servers are power/energy restricted for mobile edge computation. UAVs are utilised
as mobile edge servers or devices because they are often powered by batteries. For
quick and energy-efficient AI/ML computation, these mobile edge devices frequently
use AI accelerators. Furthermore, as shown in Fig. 1, edge servers or fog nodes are
more agreeable to the incorporation of ML accelerators such as CNN, RNN, and
MLP accelerators.
Each edge server is in charge of a cluster of sensors and IoT devices at network’s
edge. Edge servers offer edge-of-network IoT devices with apps, content, context,
services, and storage. Edge servers, for example, can help with complicated image
processing as well as storage responsibilities for imaging data collected from IoT
devices’ camera sensors. Our framework’s edge servers are linked to a top-tier
centralised cloud server. The core network sends locally processed data from edge to
cloud for a variety of reasons, including analytics, archiving, and large-scale decision-
making. High-performance servers are commonly employed in cloud since the cloud
must fulfil a large number of requests from network edge devices and edge servers.
As shown by data fusion blocks at every hierarchical layer in Fig. 1, data fusion
plays an essential role in the proposed system alongside AI. We present a high-level
overview of data fusion at several tiers of proposed method here, whereas Section
III provides a thorough study of data fusion within data fusion blocks in Fig. 1. IoT
devices then employ AI to improve accuracy, performance, and energy efficiency
of AI jobs by fusing data. The data obtained from different IoT devices is then
fused by the edge servers. The edge servers also use topological, contextual, and
environmental data in data fusion to determine topological links between sensors.
The edge servers employ AI to process the fused data as well as then send sanitised
and fused data, analytics, to cloud. Data obtained from network edge is merged in
the cloud, and then AI is used to the merged data to generate global analytics and
insights. Edge servers and fog nodes can also offload duties to the cloud.
Utilizing2 blocks: an encoder as well as decoder, we aim to learn probability
distribution (PD) of data: X in the VAE architecture. The encoder uses q(Z/X) to map
X . Decoder block will work in same way as encoder, transferring latent variables to
a new set of specifications that define a new set of related PDs: p(Xb/Z), from which
we will gather samples once more. Our network’s output will be these final samples:
Xb. The end goal is to resemble the network’s input and output, X and Xb, as closely
as feasible. However, to achieve that goal, we must map data’s internal structure to
PDs q(Z/X) and p(Xb/Z). Conditional PD p(Xb/Z) and q(Z/X) model likelihood of
Xb and Z but are dependent on their certain inputs: Z and X are the letters Z and
X, respectively. We train the PD: q(Z/X) and p(Xb/Z) in a VAE by employing a
variational technique, which converts learning method to a reduction method, which
can be simply expressed in terms of SGD in a NN. Method Specifications: θ and
ϕ are utilized to reflect architecture as well as weights of NN employed in Fig. 2.
These parameters are fine-tuned during the VAE training process as well as treated
as constants.
We strive to maximise the chance of getting the desired data as an output in the
variational approach by maximising a number known as Evidence Lower Bound
(ELBO) [22]. ELBO is made up of 2 parts: (1) a measure of distance between
PD of q(Z/X) and some similar reference PD, where distance is usually Kullback–
Leibler (KL) divergence; and (2) log likelihood of p(X) under PD p(Xb/Z), which is
probability of obtaining desired data (X) with final PD that generatesXb. All learnt
PDs are parameterized PD, which means that a set of parameters entirely defines
them. This is critical for the model’s operation because we use these parameters,
which are obtained as network node values, to represent corresponding PD: p(Xb/Z)
and q(Z/X).
Parts of loss function to be minimised using SGD for ID-CVAE method are
shown in Fig. 2. As previously stated, the loss function is divided into 2 parts: a KL
9 Trust Model Based Data Fusion … 145
divergence as well as log likelihood portion. Second portion uses the distribution
p(Xb/Z, L), which is a distance between X and Xb, to determine how likely it is to
generate X. By decreasing this distance, we are effectively preventing q(Z/X) from
deviating too far from its previous, and therefore functioning as a regularisation
term. This regularisation term has the advantage of being automatically modified,
therefore cross-validation is not required to adjust a hyper-parameter connected with
regularisation, as it is in other methods.
We utilize multivariate Bernoulli distribution for distribution p(Xb/Z, L). Because
the output specifications that describes distribution is mean, which is same as likeli-
hood of success, the Bernoulli distribution does not require a final sampling. Ground
truth X, which has already been scaled to [0–1], can be regarded as a [0–1] scaled
value for this likelihood. The distributions chosen for q(Z/X) and p(Xb/Z, L) are in
line with ones selected; they are simple as well as produce decent results. Boxes in
lower half of loss function’s certain choice. This is a specific choice for generic loss
function shown in Fig. 2.
Before entering following section, denote sample set {xi , yi }i=1 n
by a matrix form
X = [x1 ; x2 ; . . . ; xn ] ∈ R and Y = {y1 ; y2 ; . . . ; yn } ∈ R . Here, yi is an r-
nx n×r
dimensional one-hot row vector with a value of 1 in ith dimension and 0 in other
dimensions, and r is number of fault categories.
DIAE encoding method is same as AE method discussed in Sect. 3.1, in that it
uses Eq. to translate the input sample xi into
[ a hidden layer (1).] While introducing
an extra dimension to sample x i as: xi = xi1 , xi2 , . . . , xin−1 , 1 . The effect will be
the same as if you used bias term in Eq. (1).
Unlike traditional AE paradigm, DIAE decoding method is divided into two
stages. One method is to use the hidden representation h i Eq. (1) to reconstruct
input data:
( )
x̂i = g h i WxT (1)
146 D. Prabakar et al.
The other element is the input data prediction using h i Eq. (2):
α Σ 1 Σ ( ( ))
n n
J= (xi − x̃i )2 + −κσ yi , ŷi L
2n i=1 n i=1
β λ( )
+ ∥L − L T ∥2F + ∥Wx ∥2F + ∥WT ∥2F (4)
2 2
where α, β, λ are regularization parameters. Determine the optimal W X , Wr and L
by minimising J, because the matrices Wx∗ , WT: and L − are all randomly initialised.
Every term in Eq. (6) is described as follows for a better understanding:
• Reconstruction error is 2n 1 Σ n
i=1 (x i − ẋ i ) , which is the same as the traditional
2
AE. By reducing the use of this phrase. The ability of features to represent input
data is enhanced.
• Discriminant
( ( information
)) is represented by the second term
1 Σ n
n i=1 −κ σ yi , ŷi L . The extracted features’ discriminant ability would
be improved if this term was reduced. Here κσ (yi , yi L) is Gaussian kernel by
Eq. (5):
( ) (√ )(√ ) ( ( )2 )
κσ yi , ŷ L = 1/1 2πσ 2πσ exp − yi − ŷi L /2σ 2 (5)
Σ n
As pointed by n1 i=1 κσ (a, b) are viewed as an approximate evaluation of
correntropy Vσ (A, B) Eq. (6):
min J1 + J2 + J3 + J4 (7)
wx ,w1 L
where
( ( ))
J1 = −tr K σ Ŷ L , Y
α
J2 = ∥ Ẋ − X ∥2F
2
β (8)
J3 = ∥L − L T ∥2F
2
λ( )
J4 = ∥W X ∥2F + ∥W I ∥2F
2
To solve Eq. (9), we use a different optimization strategy. Fix Wγ and L first, then
optimise Wx using the gradient descent algorithm: Second, fix Wx and L then opti-
mise W I using the gradient descent approach. Finally, fix W X and WT , then optimise
L using the gradient descent approach. These three phases run in order until conver-
gence, starting with initialising W X , Wr , L. The following equations can be used to
calculate the values of W X , WT , L:
∂J
Wτ = W T + η
∂ WT
(9)
∂J
L = L +η
∂L
∂J ∂ J1 ∂ J2 ∂ y ∂1 ∂ WT
= + = +
∂ WX ∂ WX ∂ W X ωT WT WT
c T (( ( ) ) )
= 2H D Ỹ L − Y L T ⨀ dg(H WT ) + λWT (10)
σ
∂J ∂ J1 ∂ J4 C ( ) ( )
= + = 2 Ŷ T D Ŷ L − Y + β L − L T
∂L ∂L ∂L σ
Algorithm: 1—SSDAE:
{ }n
Input: Training sample set {X, Y } = x i: y i i=1 , the number of hidden layers K , neuron
{ }K
number in every hidden layer m, j −1 , maximum iteration number max_iter
{ }K
Output: Weight matrix of every hidden layer W x = W ix i=1 , and output of every hidden
{ }K
layer H i i=1
1. Evaluate H 0 = X
2. for i ← 1, K do
3. Start randomly W Xi and WTi ; Start L as a unit diagonal matrix
4. Construct DlAE network with m i hidden neurons:
5. Set iter = 1;
,
6. Utilize an alternative optimization technique to update W X , WTl , L ,
7. while iter < max iter do
8. If difference of loss function J in 2 consecutive iterations is smaller than threshold, break loop:
9. iter ← iter + 1:
10. end while
11. Build i - th hidden layer of SDIAE utilizing Wxi :
( )
12. Evaluate H i utilizing H i = g H i−1 W Xi :
13. X i+1 ÷ H i
14. end for
15. return W X
resource interaction or application service, we must analyse many trust metrics with
multiple critical attributes. The proposed approach for assessing trust is based on
Fig. 4. If there have been multiple contacts between ECUs clustered by common
tasks in past period of time, the corresponding trust degree of ECX is anticipated
using a capsule network to examine trust value of every ECU. As a result, we can
determine the overall level of confidence in the edge computing ecosystem.
The trust property of ECUs is described as Eq. (12):
where Vi, Vb, Vc represents identity trust, behavior trust and capacity trust.
(1) Identity Trust: When assessing a device’s identification trustworthiness, privacy
concerns must be taken into account. Exposing a device’s ID, for example,
exposes it to risk of identity theft, fraudulent data production, and so on. In
edge computing, public key infrastructure (PKI) has been widely employed
as a trustworthy authentication technique that can readily implement one-to-
one interaction in a distributed context. Identification trust Vi is defined as the
believability of an ECU’s identity during the resource request procedure. To
compute Vi, we assume that ECUs’ identity trust is assured by an anonymous
authentication mechanism based on ECC, which is utilized to evaluate each
ECU’s key. ECC generates a discrete elliptic curve E first, then generates public
as well as private keys based on one point O(x, y) on E in following 2 steps:
Step 1: Creating a Server Key The local edge server generates a key pair (Kp,
Ks) as shown in Eq. (13):
K p = H (Il , Ie )
( )
Ks = M K p (13)
The hash function H calculates the public key Kp based on the ID Il of the
edge server and the ID Ie of the ECUs. According to Kp, the private key Ks is
calculated using the encoding function M. Following the calculation of Kp and
Ks, the local edge server provides identity data D = {Kp, E, O} to ECUs.
Step 2: ECUs Key Generation: After the ECUs get D as Eq. (14), three keys
Kvi, Kep, and Kes are calculated:
where virtual key is used Kvi is used to safeguard ECUs’ genuine IDs, whereas
Kep and Kes are ECUs’ public and private keys, respectively, and R1() and R2()
150 D. Prabakar et al.
are PRNG functions. Then ECUs sends a message to the local edge server with
the format F = {IDe, Kv, Kep, a}, where an is a nonce integer. The public key
is used to contact edge servers and ECUs in the later authentication step, while
private key is used to complete identity matching. Vi is set to 1 if the ECUs pass
the authentication; otherwise, Vi is set to 0.
The proposed framework would calculate trust values utilizing fuzzy rules,
with fuzzy inputs being objective and subjective trust values as well as fuzzy
outputs being trust values.Low, medium, and high are the 3 states of fuzzy
outputs, and their ranges are 0 to 1. CSPj trust value maps to distrust if the
obtained trust value is low. CSPj trust value maps to trust if the acquired trust
value is high, and uncertain if the obtained trust value is low.
We define behaviour trust of ECUs as Vb as a weighted total of three terms by
Eq. (15), taking into account the process of service request as well as response:
where Vbc stands for behaviour constraint, Vbe for behaviour experience, and
Vbr for behaviour change rate. wbc+wbe+wbr = 1 are weight coefficients that
can be modified depending on the task at hand.
(2) Behavior Constraint: Varying ECUs observe different numbers of behaviour
restrictions or interaction specifications, which are found by certain tasks using
Eq. (16):
Vbe = w1 ξ1 + w2 ξ2 + · · · + we ξc (16)
Σ c
where ξi is constraint or specification that ECUs follows, i=1 wi = 1 are
weight coefficients.
Behavior Experience: Because the edge computing environment is always
evolving, it is vital to evaluate trust changes on a regular basis.ECUc is used to
define the ECUs that will be examined.Behavior experience Vbh indicates the inter-
active credibility of the recent past for interacting behaviours of ECUc with many
ECUs, where minimum engagement can also lead to effective trust development.
Assume that ECUs have interacted with k other ECUs, namely ECU1, ECU2, and
ECUk. Let’s see what happens if we use Eq. (17)
{ }
s j (Δt) = τ j1 , τ j2 , . . . , τ js , . . . , τ jt (17)
+
1 Σ H j (Δt) Ri (t)
k
Vbe = (18)
k j=1 t F̂i
Rate of Behavior Change: Because behaviour varies over time, Vbr, which indicates
the real change in behaviour trust, is defined as the rate of behaviour change. There
are last time series Δ(t − 1), mainly consider t − H j+ (Δ(t − 1)) of ECU j with
ECUc . For the time series Δ(t), the basic exponential smoothing factor is H j+ (Δt)
of ECU j with ECUc , λ.
Because ECUs typically do not keep a lot of previous data due to restricted
resources, the rate of behaviour change is based on two time series: Δ(t − 1), and
Δ(t). We believe historical data that is closer to the present moment to be more
suggestive, so recent data is given more weight Δ(t).
Capability Trust is described as degree of confidence in a device’s capability
property, such as accessibility, available bandwidth, response speed, and so on.
L = {e1 , e2 , . . . , en } is a set of ECU gives an ECUs . C = {c1 , c2 , . . . , cm } is
capability
{ property set of } ECU. Value of m of any ECU under an ECUn . is identical.
Cs = cs1 , cs2 , . . . , csm indicates any cak ∈ Cn requires based
⟨ on ck ∈ C, where cak⟩
is standard value preset based on task requirements. Vs = V⟨s = Cs1 , Cs2 , . . . ,⟩ Csn
notes a vector of ECUs which contains Cs of ECU , and V = C 1 , C 2 , . . . , C n is a
vector where C is actual C s when ECU is working.
The correlation coefficient of Vs and V is described as ρ. We propose that value
of denotes capacity trust of ECU s, which can be expressed as coefficient vectors
with the values of Vc · ωvT and ωv, T
.
CCA [12] method is used to evaluate ρ as follows:
{ }K
Wx = Wxi i=1 .
Max(N P) = α K P I A R E + β K P I L F I (19)
( ) (Σ Σ )2
M N
1 Σ n=1 In,i ϕn (t)\ϕmax
Nn
Ri (t) 1 i=1
NP = α + β( ( )2 (20)
Nn i=1 E M Σ M Σ N
i=1 n=1 I n,i ϕn (t)\ϕmax
Ri (t)
where Ê
is residual energy of a sensor node I at time t and is described as ratio of
node i’s remaining energy (Ri(t)) to each node’s initial energy (Ê) at time t. At the
moment of deployment, Ê is the same for all nodes in the network. We examine load
fairness at the edge servers for the second KPI. In, i is a binary indication, meaning
it is 1 if an nth gateway sends φn packets to ith server at time t, and 0 otherwise.
With an enhance in number of M and N, number of alternative configurations grows
exponentially. We use evolutionary methods, such as GA and DPSO, to address GE
configuration as an optimization issue. To reach an optimal setup as well as balanced
load, these algorithms go through the processes below.
1. Produce random population R 0 of size Δ. Best possible position for every
particle, i.e., Gateway, is started such that Pbest i0 = ri0 , ∀1 ≤ i ≤ Δ
2. Determine each particle’s fitness value for DPSO and every chromosome’s fitness
value for GA in R 0 and its global best position Gbest0 by Eq. (21)
( )
Gbest0 = arg max F Pbestli (21)
1≤i≤Δ
xI Igives particle I’s current position at the I th iteration, r1 and r2 are random
variables in (0, 1) range, a1 and a2 are acceleration constants utilized to pull
particles toward best position, and jw reflects inertia effect of the preceding
particle’s velocity.
4. Best accessible γ chromosomes from the present population are then extracted
for GA. Present population is R I, and probability of selection is Ps. Iteration
number is simply updated in DPSO. i.e., I = I + 1.
5. In case of GA, crossover as well as mutation are done on γ. If best candidate
solution for GE configuration is found in DPSO, search is completed; otherwise,
Step 6 is performed.
9 Trust Model Based Data Fusion … 153
6. All of the processes from Step 2 are repeated for GA. In DPSO, each particle’s
personal best position is updated utilizing Eq. (23).
⎧ ( )
PbestiI −1 , i f F riI ≤ F PbestiI −1 )
PbestiI = (23)
riI , otherwise.
7. The global optimum position for DPSO is updated using Eq. (24).
⎧ ( ) ( ) ( )
⎨ arg max F Pbestli , i f F Pbestli > F Pbestl−1
i
Gbestli = 1≤i≤Δ
(24)
⎩ Gbestl−1 , otherwise.
i
Algorithm: 2—GSLOA
Input: dynamic CPU and memory utilization threshold
values i.e.U t c and U t m , respectively—computed periodically using Eq. (12); channel
condition C t
Output: migration list map → input for load optimization
module in Fig. 3
1. compute CPU utilization level of the edge node (E Nuc );
2. compute memory utilization slevel of the edge node (E Num );
3. compute channel condition (Cc )
4. For each node edge do
5. Select application m from edge node
6. Choose edge node n as destination node
7. Map = m,n
8. End if
9. End for
10. Return map
4 Performance Analysis
been carried out for enhancing data fusion model with security and data optimization
technique in edge computing. The proposed technique obtained network security
analysis of 90% for 500 epochs since the analysis has been carried out based on
number of epochs in NN as represented in Fig. 3. Then the throughput analysis of
proposed technique has been carried out and it obtained 98% for 500 epochs as
shown in Fig. 4. The analysis of coverage fraction obtained of 81% for 500 epochs
as shown in Fig. 5. The proposed technique obtained as 65% for 500 epochs of
energy consumption as shown in Fig. 6. Delay time of proposed technique is 55%
as shown in Fig. 7 for 500 epochs. Comparative analysis has been carried out for
storage efficiency of 92% for 500 epochs as shown in Fig. 8. From the above analysis,
proposed technique obtained optimal results (Figs. 9 and 10).
5 Conclusion
This research proposes novel technique in data fusion model with security and data
optimization technique in edge computing. Here the proposed data fusion is carried
out using secure sequential discriminant auto encoder in which the improvement of
data accuracy, as well as for the maximizing of Edge-cloud based sensor networks
lifespan. The fusion of edge cloud data has been carried out using discriminant auto
158 D. Prabakar et al.
encoder which is integrated with distributed edge cloud users, where the security of
the network has been enhanced using secure sequential fuzzy based trust model. The
data optimization has been established using Genetic swarm lightweight optimization
algorithm. The experimental analysis is carried out based on data fusion and network
security model. The parametric analysis is carried out in terms of network security
analysis of 90%, throughput of 98%, coverage fraction obtained of 81%, Energy
9 Trust Model Based Data Fusion … 159
References
1. W. Mao, W. Feng, Y. Liu, D. Zhang, X. Liang, A new deep auto-encoder method with fusing
discriminant information for bearing fault diagnosis. Mech. Syst. Signal Process. 150, 107233
(2021)
2. M. Lopez-Martin, B. Carro, A. Sanchez-Esguevillas, J. Lloret, Conditional variational autoen-
coder for prediction and feature recovery applied to intrusion detection in iot. Sensors 17(9),
1967 (2017)
3. C. Jia, K. Lin, J. Deng, A multi-property method to evaluate trust of edge computing based
on data driven capsule network. In IEEE INFOCOM 2020-IEEE Conference on Computer
Communications Workshops (INFOCOM WKSHPS), (IEEE, 2020), pp. 616–621
4. M.A. Jan, M. Zakarya, M. Khan, S. Mastorakis, V.G. Menon, V. Balasubramanian, A.U.
Rehman, An AI-enabled lightweight data fusion and load optimization approach for Internet
of Things. Futur. Gener. Comput. Syst. 122, 40–51 (2021)
5. B. Mahbooba, M. Timilsina, R. Sahal, M. Serrano, Explainable artificial intelligence (xai)
to enhance trust management in intrusion detection systems using decision tree model.
Complexity (2021)
6. A.B. Arrieta, N. Díaz-Rodríguez, J. Del Ser, A. Bennetot, S. Tabik, A. Barbado, S. García, S.
Gil-López, D. Molina, R. Benjamins, R. Chatila, F. Herrera, Explainable Artificial Intelligence
(XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion
58, 82–115
160 D. Prabakar et al.
Abstract The sensor networks have attracted a numerous research attention due to
its diverse applications ranging from surveillance and monitoring applications. The
sensor nodes are usually characterized as having scarce resources; hence energy effi-
cient mechanisms which can enhance the resource utilization are of great significance.
The integration of edge computing framework with the sensor network can aid in the
data collection, dissemination and decision making. Scheduling approaches which
divide the nodes into a number of set covers and monitor the given points of interest
with the desired confidence level along with the objective of maximizing coverage
and network lifetime have been proved a prominent approach. The determination of
set covers is a NP hard problem and is dependent on different network parameters
such as node contribution, trust values and coverage probability. In this scheme, the
node has to monitor the neighboring node parameters at regular intervals, which
incurs a huge number of communication overhead. The nodes in sensor network can
employ the learning strategy to determine its best possible action to enhance the
network coverage as well as network lifetime. The chapter proposes a LSTM based
strategy for an edge computing enabled WSN to determine the status of the node
depending on the network parameters such as number of communications, number of
packets transmitted and initial energy of the nodes. The proposed protocol is imple-
mented using tensor flow and keras libraries in the python language. The keras tuner
package has been used to determine the best parameters such as number of hidden
layers and number of neurons in each layer. The obtained parameters are used to
construct a hyper model and the efficiency of the model is evaluated in terms of
the loss function. The explainability of the proposed model is investigated using the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 161
A. E. Hassanien et al. (eds.), Explainable Edge AI: A Futuristic Computing Perspective,
Studies in Computational Intelligence 1072,
https://doi.org/10.1007/978-3-031-18292-1_10
162 P. Chaturvedi et al.
1 Introduction
The machine learning based approaches are considered as efficient because it requires
less complexity. The different applications in which machine learning approaches can
be applied are: localization [3–6], coverage and connectivity [7–13], anomaly detec-
tion [14–18], fault detection [19–22], routing [23–25], MAC [26–31], data aggrega-
tion [32–35], synchronization [36–38], congestion control [39–43], and quality of
service [44–46]. The survey on the applications of machine learning technique in
different applications in WSNs can be found in [47–50].
The benefits of employing machine learning based approaches for the coverage
problem are as:
(i) The set cover consisting of minimum number of nodes can be determined
quickly and easily.
(ii) The nodes can be classified as active/idle in the dynamic environment without
losing the data.
In [51], authors have proposed a regression based interface design approach
which maintains the connectivity of the network. In [52], SVM based communi-
cation approach has been proposed which reduces the communication complexity
and hence improves the connectivity of the network. In [53], random forest coverage
and connectivity based approach is proposed which improves the coverage area by
changing the attributes of the network. In [54], reinforcement learning based coverage
optimization approach is proposed which enhances the network lifetime.
3 Network Model
Figure 1 represents the network model considered for the proposed prediction based
strategy in edge computing based environment. At the lowest level the number of
sensor nodes is deployed to monitor the given environment. The sensor nodes collect
the information from the environment and transmit it to the edge server where it is
used for further processing and analysis. The advantage of the edge server is that
the data is processed near the source of the data generation. The edge server then
forwards the data towards the application layer which usually comprise of different
applications which may utilize the sensed information.
164 P. Chaturvedi et al.
4 Proposed Protocol
End
166 P. Chaturvedi et al.
c. Identifying the size of the data set. The considered data set consist of 100
rows and 9 columns. The column names are Node id, coverage probability,
trust, no of communications, no of packets transmitted, whether the node is
cluster head or not, degree of the nodes, energy consumed, no of targets and
the status of the node as either active or sleep.
d. Determining the number of nodes in the active and sleep state using the
pandas library in python language. The number of active and sleep nodes
based on the simulation of 100 nodes is as shown in the Fig. 3. The results
show that there are 60 active and 40 sleep nodes, so we can say that the data
set is balanced.
The dataset obtained through the simulation of the sensor network is provided as
input to the LSTM based neural network. The keras tuner package is used to obtain
the optimal parameter such as number of hidden layers and number of neurons in
each layer. The number of layers is varied from 2 to 20 and number of neurons is
varied between 32 and 512 with a step size of 32. The learning rate is considered
as 0.001, 0.0001 and 0.00001. The other simulation parameters are as shown in the
Table 1.
The hypermodel is constructed based on the parameter values obtained and then it
is trained for 100 trials with 5 executions per trial. The hypermodel is then executed for
the 1000 iterations to analyze its efficiency. The results of the training and validation
are discussed in the next section.
The performance of the proposed prediction based model is evaluated using the
simulation setup and parameters defined in the previous section as discussed in the
subsections:
170 P. Chaturvedi et al.
The network parameters obtained through simulation are provided as input to LSTM
based neural network model. To determine the optimal number of layers and neurons
the keras tuner package is used. The keras tuner regularly vary the number of hidden
layers, number of neurons and learning and train the model for the different learning
rates. The mea absolute error is considered as the loss function. The validation split
is considered as the 30% and the loss function for the validation data set is obtained
as val_mean_absolute_error: 0.179. The results of 10 best iterations are summarized
in the Table 2. It can be seen from the results that the best loss function is obtained
as 0.1604 for the number of layers as 4 and learning rate at 0.001.
The number of neurons in each layer for the 10 best trials is as shown in the
Table 3.
The score of the 10 best iterations is shown in the Fig. 5.
Using the best parameters thus obtained, a hyper model is constructed and the
prediction is done for the test dataset. The hypermodel is trained for the training
dataset and tested for the test dataset by executing the model for the 1000 iterations.
The results show that after 1000th iteration, the loss function is almost stable. The
loss function using the hypermodel for the training and test dataset is as follow:
[test loss, train loss]: [0.1099, 0.1099].
The comparison results of the loss function for the training and validation dataset
is as shown in the Fig. 6. The results show that the hypermodel accurately predicts
the node status for the testing as well as the validation dataset.
Table 2 Result of 10
Trial Num_layers Learning rate Score
iterations
1 4 0.01 0.16047
2 6 0.001 0.1706
3 6 0.01 0.1752
4 3 0.001 0.1778
5 3 0.01 0.1790
6 4 0.01 0.1791
7 4 0.01 0.1793
8 7 0.001 0.1798
9 2 0.001 0.1833
10 5 0.01 0.1853
Table 3 Number of neurons in each layer in each trial
Trial/union 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
1 224 352 192 96 64 192 288 352 384 448 384 288 320 384 64 320 352 256 224 288
2 224 96 384 448 320 480 352 32 320 288 160 32 96 384 384 160 160 96 352 480
3 352 32 192 96 160 128 480 320 192 256 128 256 224 416 192 256 256 224 416 288
4 128 192 288 32 352 320 416 416 64 32 416 256 448 96 448 256 416 320 32 448
5 224 480 320 224 256 96 480 480 160 256 224 416 320 352 128 192 128 128 96 384
6 384 192 96 192 256 320 448 32 512 416 480 480 448 64
7 160 448 512 512 384 480 256 384 448 192 320 480 160 160 384 448 416 256 416 384
8 288 320 32 160 480 192 352 32 192 96 64 160 192 320 256 256 96 96
10 A Deep Learning Based Target Coverage Protocol for Edge …
9 128 384 416 320 352 448 256 256 384 416 224 64 352 416 192 288 96 256 224 192
10 320 416 320 256 320 352 480 192 352 352 256 128 512 96 384 448 320 64 384 32
171
172 P. Chaturvedi et al.
The machine learning models are usually considered as the black box which only
produces the results in the form of prediction and the impact of the different param-
eters are usually hidden from the outside world. So the LIME package is used to
understand the most significant parameter in the prediction of the node status.
a. Positive
Figure 7 represents the local explanation of the various features on the node
status being as active. The green color represents the positive correlation with
10 A Deep Learning Based Target Coverage Protocol for Edge … 173
the node status and the red color represents the negative correlation on the node
status. It can be inferred from the figure that the attributes CH or not, energy
consumed, number of packets and number of targets have positive correlation on
the node status, whereas the coverage probability, node degree, trust and number
of communications have negative correlation on the node status.
b. Negative
Figure 8 represents the local explanation of the various features on the node
status being as sleep. The green color represents the positive correlation with
the node status and the red color represents the negative correlation on the node
status. It can be inferred from the figure that the attributes CH or not, energy
consumed, number of packets and number of targets have positive correlation on
the node status, whereas the coverage probability, node degree, trust and number
of communications have negative correlation on the node status.
The correlation of various features on the node status is shown in the list form in
the Fig. 9. The active status is represented as 0 and sleep status is represented as 1.
Row 0
The local importance of various features for the first row in the dataset has been
analyzed for the node status as active or sleep as shown in the Fig. 10. The parameter
values for the row are as shown in the Table 4.
Row 1
The local importance of various features for the first row in the dataset has been
analyzed for the node status as active or sleep as shown in the Fig. 11. The parameter
values for the row are as shown in the Table 5.
In this section we have implemented the LGBM classifier for the considered dataset
to improve the training efficiency and faster convergence. The other simulation
parameters are:
Learning_rate = 0.09, max depth = 5 random state = 42, Verbose = 20,
eval_metric = ‘logloss’.
The results of the model prediction for the training and validation dataset are as
shown in the Table 6.
The binary logloss function is used to evaluate the efficiency of the LGBM clas-
sifier as shown in the Fig. 12. The loss value is approaching to minimum in 200
iterations for the training as well as the validation, which proves the efficiency of the
proposed model.
10 A Deep Learning Based Target Coverage Protocol for Edge …
Fig. 10 Local impact of feature on the node status for the row 0
175
176 P. Chaturvedi et al.
The global feature importance on the node prediction is plotted as shown in the Fig.
13. The results show that the most dominant features for the node status prediction are
Column 0–coverage probability, Column 1–trust, column 3–CH or not and Column
4–node degree.
The paper presented a LSTM based data prediction approach which can be used
to determine the node status in the scheduling approach to monitor a given set of
targets. The keras tuner package has been used to determine the optimal number
of neurons and number of hidden layers. The results are then applied to construct
the hypermodel which is then used to predict the node status. The loss function for
the model is determined for the different learning rate, number of hidden layers and
number of neurons. The lime package is used to determine the local feature impact
on the target attribute. The LGBM based classifier is used as a boosting algorithm
to improve the prediction model. The simulation results show that the LGBM based
boosting strategy achieves the accuracy of 100% and 97% for the training and testing
dataset.
As a future work, we aim to determine the different performance parameters for
the proposed prediction model such as network lifetime, energy consumption and
other QoS parameters.
10 A Deep Learning Based Target Coverage Protocol for Edge …
Fig. 11 Local impact of feature on the node status for the row 1
177
178 P. Chaturvedi et al.
Table 6 Loss values for the LGBM classifier for training and validation dataset for 100 Epochs
Epochs Training loss Validation loss
0 1 0 1
20 0.1118 0.1118 0.1277 0.1277
40 0.0356 0.0356 0.0618 0.0618
60 0.0118 0.0118 0.0554 0.0554
80 0.0038 0.0038 0.0637 0.0637
100 0.0014 0.0014 0.0747 0.0747
References
1. J. Yick, B. Mukherjee, D. Ghosal, Wireless sensor network survey. Comput. Netw. 52(12),
2292–2330 (2008)
2. R. Mulligan, H.M. Ammari, Coverage in wireless sensor networks: a survey. Netw. Protoc.
Algorithms 2(2), ISSN: 1943–3581
3. P. Chaturvedi, A.K. Daniel, An energy efficient node scheduling protocol for target coverage in
wireless sensor networks. Paper presented at the 5th international conference on communication
system and network technologies (CSNT-2015), 2015
4. J. Kuriakose, S. Joshi, R.V. Raju, A. Kilaru, A review on localization in wireless sensor
networks. Adv. Sig. process. Intel. Recogn. Syst. 599–610 (2014)
5. P. Cottone, S. Gaglio, G.L. Re, M. Ortolani, A machine learning approach for user localization
exploiting connectivity data. Eng. Appl. Artif. Intell. 50, 125–134 (2016)
6. M. Bernas, B. Paczek, Fully connected neural networks ensemble with signal strength clustering
for indoor localization in wireless sensor networks. Int. J. Distrib. Sens. Netw. 11(12), 1–10
(2015)
7. I. Ullah, Y. Shen, X. Su, C. Esposito, C. Choi, A localization based on unscented Kalman filter
and particle filter localization algorithms. IEEE Access 8, 2233–2246 (2020)
8. S M. Mohamed, H.S. Hamza, I.A. Saroit, Coverage in mobile wireless sensor networks (M-
WSN): a survey, Comput. Commun. 110(C), 133–150 (2017)
9. W. Fang, X. Song, X. Wu, J. Sun, M. Hu, Novel efficient deployment schemes for sensor
coverage in mobile wireless sensor networks. Inf. Fusion 41, 25–36 (2018)
10. H. Li, S. Wang, M. Gong, Q. Chen, L. Chen, IM2DCA: Immune mechanism based multipath
decoupling connectivity algorithm with fault tolerance under coverage optimization in wireless
sensor networks. Appl. Soft. Comput. 58, 540–552 (2017)
11. M. Abo-Zahhad, N. Sabor, S. Sasaki, S.M. Ahmed, A centralized immune-voronoi deploy-
ment algorithm for coverage maximization and energy conservation in mobile wireless sensor
networks. Inf. Fusion 30, 36–51 (2016)
12. W. Wang, H. Huang, Q. Li, F. He, C. Sha, Generalized intrusion detection mechanism for
empowered intruders in wireless sensor networks. IEEE Access 8, 25170–25183 (2020)
13. C. Miranda, G. Kaddoum, E. Bou-Harb, S. Garg, K. Kaur, A collaborative security framework
for software-defined wireless sensor networks. IEEE Trans. Inf. Forensics Secur. 15, 2602–2615
(2020)
14. Y. Gao, F. Xiao, J. Liu, R. Wang, Distributed soft fault detection for interval type-2 fuzzy-
model-based stochastic systems with wireless sensor networks. IEEE Trans. Ind. Informat.
15(1), 334–347 (2019)
15. J. Tan et al., An efficient information maximization based adaptive congestion control scheme
in wireless sensor network. IEEE Access 7, 64878–64896 (2019)
16. F. Hajjej, M. Hamdi, R. Ejbali, M. Zaied, A distributed coverage hole recovery approach based
on reinforcement learning for wireless sensor networks. Ad Hoc Netw. 101 (2020). 102082
17. Z. Sun, L. Wei, C. Xu, Z. Lv, An event-driven mechanism coverage algorithm based on sensing-
cloud-computing in sensor networks. IEEE Access, 7, 84668–84679 (2019)
18. B. Khalifa, A.M. Khedr, Z. Al Aghbari, A coverage maintenance algorithm for mobile WSNs
with adjustable sensing range. IEEE Sens. J. 20(3), 1582–1591 (2020)
19. A. Farhat, C. Guyeux, A. Makhoul, A. Jaber, R. Tawil, A. Hijazi, Impacts of wireless sensor
networks strategies and topologies on prognostics and health management. J. Intell. Manuf.
1–27 (2017)
20. P. Oluwasanya, Anomaly detection in wireless sensor networks. arXiv preprint arXiv:1708.
08053.
21. A.D. Paola, S. Gaglio, G.L. Re, F. Milazzo, M. Ortolani, Adaptive distributed outlier detection
for WSNs. IEEE Trans. Cybern. 45(5), 902–913 (2015)
22. M. Wazid, A.K. Das, An efficient hybrid anomaly detection scheme using K-means clustering
for wireless sensor networks. Wireless Pers. Commun. 90(4), 1971–2000 (2016)
180 P. Chaturvedi et al.
23. H.H. Bosman, G. Iacca, A. Tejada, H.J. Wörtche, A. Liotta, Spatial anomaly detection in sensor
networks using neighborhood information. Inf. Fusion, 33, 41–56 (2017)
24. C. O’Reilly, A. Gluhak, M.A. Imran, S. Rajasegarar, Anomaly detection in wireless sensor
networks in a non-stationary environment. IEEE Commun. Surv. Tutorials 16(3), 1413–1432
(2014)
25. T. Muhammed, R.A. Shaikh, An analysis of fault detection strategies in wireless sensor
networks. J. Netw. Comput. Appl. 78, 267–287 (2017)
26. H. Geng, Y. Liang, F. Yang, L. Xu, Q. Pan, Model-reduced fault detection for multi-rate sensor
fusion with unknown inputs. Inf Fusion 33, 1–14 (2017)
27. R.R. Swain, P.M. Khilar, S.K. Bhoi, Heterogeneous fault diagnosis for wireless sensor
networks. Ad Hoc Netw. 69, 15–37 (2018)
28. R. Palanikumar, K. Ramasamy, Effective failure nodes detection using matrix calculus
algorithm in wireless sensor networks. Cluster Comput. 1–10 (2018)
29. M. Hammoudeh, R. Newman, Adaptive routing in wireless sensor networks: QoS optimisation
for enhanced application performance. Inf. Fusion 22, 3–15 (2015)
30. X. Liu, Routing protocols based on ant colony optimization in wireless sensor networks: a
survey. IEEE Access 5, 26303–26317 (2017)
31. M. Asif, S. Khan, R. Ahmad, M. Sohail, D. Singh, Quality of service of routing protocols in
wireless sensor networks: a review. IEEE Access 5, 1846–1871 (2017)
32. K.-L.A. Yau, H.G. Goh, D. Chieng, K.H. Kwong, Application of reinforcement learning to
wireless sensor networks: models and algorithms. Computing 97(11), 1045–1075 (2015)
33. J. Kabara, M. Calle, MAC protocols used by wireless sensor networks and a general method
of performance evaluation. Int. J. Distrib. Sens. Netw. 8(1), 1–11 (2012)
34. I. Mustapha, B.M. Ali, A. Sali, M.F.A. Rasid, H. Mohamad, An energy efficient reinforcement
learning based cooperative channel sensing for cognitive radio sensor networks. Pervasive Mob.
Comput. 35, 165–184 (2017)
35. S. Kosunalp, Y. Chu, P.D. Mitchell, D. Grace, T. Clarke, Use of Q-learning approaches for
practical medium access control in wireless sensor networks. Eng. Appl. Artif. Intell. 55,
146–154 (2016)
36. M. Rovcanin, E. De Poorter, I. Moerman, P. Demeester, A reinforcement learning based solu-
tion for cognitive network cooperation between co-located, heterogeneous wireless sensor
networks. Ad Hoc Netw. 17, 98–113 (2014)
37. M. Rovcanin, E. De Poorter, D. van den Akker, I. Moerman, P. Demeester, C. Blondia, Exper-
imental validation of a reinforcement learning based approach for a service-wise optimisation
of heterogeneous wireless sensor networks. Wireless Netw. 21(3), 931–948 (2015)
38. M. Ambigavathi, D. Sridharan, Energy-aware data aggregation techniques in wireless sensor
network. Adv. Power Syst. Energy Manag. 165–173 (2018)
39. k. xie, L. Wang, X. Wang, G. Xie, J. Wen, Low cost and high accuracy data gathering in WSNs
with matrix completion. IEEE Trans. Mob. Comput. 17(7), 1595–1608 (2017)
40. H. Lin, D. Bai, Y. Liu, Maximum data collection rate routing for data gather trees with data
aggregation in rechargeable wireless sensor networks. Cluster Comput. 1–11 (2017)
41. E. Kanjo, E.M. Younis, N. Sherkat, Towards unravelling the relationship between on-body,
environmental and emotion data using sensor information fusion approach. Inf. Fusion 40,
18–31 (2018)
42. D. Capriglione, D. Casinelli, L. Ferrigno, Analysis of quantities influencing the perfor-
mance of time synchronization based on linear regression in low cost WSNs. Measurement
77(Supplement C), 105–116 (2016)
43. J.J. Prez-Solano, S. Felici-Castell, Adaptive time window linear regression algorithm for accu-
rate time synchronization in wireless sensor networks. Ad Hoc Netw. 24 (Part A), 92–108
(2015)
44. G. Betta, D. Casinelli, L. Ferrigno, Some Notes on the Performance of Regression-based Time
Synchronization Algorithms in Low Cost WSNs (Springer International Publishing, Cham,
2015)
10 A Deep Learning Based Target Coverage Protocol for Edge … 181