Internet of Things: Advances in Information and Communication Technology
Internet of Things: Advances in Information and Communication Technology
Deepak Puthal
Saraju Mohanty
Baek-Young Choi (Eds.)
Internet of Things
Advances in Information and
Communication Technology
123
IFIP Advances in Information
and Communication Technology 683
Editor-in-Chief
Baek-Young Choi
Editors
Internet of Things
Advances in Information and
Communication Technology
6th IFIP International Cross-Domain Conference, IFIPIoT 2023
Denton, TX, USA, November 2–3, 2023
Proceedings, Part I
123
Editors
Deepak Puthal Saraju Mohanty
Khalifa University University of North Texas
Abu Dhabi, United Arab Emirates Denton, TX, USA
Baek-Young Choi
University of Missouri at Kansas City
Kansas City, MO, USA
This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the
material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation,
broadcasting, reproduction on microfilms or in any other physical way, and transmission or information
storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now
known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice and information in this book are
believed to be true and accurate at the date of publication. Neither the publisher nor the authors or the editors
give a warranty, expressed or implied, with respect to the material contained herein or for any errors or
omissions that may have been made. The publisher remains neutral with regard to jurisdictional claims in
published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
The rapid evolution of technology has led to the development of the Internet of Things
(IoT), a network of physical objects that are embedded with sensors, software, and
network connectivity, enabling them to collect and exchange data. The IoT is trans-
forming our digital landscape, and the IFIP Internet of Things (IFIP-IoT) 2023 con-
ference is a crucial platform for scholars, researchers, and practitioners to come
together, share ideas, and advance this transformative field.
This edited book is a compilation of cutting-edge research and developments pre-
sented at the IFIP-IoT conference. The conference serves as a dynamic hub where
experts from diverse backgrounds come together to explore the multifaceted aspects of
IoT, from its technological foundations to its far-reaching implications for society,
industry, and beyond.
The chapters in this book are a testament to the collaborative spirit of the IFIP-IoT
community. They offer insights into the latest innovations, challenges, and opportu-
nities in IoT, covering a wide array of topics, including IoT architectures, security and
privacy, data analytics, edge computing, and applications in various domains. These
contributions not only reflect the state of the art in IoT research but also provide
valuable perspectives that pave the way for future breakthroughs.
The IFIP-IoT Conference is an annual IFIP event dedicated to IoT research, inno-
vation, and applications, emphasizing the multidisciplinary nature of IoT. IoT
encompasses topics from network protocols and embedded systems to analytics,
machine learning, and social, legal, ethical, and economic considerations, enabling
services in e-health, mobility, energy, manufacturing, smart cities, agriculture, and
more. Security, privacy, and societal aspects are essential in IoT deployment. IFIP-IoT
covers these diverse areas, seeking papers showcasing technical advancements,
research, innovation, pilot results, and policy discussions. Contributors include
researchers, users, organizations, ICT industry experts, authorities, and regulators.
IFIP-IoT welcomed full and short paper submissions, with full papers being original
and unpublished elsewhere. Poster presentations were limited to student papers. The
conference program featured keynotes, plenary talks, tutorials, technical sessions,
special sessions, expert panels, a research demo session (RDS), and a student research
forum (SRF). New tracks like “SRF” and “RDS” aimed to enhance event participation.
The paper submission guidelines include an 18-page limit for full papers, which
applied to both regular and special sessions, as well as an 8-page limit for short papers,
applicable to any session, including SRF and RDS. To ensure a thorough review
process, we implemented a four-tier review mechanism within EDAS, consisting of
TPC-Chairs, Track Chairs, TPC members, and dedicated reviewers. We took measures
to address conflicts of interest by appointing multiple TPC chairs and multiple track
vi Preface
chairs for each track. Additionally, we imposed a limit of 2 papers maximum for PC
members. SRF encouraged student first-author papers with an 8-page limit, while RDS
papers also had an 8-page limit and may or may not feature student first authors. It’s
important to note that TPC members were permitted to co-author papers with their
students in both SRF and RDS. Furthermore, our conference included Regular tracks/
sessions that accept submissions from any authors, as well as Special Sessions/Tracks
proposed by established researchers, with submissions received by invitation.
The IFIP-IoT conference had six regular tracks, each focusing on a different aspect
of IoT:
• Hardware/Software Solutions for IoT and CPS (HSS): This track covered the
design, development, and implementation of hardware and software solutions for
IoT and cyber-physical systems (CPS).
• Electronics and Signal Processing for IoT (ESP): This track focused on the use of
electronics and signal processing techniques for IoT applications.
• Artificial Intelligence and Machine Learning Technologies for IoT (AMT):
This track explored the use of artificial intelligence (AI) and machine learning (ML)
technologies for IoT applications.
• Cyber Security/Privacy/Trust for IoT and CPS (SPT): This track addressed the
security, privacy, and trust challenges of IoT and CPS systems.
• IoT or CPS Applications and Use Cases (APP): This track presented case studies
and applications of IoT and CPS technologies.
• Networking and Communications Technology for IoT (NCT): This track
focused on the networking and communication technologies used in IoT systems.
Leading IoT experts from around the globe proposed special sessions on cutting-
edge IoT topics. These session organizers then invited other established researchers to
submit papers to their sessions. We are pleased to announce that the following special
sessions took place and contributed excellent research papers to the IFIP-IoT 2023
conference:
• AI and Big Data for Next-G Internet of Medical Things (IoMT): This special
session explored the use of AI and big data for the next generation of IoMT
systems.
• Blockchain for IoT-Driven Systems (BIoT): This special session examined the
use of blockchain for IoT-driven systems.
• Edge AI for Smart Wearables (EAW): This special session focused on the use of
edge AI for smart wearables.
• Energy-Aware Security for IoT (EAS): This special session addressed the
security challenges of IoT systems, with a focus on energy efficiency.
• IoT for Smart Healthcare (SHC): This special session explored the use of IoT for
smart healthcare applications.
• IoT for Wearables and Smart Devices (IWS): This special session focused on the
use of IoT for wearables and smart devices.
• Metaverse for IoT (MIoT): This special session examined the use of the metaverse
for IoT applications.
Preface vii
• Security by Design for IoT (SbD): This special session discussed the importance
of security by design for IoT systems.
• Technologies for Smart Agriculture (TSA): This special session explored the use
of IoT technologies for smart agriculture.
In addition to the regular tracks and special sessions mentioned earlier, we intro-
duced two sessions to support graduate students, early career researchers, and ongoing
projects through short papers:
• Student Research Forum (SRF): This session was designed to provide valuable
opportunities for research scholars and graduate students. Presentations in this
session were in a concise oral or poster format.
• Research Demo Session (RDS): Authors in this session had the chance to show-
case live demonstrations and hardware prototypes of their research.
We are grateful to the authors who contributed their expertise to this volume, and we
commend their dedication to advancing the field of IoT. We would also like to
acknowledge the reviewers whose insightful feedback ensured the quality and rigor
of the included chapters.
We hope that this edited book will serve as a valuable resource for researchers,
educators, policymakers, and industry professionals alike, fostering a deeper under-
standing of IoT and inspiring further innovation in this transformative domain. As the
IFIP-IoT conference continues to evolve and grow, we look forward to witnessing the
continued impact of this vibrant community on the ever-expanding Internet of Things.
Deepak Puthal
Saraju Mohanty
Baek-Young Choi
Organization
General Chairs
Saraju Mohanty University of North Texas, USA
Shengli Fu University of North Texas, USA
Program Chairs
Baek-Young Choi University of Missouri Kansas City, USA
Deepak Puthal Khalifa University, UAE
Carlos Pereira Federal University of Rio Grande do Sul, Brazil
Publication Chairs
Hui Zhao University of North Texas, USA
Sharad Sharma University of North Texas, USA
Web Chairs
Laavanya Rachakonda University of North Carolina at Wilmington, USA
Omkar Dokur University of South Florida, USA
Media Chair
Satish Maheshwaram National Institute of Technology Warangal, India
Finance Chair
Bibhudutta Rout University of North Texas, USA
Registration Chair
Alakananda Mitra University of Nebraska Lincoln, USA
Publicity Chairs
Umamaheswara Tida North Dakota State University, USA
Uttam Ghosh Meharry Medical College, USA
Hemanta Mondal National Institute of Technology Durgapur, India
Sudeendra Kumar PES University, India
Dhruva Ghai Oriental University, India
Sujit Biswas University of East London, UK
Theocharis Theocharides University of Cyprus, Cyprus
Panel Chairs
Hao Zheng University of South Florida, USA
Alex Chiumento University of Twente, The Netherlands
Track Chairs
Regular Track - Artificial Intelligence and Machine Learning Technologies
for IoT (AMT)
Sejun Song University of Missouri Kansas City, USA
Yu Chen Binghamton University, USA
Special Track - AI and Big Data for Next-G Internet of Medical Things (IoMT)
Uttam Ghosh Meharry Medical College, USA
Layer-Wise Filter Thresholding Based CNN Pruning for Efficient IoT Edge
Implementations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
Lakshmi Kavya Kalyanam, Rajeev Joshi, and Srinivas Katkoori
IoT Based Real Time Monitoring of Delhi-NCR Air Pollution Using Low
Power Wide Area Network. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Prem Chand Jain, Vandavasi Satya Charan, Chitumalla Sahith,
and Rohit Singh
1 Introduction
all cases, in-vehicle intrusion attacks [1] can have serious consequences, including
loss of control over the vehicle, compromise of sensitive data, and potentially life-
threatening situations. It is important for vehicle manufacturers to take steps to
secure their systems and protect against these types of attacks.Figure 1 shows
the general scenario that is carried out by hackers to inject cyber-attack in vehi-
cles. Because of this, a lot of researchers have tried to come up with reliable
ways to find and analyze attacks. In the identification of the potential attacks,
a number of IDS based on Many classification approaches are proposed by dif-
ferent authors which include the use of machine learning (ML) methods such
as support vector machine (SVM), decision tree (DT), naive Bayes classifier
(NB), logistic regression (LR), and deep convolutional neural network (CNN)
[2–5]to detect network attacks. Current machine learning approaches for detect-
ing attacks on moving cars often utilize cyber threat studies that pair routing
data with potential attack profiles derived from behavioral analysis techniques
including packet collecting, feature comparisons, and packet filtering. In order
to classify potential attacks, determine the true nature of the attack, identify
the attacked ECU, and initiate countermeasures, routing data of CAN devices
is used. Hence, security administrators have turned to DNN approaches such as
LSTM, RNN, and CNN to identify complex threats from a variety of vectors.
In most cases, deep CNN provides a novel approach to enhancing the accuracy
of threat identification in network intrusion detection (NID) and decreasing the
occurrence of false positives (FPR). Nevertheless, a single base classifier doesn’t
adequately fit the data distribution, either because of an excessive amount of
bias or an excessive amount of variance [5].
The rest of the paper is organized as follows; In the Sect. 2, we present a literature
review of ML, DL, and graph network modes. In Sect. 3 the detailed proposed
system is discussed which includes text-to-image conversion of dataset and DL
models namely CNN and GNN design pipeline. In Sect. 4, we review research
on the theoretical and empirical outcomes and analyses of CNNs and GNNs, as
well as related research efforts. We finish with the suggested system in Sect. 5
6 P. Zambare and Y. Liu
2 Literature Review
2.1 Traditional IDS
Most of the traditional ways to find intrusions are based on statistical analysis
[9], threshold analysis [10], and signature analysis [11]. While these techniques
are effective at revealing malicious traffic behavior, they need the input of expert
knowledge from security researchers. This renders their predetermined rules and
limits mostly useless. In addition, this “experience” is really simply a description
of the hitherto unquantifiable behavior of hostile traffic. Because of this, these
methods can’t be easily adapted to the Internet of today, which has a lot more
network data and more unpredictable network attacks.
protocols if they are trained quickly. Fan and Ling-zhi [21] used a multilayer
CNN to get very accurate features. The convolution layer was connected to the
sampling layer below, and this model did better on the KDD99 dataset than
traditional detection algorithms like SVM. But CNN can only look at one input
package at a time; it can’t look at information about timing in a given traffic
scenario. In a real attack traffic scenario, a single packet is just normal data.
When a lot of packets are sent at once or in a short amount of time, this packet
is considered to be malicious traffic. In this situation, the CNN doesn’t work,
which could mean that a lot of alerts are missed.
3 Proposed System
In this study, we provide an Analytical Model for In-Vehicle NID. In order to
anticipate the stability of NID once suspicious network flows have been gathered,
the proposed NID model employs the GNN model with fine-tuned parameters.
The three steps of the process for classifying behavior are displayed in Figure 2
together with the general structure of the model: (1) Pre-processing of data, (2)
Text to Image Conversion, (3) applying DL-based model (CNN and GNN) (4)
model validation and performance analysis.
Normalizing the data is the first step in the process of transforming it. Con-
sidering the values of pixels in an image range from 0 to 255, so the scale for
network data should also range from 0–255. Quantile normalization is one of the
ways that data values can be changed so that they all fall in the same range. We
used quantile normalization in the proposed framework because it is very good
at handling outliers. Using the quantile normalization approach, feature distri-
butions are transformed into normal distributions, and the values of the features
are recalculated based on the normal distribution. Hence, most of the values
for the variables are quite near to the median values, which makes it simple to
handle any outliers [15]. Upon the completion of the normalization process, the
data is next divided into chunks according to the timestamps and sizes of the
features included within the network traffic datasets. Figure 4 shows the steps
that are taken to get the feature map from the dataset. These steps are feature
extraction and normalization. Each segment of 27 consecutive samples from the
Car-Hacking dataset, which includes 9 significant features are converted into
an image with the form 28283 and a total of 243 feature values. As a result,
each altered image is now a three-channel, square, color image (RGB). Figure 5
demonstrates the stark differences between the feature maps created by malware
and by legitimate applications utilizing the HRCL dataset.
C Performance Parameters
In this paper, detecting intrusions in a vehicle network is a multiclass problem,
and the result of the detection is a DOS attack on the vehicle network. We use
the F1-score, precision, recall, and accuracy as evaluation metrics to figure
out how well our proposed model works. In the confusion matrix, the variables
TP, FP, TN, and FN stand for the following: true negative (TN), true positive
(TP), false negative (FN), and false positive (FP).
we used the below formula for the calculation;
Accuracy (ACC):
TP + TN
ACC = (2)
TP + TN + FP + FN
Precision (PRE):
TP
P RE = (3)
TP + FP
Recall (REC):
TP
REC = (4)
TP + FN
F1 -Measure (F1):
P RE ∗ REC
F1 = 2 ∗ (5)
P RE + REC
D Result Analysis
We used the proposed GNN classifier to ensure that the prediction accuracy
of a DL learning classifier, like GNN, is better than that of other DL classi-
fiers. This was done to prove that the developed model was good at making
predictions. In this study, the model parameters for binary classification on
GNN and CNN were set using a 10-fold cross-validation scheme.
Table 1 provides the fine-tuned parameters for two classifiers: the CNN
Classifier and the GNN Classifier. The CNN Classifier uses ReLU as the acti-
vation function and employs the Adam optimizer with a batch size of 64. It
undergoes 25 epochs during fine-tuning. On the other hand, the GNN Clas-
sifier uses ReLU and Tanh as the activation function, and it uses SGD as the
optimizer with a batch size of 28. Similar to the CNN Classifier, it also under-
goes 25 epochs during fine-tuning. These fine-tuned parameters help adapt
the pre-trained models to specific tasks, leading to improved performance and
accuracy.
14 P. Zambare and Y. Liu
Fig. 8. Performance comparison graph of CNN (a) accuracy graph (b) loss graph.
Fig. 9. Performance comparison graph of CNN (a) accuracy graph (b) loss graph.
Table 1 shows the parameters that can be used to fine-tune the DL classifier.
Figure 8 shows a comparison of the accuracy and loss graphs of CNN algorithms
with hyperparameters that have been fine-tuned.
Figure 9 shows the accuracy and loss graph comparison of CNN algorithms
with fine-tuning of hyperparameters. Figure 9 shows the accuracy and loss graph
comparison of GNN algorithms with fine-tuning of hyperparameters. GNN gives
better accuracy compared to CNN as can be seen in Fig. 10
GNN Based IDS in Smart Vehicles 15
Fig. 10. Accuracy and F1-Score comparison graph of CNN and GNN
5 Conclusion
The use of DL techniques (CNN) and proposed graph-based techniques (GNN)
for attack detection in in-Vehicle Network systems has shown promising results.
With the increasing use of connected vehicles, the security of these systems is
paramount to ensure the safety of passengers and prevent malicious attacks.
We have used the car-hacking dataset in experiments where we have converted
text data into image data and passed it to the proposed classifier. Models such
as CNNs, GNN have been employed to identify different types of attacks in
vehicle network systems, such as DoS attacks, fuzzy attacks, spoofing the drive
gear, and spoofing the RPM gauge with the data stream. These models analyze
large amounts of data in real time, detect anomalous behavior, and provide
accurate predictions with high accuracy rates. We achieve an accuracy of 95 %
using the GNN classifier with fine-tuning of hyperparameters. Future research
should concentrate on enhancing these models’ scalability and efficiency, as well
as inventing new ways for detecting complex threats in real-time.
References
1. Khan, J., Lim, D.W., Kim, Y.S.: Intrusion detection system can-bus in-vehicle net-
works based on the statistical characteristics of attacks. Sensors 23(7), 3554 (2023)
2. Hossain, M.D., Inoue, H., Ochiai, H., Fall, D., Kadobayashi, Y.: LSTM-based in-
trusion detection system for in-vehicle can bus communications. IEEE Access 8,
185489–185502 (2020)
3. Yu, Y., Si, X., Hu, C., Zhang, J.: A review of recurrent neural networks: LSTM
cells and network architectures. Neural Comput. 31(7), 1235–1270 (2019)
16 P. Zambare and Y. Liu
4. Javed, A.R., Ur Rehman, S., Khan, M.U., Alazab, M., Reddy, T.: Canintelliids:
detecting in-vehicle intrusion attacks on a controller area network using CNN and
attention-based GRU. IEEE Trans. Netw. Sci. Eng. 8(2), 1456–1466 (2021)
5. Zhu, H., et al.: Space-efficient optical computing with an integrated chip diffractive
neural network. Nat. Commun. 13(1), 1044 (2022)
6. Song, H.M., Woo, J., Kim, H.K.: In-vehicle network intrusion detection using deep
convolutional neural network. Veh. Commun. 21, 100198 (2020)
7. Zhou, J., et al.: Graph neural networks: a review of methods and applications. AI
open 1, 57–81 (2020)
8. Zhang, H., Xu, M.: Graph neural networks with multiple kernel ensemble attention.
Knowl.-Based Syst. 229, 107299 (2021)
9. Verma, A., Ranga, V.: Statistical analysis of cidds-001 dataset for network intrusion
detection systems using distance-based machine learning. Procedia Comput. Sci.
125, 709–716 (2018)
10. Xu, H., Mueller, F.: Machine learning enhanced real-time intrusion detection using
timing information. In: International Workshop on Trustworthy and Real-time
Edge Computing for Cyber-Physical Systems (2018)
11. Wang, Y., Meng, W., Li, W., Li, J., Liu, W.X., Xiang, Y.: A fog-based privacy-
preserving approach for distributed signature-based intrusion detection. J. Parallel
Distrib. Comput. 122, 26–35 (2018)
12. Xu, H., Fang, C., Cao, Q., Fu, C., Yan, L., Wei, S.: Application of a distance-
weighted knn algorithm improved by moth-flame optimization in network intru-
sion detection. In: 2018 IEEE 4th International Symposium on Wireless Systems
within the International Conferences on Intelligent Data Acquisition and Advanced
Computing Systems (IDAACS-SWS), pp. 166–170. IEEE (2018)
13. Teng, S., Wu, N., Zhu, H., Teng, L., Zhang, W.: Svm-dt-based adaptive and collab-
orative intrusion detection. IEEE/CAA J. Automatica Sinica 5(1), 108–118 (2017)
14. Liu, J., Xu, L.: Improvement of SOM classification algorithm and application effect
analysis in intrusion detection. In: Patnaik, S., Jain, V. (eds.) Recent Developments
in Intelligent Computing, Communication and Devices. AISC, vol. 752, pp. 559–
565. Springer, Singapore (2019). https://doi.org/10.1007/978-981-10-8944-2 65
15. Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional
networks. arXiv preprint arXiv:1609.02907 (2016)
16. Ma, T., Wang, F., Cheng, J., Yu, Y., Chen, X.: A hybrid spectral clustering and
deep neural network ensemble algorithm for intrusion detection in sensor networks.
Sensors 16(10), 1701 (2016)
17. Sun, P., et al.: Dl-ids: Extracting features using CNN-LSTM hybrid network for
intrusion detection system. Secur. Commun. Netw. 2020, 1–11 (2020)
18. Eesa, A.S., Orman, Z., Brifcani, A.M.A.: A novel feature-selection approach based
on the cuttlefish optimization algorithm for intrusion detection systems. Expert
Syst. Appl. 42(5), 2670–2679 (2015)
19. Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT press, Cambridge
(2016)
20. Wang, Z.: The applications of deep learning on traffic identification. BlackHat USA
24(11), 1–10 (2015)
21. Jia, F., Kong, L.: Intrusion detection algorithm based on convolutional neural
network. Trans. Beijing Inst. Technol. 37(12), 1271–1275 (2017)
22. Peng, Y., Choi, B., Xu, J.: Graph learning for combinatorial optimization: a survey
of state-of-the-art. Data Sci. Eng. 6(2), 119–141 (2021)
GNN Based IDS in Smart Vehicles 17
23. Niepert, M., Ahmed, M., Kutzkov, K.: Learning convolutional neural networks for
graphs. In: International Conference on Machine Learning, pp. 2014–2023. PMLR
(2016)
24. Zhang, Y., Liu, Q., Song, L.: Sentence-state LSTM for text representation. arXiv
preprint arXiv:1805.02474 (2018)
25. Yao, L., Mao, C., Luo, Y.: Graph convolutional networks for text classification.
In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp.
7370–7377 (2019)
26. Tai, K.S., Socher, R., Manning, C.D.: Improved semantic representations from
tree-structured long short-term memory networks. arXiv preprint arXiv:1503.00075
(2015)
27. HCRL Dataset. https://ocslab.hksecurity.net/Datasets/car-hacking-dataset
Evaluation of the Energy Viability
of Smart IoT Sensors Using TinyML
for Computer Vision Applications:
A Case Study
1 Introduction
The Internet is a global network that connects millions of computers [1]. Its
technological advancement has brought significant transformations to society,
with impacts on both the public and private sectors, as well as social, political,
and economic contexts [2]. According to the “Digital: Global Overview Report”
study, published in April 2023, the number of active users worldwide reached
5.18 billion, representing approximately 63% of the global population. The report
indicates a significant increase in the number of internet users worldwide over
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 18–33, 2023.
https://doi.org/10.1007/978-3-031-45878-1_2
Evaluation of the Energy Viability of Smart IoT Sensors Using TinyML 19
the past decade, with a doubled growth rate. Based on this trend, it is expected
that this number will continue to grow in the coming years [3]. It is believed that
the growth potential of the internet is moving towards a pervasive computing
ratio [4].
With the advancement of the Internet and the potential for growth of con-
nected objects, the term Internet of Things (IoT) has emerged over the years.
IoT is a network in which “things” or devices embedded with sensors are inter-
connected through private or public networks. These devices can be remotely
controlled to perform specific functionalities, and information can be shared
through the networks using communication protocols. Smart things or devices
can range from simple accessories to large machines, all equipped with some type
of sensor [5].
This technology is increasingly present in everyday life. There are various
application solutions in the development of IoT. For example, in smart homes,
we have the Smart Home, which includes security systems and smart energy
applications. In the context of intelligent transportation, we find fleet tracking as
an example. In the field of smart healthcare, we have the surveillance of patients
with chronic diseases. Finally, we can mention Smart City projects with real-time
monitoring solutions, such as parking availability and smart street lighting” [6].
“A recent example occurred in the city of Christchurch, New Zealand, after an
earthquake in 2011. The city was rebuilt and took the opportunity to implement
a Smart City solution, installing sensors that collect real-time data on various
aspects, from traffic flow to water quality. This provided greater urban efficiency
and productivity, along with relatively low implementation costs for these IoT
devices [7].
These intelligent solutions mentioned above, connecting so many ‘things,’
lead us to a new generation, an evolution of IoT called Internet of Things 2.0
(IoT 2.0), which involves data inference with artificial intelligence in devices,
sensors, and actuators. IoT 2.0 will be a key to the digital transformation of
an internet-connected society, aiming to generate intelligence from devices and
enable real-time data sharing” [8]. “One of the possibilities for the use of Internet
of Things 2.0 is Tiny Machine Learning1 (TinyML). This technology involves the
use of machine learning on low-cost and low-power microcontrollers that have
limitations in memory and storage. In other words, extremely small devices
execute machine learning models” [9]. “TinyML emerges with the purpose of
addressing the challenges of memory and storage, enabling analysis of sensor data
on the device and involving hardware, algorithm, and software considerations
[10].
A recent application developed using TinyML technology is capable of detect-
ing precise alcohol concentrations through temperature, humidity, and alcohol
sensors on a low-power, small-sized microcontroller, resulting in high latency.
This embedded device collects data and implements a machine learning model,
processing the data directly on the device during testing. The study used Google
Colab to build a TinyML model, with the assistance of a micro-library called
1
https://www.tinyml.org/home/.
20 A. M. De Nardi and M. E. Monteiro
TensorFlow Lite. The objective of this study was to improve the accuracy in
alcohol detection, considering the variation in environmental conditions [11].
Over the past few decades, there has been dedicated research effort towards
improving embedded technologies for use in resource-constrained environments
on the internet” [12]. “A successful implementation of IoT requires, among other
desirable requirements, efficient power consumption by embedded devices, which
is particularly important in the IoT application field. With the recent emergence
of TinyML, this technology offers a potential application for IoT, standing out
for its use of machine learning on small-scale embedded devices operating with
limited battery resources and requiring optimization. This study can contribute
to exploring and providing innovative solutions for the use of TinyML in the
development of future applications [13].
The study of energy consumption, among other factors, is of great interest
to the TinyML community, aiming for better development and improvement of
this technology” [14]. “Embedded devices typically consist of a microcontroller
with a processing unit connected to sensors. These sensors, responsible for data
collection, can affect the battery life, especially when it comes to camera sensors.
In this work, a camera sensor will be used, which poses a significant challenge for
embedded systems. While the literature on TinyML demonstrates the possibility
of using this technology in building IoT devices, in my research, I have not
found studies that provide practical support in a real IoT scenario, utilizing
TinyML and exploring its energy behavior. Therefore, this research aims to
test the viability of TinyML as a sustainable device for IoT, analyzing battery
consumption and durability, as well as proposing techniques and methods to
enhance its performance [15].
However, in practice, in a case study and field test research, what would
be the performance of this embedded device with TinyML technology? What
would be the battery consumption of the device when operating in a specific
application? What would be the durability of the device when operating in a
real network? What would be the operational availability profile of the TinyML
device? External environmental factors such as sunlight and rain, would they
hinder its usefulness and application? Would the energy consumption results be
significant?
Currently, researchers and experts are working on enhancing TinyML-based
IoT for devices. One recommended device for this purpose is the microcontroller
(MCU), which offers an increasingly popular option due to its affordability, con-
venience, and portability“ [16]. “This work aims to contribute to the advance-
ment of future studies on TinyML in these intelligent IoT devices, providing
beneficial indicators and identifying areas for improvement. An example is the
evaluation of device durability in a real-world environment.
As mentioned earlier, the integration of machine learning (ML) into IoT
devices aims to reduce the demand for memory, storage, and processing power,
Evaluation of the Energy Viability of Smart IoT Sensors Using TinyML 21
with the goal of reaching a broader audience and offering new options for per-
sonal and professional services and applications. According to the reviewed lit-
erature, the purpose of this study is to conduct a case study of a ready-made
computer vision solution, implementing it on an embedded device using TinyML
technology and applying it in an IoT scenario. Real-time data will be captured
through a camera sensor and stored in the network. Subsequently, an energy
feasibility study will be conducted, analyzing the lithium battery consumption
of the embedded device and investigating its behavior and utilized techniques.
Furthermore, improvements will be proposed and presented through results and
graphs. By the end of this work, it will be possible to evaluate whether the solu-
tion meets the project’s expectations and demands, as well as provide insights
for future research.
According to the study by Khan et al. [5], TinyML has the capability to
run computer vision models. The study highlights research in Machine Learning
(ML) and computer vision across various domains, divided into different fields.
The research fields include Weather Forecasting, Agriculture, Medical Sciences,
Evaluation of the Energy Viability of Smart IoT Sensors Using TinyML 23
Professional Sports, and Facial Recognition, each representing 6%. The fields of
Industries, Professional Sports, and City Traffic Levels have a representation of
12% each, while Biological Sciences and Human Activity are the most represen-
tative fields, each with 19%.
Regarding ML applications in computer vision, it is observed that they have
been successfully used in various areas such as weather forecasting, biological
sciences, expression reading, food security, species classification, sports, traffic
flow monitoring, and predictive maintenance in industries. Some areas, such
as biological sciences, human activity interpretation, traffic management, and
professional sports, are emerging. The study concluded that object detection,
classification, and prediction are the most frequently performed tasks in com-
puter vision with ML. All these areas mentioned in the research can be applied
in the context of the Internet of Things (IoT), emphasizing the importance of
studies on energy efficiency in this field.
The study by Jolly et al. [18] addresses the end-of-life behavior of batteries
in IoT devices. The study utilizes measurements from the battery of the IoT
device Keysight X8712A at different voltage levels. It was observed that the
charge consumption varies as the voltage decreases, providing valuable insights
for optimizing battery life.
Several strategies to extend battery life are presented, such as disabling
non-essential features, reducing sensor measurement frequency, minimizing data
transmission, and alerting users when the battery level is low. These measures
aim to improve customer satisfaction and add value to IoT devices.
Overall, these studies highlight the importance of energy management and
conservation in IoT devices, offering insights and strategies to optimize energy
consumption and extend battery life. This is crucial for the performance and
efficiency of these devices, providing benefits to end-users.
2 Method
To achieve the objectives of this work and assess the energy availability and
behavior of TinyML technology in IoT devices using a computer vision model,
this chapter discusses the materials and methods that will be employed. It begins
by describing the methods, which include field research and the project devel-
opment flow, followed by an explanation of how the TinyML application was
developed. Lastly, the materials and technologies used in the development of
this work will be presented in detail.
and improvements for its study will be proposed in this work. Regarding field
research, guidance is provided on how to conduct field research with market-
oriented research lessons, which can be divided into four activities: choosing
what to study, whom to study with, conducting the research, and elaborating
the results [20]. Based on these four activities, we have chosen what to study and
whom to study with, focusing on the energy efficiency of a pre-trained computer
vision TinyML model. Some changes were made to the code, using appropriate
tools for inference on embedded devices and applying it in a real IoT scenario,
which was tested at the Federal Institute of Espirito Santo. After conducting
the research, the results were elaborated and discussed. Various different tools
are used for experimental research in IoT systems, including the implementation
of the approach itself, software platforms, frameworks, and hardware. Addition-
ally, tools are used to schedule, execute, control, and evaluate the necessary
experiment [21].
Defining the field research methodology for the prototype to run its experi-
ment in the case study, there is a flow to be followed before developing ML to
TinyML. The following flowchart in Fig. 2 addresses each stage of the develop-
ment flow of the chosen ML model until its implementation on an embedded
device. The first stage is “Choose a model,” which will be an existing applica-
tion model for computer vision in the TensorFlow Lite version. The next stage
is “Convert the model,” but this work does not aim to convert the Tensor-
Flow model to the lite version, as the lite version of the model has already been
selected, skipping this part. In the third stage, “Perform inference with the model
on the microcontroller,” a custom framework will be used for this procedure to
bring intelligence to the embedded system. Finally, we have the optional last
stage, “Optimize your model.” In this stage, the model was customized to count
people, aiming to have a computer vision application to verify the battery’s via-
bility. In the optimization development stage, the C++ language was used, and
then the model inference was executed on the microcontroller through a Linux
terminal [22].
IoT. Therefore, a TinyML computer vision model was chosen that detects peo-
ple using the camera sensor. The code was optimized to count the number of
people once the detection is performed, and this technology was inferred in the
embedded system. The entire process, from model selection to inference on the
embedded system, will be detailed. The study of the lithium battery’s perfor-
mance over time is achieved through a voltage divider, and the development of
another controller that retrieves information about the people count and current
battery level, and sends the data over the network to be saved in a non-relational
database.
Data Controller for Network Transmission: For the data controller for
network transmission, the Arduino IDE was used to develop the code. When
the computer vision application detects and counts a person, the output pin on
the ESP32-CAM signals this by activating an LED, indicating that a person
has been detected. This information is then sent to the ESP8266. In addition
to obtaining a people count, this microcontroller also includes a voltage divider
to calculate the level of the lithium battery, with the value obtained from an
26 A. M. De Nardi and M. E. Monteiro
ESP8266: In the proposed project, the NodeMCU is used to capture the battery
level at the time of person detection, total people count, and the runtime of the
application. The NodeMCU is a low-cost microcontroller that runs firmware
on the ESP8266 WiFi SOC, developed by Espressif Systems. This hardware is
an opensource platform and can be worked with IoT due to its features. The
NodeMCU is based on the ESP-12 module and has 16 digital I/O pins and
one analog pin. Additionally, it has Wi-Fi connectivity features and operates
at a voltage of 3.3V. Using the Wi-Fi connection, it sends this data through a
gateway to the network, where the database receives them in real-time, allowing
them to be viewed through a computer or smartphone [28]
3 Case Study
For the assembly of the prototype, in each case study, all the hardware com-
ponents were mounted on a sturdy clipboard material using wires and screws
to provide additional stability. In addition to some parts, hot glue was used
for enhanced protection. A protoboard was used, along with jumpers, to inter-
connect the microcontrollers and the 18650 lithium batteries, which power the
system. Diodes and resistors were also utilized.
To generate the results, we selected the Federal Institute of Espı́rito Santo (IFES)
as the field research location. IFES is an academic center that offers courses rang-
ing from technical programs to doctoral degrees. Currently, it has 22 campuses,
with a total of 1300 professors and 36,000 students. Among these campuses, we
chose the Serra Unit to install the prototype. According to information from the
Evaluation of the Energy Viability of Smart IoT Sensors Using TinyML 29
Second Prototype: In this prototype, three 18650 battery models were used,
connected in series through a battery case, allowing an output of up to 12V. To
power the circuit, a voltage regulator was used to convert the 12V voltage to
5V. Two tests were performed, one with the camera sensor module connected
to detect people, and another without the camera sensor module to compare
battery consumption. A graph was also generated to analyze the consumption
of the three batteries over a six-hour interval, which was the maximum time the
circuit with the camera sensor module was able to support, generating significant
results.
30 A. M. De Nardi and M. E. Monteiro
Although existing literature on TinyML points to its potential use in IoT devices,
no studies were found that prove its energy viability in this field test con-
text. Given this gap, the present study was developed with the aim of evalu-
ating the energy viability of intelligent IoT sensors using TinyML for computer
vision applications in a real-world environment. TinyML technology emerges as
a promising solution to reduce memory, storage, and computational processing
in IoT devices. Additionally, computer vision applications in IoT environments
have significant potential in various areas, as indicated by previous studies in
this work, both in the IoT and computer vision scenarios. Field research con-
ducted at a Federal Institute revealed the voltage expenditures of the prototype
and allowed for the identification of improvements to advance the use of TinyML
technology. To increase battery life in application usage, for future studies, it is
32 A. M. De Nardi and M. E. Monteiro
References
1. Vangie, B.: What is the Internet?—Webopedia (2021). https://www.webopedia.
com/definitions/internet/. Accessed 11 Feb 2021
2. Zimmer, E.F., et al.: A influência da internet na saúde biopsicossocial do adoles-
cente: revisão integrativa. Revista Brasileira de Enfermagem, SciELO Brasil 73(2)
(2020)
3. Datare Portal. Digital 2023: Global Overview Report (2023). https://datareportal.
com. Accessed 01 Jan 2023
4. Umar, F.M., et al.: A review on internet of things (iot). Int. J. Comput. Appl.
113(1), 1–7 (2015)
5. Ahmad, K.M., Khaled, S.: IoT security: review, blockchain solutions, and open
challenges. Future Gener. Comput. Syst. 82, 395–411 (2018)
6. Felix, W., Kristina, F.: Internet of Things. Bus. Inf. Syst. Eng. 57(3), 221–224
(2015)
7. Sarah, M.: Lamppost shines a light on smart cities (2015). https://www.ft.com/
content/53b285c8-851d-11e4-ab4e-00144feabdc0. Accessed 11 Feb 2021
8. Stefano, N., et al.: Iot 2.0 and the internet of transformation (2020)
9. Ji, L., et al.: Mcunet: tiny deep learning on IoT devices. arXiv preprint
arXiv:2007.10319 (2020)
10. de Prado, M., et al.: Robust navigation with tinyml for autonomous mini-vehicles.
arXiv preprint arXiv:2007.00302 (2020)
11. Vitthalrao, L.S., et al.: Alcohol sensor calibration on the edge using tiny machine
learning (tiny-ml) hardware. In: ECS Meeting Abstracts, p. 1848. IOP PUBLISH-
ING (2020)
12. Lachit, D., Swapna, B.: Tinyml meets IoT: a comprehensive survey. Internet Things
16, 100461 (2021)
Evaluation of the Energy Viability of Smart IoT Sensors Using TinyML 33
13. Somayya, M., et al.: Internet of things (IoT): a literature review. J. Comput. Com-
mun. 3(05), 164 (2015)
14. Stanislava, S.: TinyML for Ubiquitous Edge AI (2021)
15. Chandrasekar, V., et al.: Democratization of AI, albeit constrained iot devices &
tiny ml, for creating a sustainable food future. In: IEEE 2020 3rd International Con-
ference on Information and Computer Technologies (ICICT), pp. 525–530 (2020)
16. Hiroshi, D., Roberto, M., Jan, H.: Bringing machine learning to the deepest IoT
edge with tinyml as-a-service (2020)
17. Kalliopi, K., Theodoros, K., An, P.: Towards an iot-enabled intelligent energy
management system. In: IEEE 2017 18th International Symposium on Electro-
magnetic Fields in Mechatronics, Electrical and Electronic Engineering (ISEF)
Book of Abstracts, pp. 1–2 (2017)
18. Brad, J.: The last thing IoT device engineers think about: end of battery life behav-
ior for IoT devices. In: IEEE 2019 IEEE 62nd International Midwest Symposium
on Circuits and Systems (MWSCAS), pp. 837–840 (2019)
19. Colby, R.B., et al.: Benchmarking tinyml systems: challenges and direction. arXiv
preprint arXiv:2003.04821 (2020)
20. Jaworski, B.J., Kohli, A.K.: Conducting field-based, discovery-oriented research:
lessons from our market orientation research experience. AMS Rev. 7, 4–12 (2017)
21. Bin, Z.Y., et al.: Internet of things (IoT): operating system, applications and pro-
tocols design, and validation techniques (2018)
22. TensorFlow. TensorFlow Lite (2022). https://www.tensorflow.org/lite/guide?
hl=pt-br. Accessed 21 Mar 2022
23. Espressif ESP-IDF (2022). https://docs.espressif.com/projects/esp-idf/en/latest/
esp32/get-started/index.html. Accessed 22 Mar 2022
24. Ubuntu, U.: https://ubuntu.com/download. Accessed 21 July 2021
25. Chugh, S.: https://github.com/sunnychugh/tensorflow/. Accessed 01 Aug 2021
26. TensorFlow Blog (2021). https://blog.tensorflow.org/2020/08/announcing-
tensorflow-lite-micro-esp32.html. Accessed 01 Aug 2021
27. EzContents blog (2021). https://ezcontents.org/esp8266-battery-level-meter.
Accessed 01 Aug 2021
28. Abhinab, S., Diwan, R.: IoT based load automation with remote access surveillance
using esp 32 camand esp 8266 module. Ann. Romanian Soc. Cell Biol., 6904–6914
(2021)
29. Espressif ESP-IDF. Disponı́vel em (2022). https://docs.espressif.com/projects/
esp-idf/en/latest/esp32/get-started/index.html. Acesso em 22 Mar 2022
30. Firebase (2022). https://firebase.google.com/docs/database?hl=pt. Accessed 25
Mar 2022
31. Ministério da Educação (2022). https://www.ifes.edu.br/. Accessed 13 May 2022
32. Welcome to Python (2022). https://www.python.org/ Accessed 13 May 2022
33. Google Colab. (2022). https://colab.research.google.com/. Accessed 13 May 2022
34. Xian-Da, Z.: Machine learning. In: A Matrix Algebra Approach to Artificial Intel-
ligence, pp. 223–440. Springer, Heidelberg (2020)
35. Korablyov, D.: https://medium.com/@dmytro.korablyov/first-steps-with-esp32-
and-tensorflow-lite-for-microcontrollers-c2d8e238accf. Accessed 02 May 2022
Simulated Annealing Based Area
Optimization of Multilayer Perceptron
Hardware for IoT Edge Devices
1 Introduction
Deep Neural Networks’ success has been largely attributed to the construction
of highly complex larger neural networks (NNs). The excessive parameterization
of these NNs results in extremely accurate models. This enables these models
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 34–47, 2023.
https://doi.org/10.1007/978-3-031-45878-1_3
SA Based Area Optimization of MLP Hardware for IoT Edge Devices 35
Neural networks mimic the activity of the human brain, enabling AI, machine
learning, and deep learning systems to spot patterns and solve common issues.
These networks, also known as “artificial neural networks” are subfields of
machine learning. These networks have nodes that are interconnected and
arranged in layers. In general, they have three layers: an input layer, single
or multiple hidden layers, and an output layer. The term “Multilayer Percep-
tron” refers to a fully connected multilayer neural network. Figure 1 shows an
MLP with a single hidden layer. Each of these layers computes and passes on
SA Based Area Optimization of MLP Hardware for IoT Edge Devices 37
The nodes, or neurons, are assigned a weight W and a threshold, and after
performing the necessary computation, they are forwarded further [4,13]. The
hidden layers are computationally heavy, and all the neurons are connected to
all the layers in the previous layer and the subsequent layers, to be called fully
connected layers.
x1 h1
y1
x2 h2
y2
x3 h3
solves the compression allocation analytically using weight parameters that have
already been trained. This method avoids the need for manual tuning or iterative
optimization. Unstructured and structured pruning are the two main subcate-
gories of pruning methods. Lee et al. [15] presented an unstructured pruning
approach that prunes the unimportant weights of a NN once during initializa-
tion before training. Similarly, Park et al. [18] presented a simple unstructured
magnitude-based pruning method that extends the single-layer optimization to
a multi-layer optimization. A structured pruning-based general framework for
compressing the NN based on neuron importance levels is presented in [21].
Lin et al. [16] proposed a structured pruning based global and dynamic prun-
ing (GDP) approach to prune unnecessary filters for CNN acceleration. A shift
operation-based spatial information collection strategy as an alternative to spa-
tial convolutions is presented in [20]. In this method, end-to-end trainable shift-
based modules are created by combining shifts and point-wise convolutions. Some
of the works that have been studied regarding optimizing the micro-architecture
and macro-architecture of NN for designing efficient NN model architectures are
presented in [17].
3 Proposed Work
– First, we prepare the training dataset, D, along with the pre-trained single-
layered MLP model with weights, W and biases, B. The pre-trained MLP
model contains its parameters in IEEE-754 single-precision FP32 format.
– The SA algorithm’s various parameters are subsequently initialized. As a
starting point, we choose the random solution along with the starting anneal-
ing temperature Tinit , and the temperature reduction function, α. The start-
ing annealing temperature is kept at 100. As we run the SA algorithm, the
temperature T decays with α, which is given by T = α * T .
– Then, we select the specific percentage, (perturbation amount (p)) of the
hidden layer neuron weights, Wp for the perturbation of all the neuron weights
of the hidden layer, Wh at random, where Wp ⊆ Wh .
– We also specify the number of iterations, N , before running the SA algorithm.
Once the SA algorithm is run, we perturb each neuron’s weight of the hidden
layer at random in each iteration of the training. The Wp is proportional to
the T in our proposed methodology.
– For each iteration, the analysis of the newly generated hidden layer neuron
weights, W is performed. If some of the W are proximate to the integer
value, we round them to the nearest neighbor integer.
40 R. Joshi et al.
Dataset MLP # of
Configurations Parameters
Iris 4-4-3 35
Heart Disease 13-10-2 162
Breast Cancer Wisconsin 30-10-2 332
Credit Card Fraud Detection 29-15-2 482
Fetal Health 21-21-3 528
4 Experimental Results
In this section, we discuss the experimental flow and findings to evaluate our
proposed method. We evaluate our proposed method using five well-known clas-
sification datasets, i.e., Iris [5], Heart Disease [3], Breast Cancer Wisconsin (Diag-
nostic) [19], Credit Card Fraud Detection [2], and Fetal Health Classification [1],
42 R. Joshi et al.
respectively. The experiment flow for this work consists of training the classifi-
cation datasets using contemporary methods by randomly dividing the training
and testing data in an 80:20 ratio to generate the MLP model. In the proposed
work, we train all the datasets using a single hidden layer MLP. Once the MLP
model is generated, we use the same parameters as the pre-trained MLP model
along with the dataset as an input to the custom-modified SA algorithm. After
running the SA algorithm for several iterations, the optimized version of the
MLP model is obtained.
We evaluate our hardware MLP model inference architecture based on an
estimation of the hardware resources utilized by a single unit of IEEE-754 single-
precision FP32 multiplier and adder circuit architecture. The Xilinx Vivado
v.2019.2 tools is used to synthesize the single unit of an IEEE-754 single-
precision FP32 multiplier and adder for the Virtex-7vx485tffg1157-1
R FPGA.
The resource consumption of a single unit of a multiplier and adder is 60 LUTs
and 51 LUTs, respectively. We also evaluate the accuracy of the SA-optimized
MLP model using both training and testing datasets.
The model configuration of single hidden layered MLP for five different clas-
sification datasets is shown in Table 1. The Iris dataset consists of 150 data
instances. The MLP configuration for the Iris dataset [5] consists of 4 input
layer units, 4 hidden layer units, and 3 output layer units. The Heart Disease
dataset [3] consists of 1025 instances, and its MLP configuration is 13 input
layer units, 10 hidden layer units, and 2 output layer units. The Breast Cancer
Wisconsin (Diagnostic) [19] consists of 569 instances and its MLP configuration
is 30 input layer units, 10 hidden layer units, and 2 output layer units. The
Credit Card Fraud Detection dataset [2] consists of 284,807 instances, and its
MLP configuration is 29 input layer units, 15 hidden layer units, and 2 output
layer units. Similarly, the Fetal Health Classification dataset [1] consists of 2,126
instances, and its MLP configuration is 21 input layer units, 21 hidden layer
units, and 3 output layer units.
Table 2. Optimized MLP model configurations for the best case solution with p = 10%
and N = 10,000
Fig. 3. Comparison of LUTs savings (%) vs. N , varying the p between regular model
and SA-optimized model for different datasets.
Fig. 4. Comparison of FFs savings (%) vs. N , varying the p between regular model
and SA-optimized model for different datasets.
and 19.58% (FFs). For the Breast Cancer Wisconsin (Diagnostic), we observe
savings of 20.83% (LUTs) and 20.08% (FFs). For the Credit Card Fraud Detec-
tion dataset, we record savings of 20.65% (LUTs) and 22.21% (FFs). Lastly,
for the Fetal Health dataset, we achieve savings of 23.49% (LUTs) and 23.77%
(FFs). Furthermore, we also observe that the savings in terms of LUTs and
FFs increase with the increase in N in our proposed approach. Although it
takes more time to run the experiment with a higher N , it generates an opti-
mal solution. Figure 5(c) shows the plot of time vs. N , with p = 10%. Table 2
shows the optimized MLP model configurations for all the datasets when p =
10% and N = 10,000 (best case). We also analyze the accuracy of each of the
MLP models generated using our proposed approach. Table 3, 4 and 5 presents
the accuracy comparison between the regular and SA optimized models with
α = 0.95 and p = 10%. We find that the variation in accuracy of the best solution
produced using our proposed methodology is on average −1.64% as compared
to the regular MLP model’s accuracy.
Fig. 5. Best case plots: when p = 10% (a) FFs saving vs. N . (b) LUTs saving vs. N .
(c) Time vs. N .
Table 3. Accuracy comparison between the regular & SA optimized models with α =
0.95, p = 10%, and N = 100.
N = 100
Dataset Accuracy (%) Accuracy (%) Gain/Loss (%)
Regular SA optimized in accuracy
model model
Iris 96.67 95.48 −1.19
Heart Disease 98.83 96.09 −2.74
Breast Cancer 98.70 95.88 −2.82
Wisconsin
Credit Card 96.82 95.98 −0.84
Fraud
Detection
Fetal Health 96.06 94.78 −1.28
Average −1.77
SA Based Area Optimization of MLP Hardware for IoT Edge Devices 45
Table 4. Accuracy comparison between the regular & SA optimized models with α =
0.95, p = 10% N = 1,000.
N = 1,000
Dataset Accuracy (%) Accuracy (%) Gain/Loss (%)
Regular SA optimized in accuracy
model model
Iris 96.67 94.22 −2.45
Heart Disease 98.83 96.11 −2.72
Breast Cancer 98.70 95.55 −3.15
Wisconsin
Credit Card 96.82 95.39 −1.43
Fraud
Detection
Fetal Health 96.06 95.12 −0.94
Average −2.14
Table 5. Accuracy comparison between the regular & SA optimized models with α =
0.95, p = 10% N = 10,000.
N = 10,00
Dataset Accuracy (%) Accuracy (%) Gain/Loss (%)
Regular SA optimized in accuracy
model model
Iris 96.67 95.81 −0.86
Heart Disease 98.83 95.67 −3.16
Breast Cancer 98.70 95.43 −3.27
Wisconsin
Credit Card 96.82 96.14 −0.68
Fraud
Detection
Fetal Health 96.06 95.83 −0.23
Average −1.64
46 R. Joshi et al.
5 Conclusions
In this work, we propose an SA algorithm-based MLP optimization approach to
build a lightweight, energy-efficient, and compressed hardware MLP model. We
finetune the micro-architectural parameters (weights) of the single hidden layer
of MLP to generate the optimized model. The hardware optimization techniques
are further employed to generate the optimum compressed MLP hardware model.
Utilizing five well-known datasets, we conduct comprehensive experiments to
confirm our proposed methodology. Experimental observations demonstrate that
our proposed method produces superior results in terms of hardware resource
(LUTs and FFs) reductions when compared to the regular NN model. For all
the datasets, we find the best optimal solution when the p = 10% and N =
10,000. The purpose of this research work is to validate our proof-of-concept.
In subsequent research work, we will investigate generating highly optimized
compressed NN models with the SA algorithm by varying both the weight and
bias parameters of the NN. We will employ a more complex NN, consisting of a
large number of hidden layers, to produce optimized compressed NN models on
practical hardware platforms.
References
1. Ayres-de Campos, D., Bernardes, J., Garrido, A., Marques-de Sa, J., Pereira-Leite,
L.: Sisporto 2.0: a program for automated analysis of cardiotocograms. J. Maternal-
Fetal Med. 9(5), 311–318 (2000)
2. Dal Pozzolo, A.: Adaptive Machine Learning for Credit Card Fraud detection
(2015)
3. Detrano, R.: UCI Machine Learning Repository: Heart Disease Data Set (2019)
4. Fausett, L.V.: Fundamentals of Neural Networks: Architectures, Algorithms and
Applications. Pearson Education India, Noida (2006)
5. Fisher, R.A.: UCI Machine Learning Repository: Iris Data Set (2011)
6. Gholami, A., Kim, S., Dong, Z., Yao, Z., Mahoney, M.W., Keutzer, K.: A sur-
vey of quantization methods for efficient neural network inference. arXiv preprint
arXiv:2103.13630 (2021)
7. Hu, P., Peng, X., Zhu, H., Aly, M.M.S., Lin, J.: OPQ: compressing deep neural
networks with one-shot pruning-quantization. In: Proceedings of the AAAI Con-
ference on Artificial Intelligence, vol. 35, pp. 7780–7788 (2021)
8. Joshi, R., Zaman, M.A., Katkoori, S.: Novel bit-sliced near-memory computing
based VLSI architecture for fast sobel edge detection in IoT edge devices. In: 2020
IEEE International Symposium on Smart Electronic Systems (iSES)(Formerly
iNiS), pp. 291–296. IEEE (2020)
9. Joshi, R., Kalyanam, L.K., Katkoori, S.: Simulated annealing based integerization
of hidden weights for area-efficient IoT edge intelligence. In: 2022 IEEE Inter-
national Symposium on Smart Electronic Systems (iSES), pp. 427–432 (2022).
https://doi.org/10.1109/iSES54909.2022.00093
10. Joshi, R., Kalyanam, L.K., Katkoori, S.: Area efficient VLSI ASIC implementation
of multilayer perceptrons. In: 2023 International VLSI Symposium on Technology,
Systems and Applications (VLSI-TSA/VLSI-DAT), pp. 1–4. IEEE (2023)
SA Based Area Optimization of MLP Hardware for IoT Edge Devices 47
11. Joshi, R., Zaman, M.A., Katkoori, S.: Fast Sobel edge detection for IoT edge
devices. SN Comput. Sci. 3(4), 302 (2022)
12. Kaiming, H., Xiangyu, Z., Shaoqing, R., Jian, S.: Deep residual learning for image
recognition. In: Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, pp. 770–778 (2016)
13. Kalyanam, L.K., Joshi, R., Katkoori, S.: Range based hardware optimization of
multilayer perceptrons with RELUs. In: 2022 IEEE International Symposium on
Smart Electronic Systems (iSES), pp. 421–426 (2022). https://doi.org/10.1109/
iSES54909.2022.00092
14. Kirkpatrick, S., Gelatt, C.D., Jr., Vecchi, M.P.: Optimization by simulated anneal-
ing. Science 220(4598), 671–680 (1983)
15. Lee, N., Ajanthan, T., Torr, P.H.: Snip: single-shot network pruning based on
connection sensitivity. arXiv preprint arXiv:1810.02340 (2018)
16. Lin, S., Ji, R., Li, Y., Wu, Y., Huang, F., Zhang, B.: Accelerating convolutional
networks via global & dynamic filter pruning. In: IJCAI, vol. 2, p. 8. Stockholm
(2018)
17. Mamalet, F., Garcia, C.: Simplifying ConvNets for fast learning. In: Villa, A.E.P.,
Duch, W., Érdi, P., Masulli, F., Palm, G. (eds.) ICANN 2012. LNCS, vol. 7553,
pp. 58–65. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-33266-
1_8
18. Park, S., Lee, J., Mo, S., Shin, J.: Lookahead: a far-sighted alternative of
magnitude-based pruning. arXiv preprint arXiv:2002.04809 (2020)
19. Wolberg, W., Street, W., Mangasarian, O.: Breast cancer wisconsin (diagnostic).
UCI Machine Learning Repository (1995)
20. Wu, B., et al.: Shift: a zero flop, zero parameter alternative to spatial convolu-
tions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 9127–9135 (2018)
21. Yu, R., et al.: NISP: pruning networks using neuron importance score propaga-
tion. In: Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 9194–9203 (2018)
Machine Learning-Based Multi-stratum
Channel Coordinator for Resilient Internet
of Space Things
School of Science and Engineering, University of Missouri-Kansas City, Kansas City, MO, USA
{mi8rd,songsej,choiby}@umsystem.edu
Abstract. Sensing and transferring data are critical and challenging tasks for
space missions, especially in the presence of extreme environments. Unlike ter-
restrial environments, space poses unprecedented reliability challenges to wire-
less communication channels due to electromagnetic interference and radiation.
The determination of a dependable channel for exchanging critical data in a
highly intemperate environment is crucial for the success of space missions. This
paper proposes a unique Machine Learning (ML)-based multi-stratum channel
coordinator in building the Resilient Internet of Space Things (ResIST). ResIST
channel coordinator accommodates a lightweight software-defined wireless com-
munication topology that allows dynamic selection of the most trustworthy chan-
nel(s) from a set of disparate frequency channels by utilizing ML technologies.
We build a tool that simulates the space communication channel environments
and then evaluate several prediction models to predict the bandwidths across a set
of channels that experience the influence of radiation and interference. The exper-
imental results show that ML-prediction technologies can be used efficiently for
the determination of reliable channel(s) in extreme environments. Our observa-
tions from the heatmap and error analysis on the various ML-based methods show
that Feed-Forward Neural Network (FFNN) drastically outperforms other ML
methods as well as the simple prediction baseline method.
1 Introduction
Sensing is the basis for the monitoring and operation of various space vehicles, satel-
lites, payloads, CubeSats, surface exploration systems, ground testing systems, space
habitats examinations, etc. The necessity lies in a wide range including i) physical sens-
ing for temperature, humidity, and pressure inside or around rocket modules, ii) chemi-
cal sensing of crew capsule or space habitats, mold, mildew, or other airborne bacteria,
iii) crew health monitoring on vital signs and sleep behavior, etc. So, this sensing and
transmitting data reliably and efficiently are integral parts of the National Aeronautics
and Space Administration (NASA)’s recent vision for private astronaut missions, the
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 48–61, 2023.
https://doi.org/10.1007/978-3-031-45878-1_4
Resilient Internet of Space Things (ResIST) 49
For facilitating the benefits of transferring sensor data in space environment through
wireless medium, this paper proposes a reliable wireless communication mechanism
named the Resilient Internet of Space Things (ResIST) for space sensor and data trans-
fer applications. We build a unique Machine Learning (ML)-based multi-stratum chan-
nel coordinator, which employs a software-defined wireless communication approach,
allowing dynamic selection of the most reliable channel(s) from a set of divergent fre-
quency channels for direct transmission. Expanding the traditional Radio Frequency
50 M. T. Islam et al.
ter than the baseline method, EWMA, and two other MLDL-based (i.e., AlexNet and
ConvNet).
The rest of the paper is organized as follows. Section 2 discusses related works in
this subject. We have provided a brief idea of ML and DL methods that have been used
in this work in Sect. 3. Our observation and evaluation results are described in Sect. 4.
Finally, Sect. 5 concludes the work.
2 Related Work
We discuss some notable work related to ML-based wireless communication in this
section as Machine Learning (ML) or Deep Learning (DL) methods have gained signif-
icant attention in terrestrial wireless communication for their data-driven pattern recog-
nition capability in physical layer operations, including channel estimation and detec-
tion, channel encoding and decoding, modulation recognition, etc. [23].
Sumathi et al. represent the channel selection method for cognitive radio networks
in [20], where they utilize the machine learning algorithm the support vector machine -
SVM for best possible channel selection for the secondary user with maximum through-
put using transmitted and received power, the data rate, the maximum vacancy time
and the service time. A deep learning-based partially overlapped channel assignment
(POCA) method for wireless SDN-IoT system is proposed in [22]. Bitar et al. in [5]
mention the identification method of sensing wireless ISM band in a heterogeneous
environment with a deep convolutional neural network with improved accuracy. Tang
et al. in [21] propose a deep learning-based prediction algorithm to forecast future
traffic load and congestion in a network with a partial channel assignment algorithm
and intelligently allocate channels to each link in the SDN-IoT network. Authors in
the article [14, 26] proposed a dynamic resource allocation and multiple access control
(MAC) protocol method for a heterogeneous network with deep reinforcement learn-
ing. As for wireless resource management for space communications. Authors in [25]
propose a deep learning technique called long short-term memory (LSTM) network for
anticipating the spectrum accessibility of cognitive aerospace communication networks
without prior knowledge of the user activities. Ferreira et al. in [9] propose a hybrid
algorithm for radio resource management for cognitive space communication network
where they integrate multi-objective reinforcement learning, and a deep neural network,
which significantly reduces the cost of spending time and resources on learning action-
performance mapping. Kato et al. in [13] proposed an ML-based solution to address
different problems on a Space-Air-Ground Integrated Network (GIN) and evaluated a
deep learning-based method to improve traffic control performance. Authors in [24]
presented the performance of deep learning model for the channel estimation and sig-
nal detection in orthogonal frequency-division multiplexing (OFDM) systems, which
can perform better for addressing channel distortion and identifying the transmitted
symbols rather than conventional methods.
There have been several studies on channel selections for wireless cellular net-
work related. Kato et al. in [13] proposed an ML-based solution to address differ-
ent problems on a Space-Air-Ground Integrated Network (GIN) and evaluated a deep
52 M. T. Islam et al.
learning-based method to improve traffic control performance. Studies in [11, 16] con-
sider resource handover issues for multi Radio Access Technology (multi-RAT) net-
works, traffic flow prediction, and enhance the results of network planning tools by
utilizing various deep learning methods and their results demonstrate that deep learning
can be promising aspects on those aspects. Authors in [12] proposed a deep learning-
based channel estimation technique where they came about a channel learning scheme
using deep auto-encoder, which learned the channel state information (CSI) and min-
imize the mean square error (MSE) of the channel estimation for wireless energy
transfer. Authors in [19] formulate channel rank measurement (CRM) metric data-
set for normal-equation-based supervised machine learning algorithm (NEC algorithm)
which performs channel rank estimation (CRE) of any channel based on average packet
receive and instantaneous values of RSSI. Bai et al. in [4] propose a channel charac-
teristics prediction mechanism based on machine learning (ML) algorithm and convo-
lutional neural network (CNN) for three-dimensional (3D) millimeter wave (mmWave)
massive multiple inputs multiple outputs (MIMO) indoor channels. Authors in [15, 27]
proposed a learning-based channel estate information (CSI) prediction method for 5G
wireless communication where they pointed out several important features like fre-
quency band, location, time, humidity, weather, and temperature, etc. that affect the CSI
and after that, they designed a learning mechanism with the integration of CNN (convo-
lutional neural network). Huang et al. in [10] proposed a big data and machine learning-
enabled wireless channel model framework based on artificial neural networks (ANNs),
including feed-forward neural network (FNN) and radial basis function neural network
(RBF-NN). They leveraged the input parameters like transmitter (Tx) and receiver (Rx)
coordinates, Tx-Rx distance, and carrier frequency, while the output parameters are
channel statistical properties, including the received power, root mean square (RMS)
delay spread (DS), and RMS angle spreads (ASs).
Our work differs from the above-mentioned work as we have focused on generat-
ing channel conditions with electromagnetic interference similar to the space or harsh
environment. We have leveraged ML techniques to predict reliable and high band-
width channels among the multi-dimensional and heterogeneous channel conditions
with a software-defined approach. Our model is not dependent on any specific proto-
col, medium, or channel, rather it can predict or choose the most effective and feasible
one through available bandwidth prediction of the channel. By focusing on the physical
channel condition rather than protocol or technology, our model ensures the reliability
and effectiveness of the channel selection/prediction process in extreme environmental
conditions dynamically.
algorithm, FFNN, AlexNet, and CNN, for supporting the coordination-based dynamic
selection method’s feasibility in space communication. ML-SDC’s coordination meth-
ods coexist in the middleware as running models and can be picked dynamically based
on performance. We will first go with the prediction approach of EWMA and then go
through the mathematical and parametric definitions of different layers and their control
function. Table 1 summarizes the various notations used in the mathematical formula-
tions of this paper.
Symbols Descriptions
St Newly observed sample value
α Sample Weight parameter
w Weight matrix
x Input activity
a Output activity
f Activation function
Ek A simple loss function
s The number of pixels moved upon each application
p the number of zeroes used to pad the borders
m, n, u, v, g, h Dimension of Matrices
γ, β Additional parameters for normalized input
ML-SDC adopts each ML-based model, including FFNN, AlexNet, and CNN (or
ConvNet) using different numbers, combinations, and layers in the form of objects
imported from Keras framework [1]. Each layer of neural network have a specific math-
ematical structure, the corresponding gradient to train the parameters, the theoretical
advantages and pitfalls. The layer types are dense or fully-connected, two-dimensional
convolution, max pooling, and batch normalization. We do not alter bias terms for the
benefits of simplicity.
A dense or fully-connected layer refers to a layer consisting of a weight matrix
w ∈ Rm×n , which is multiplied by an input vector x ∈ Rm before a nonlinear activation
function f is applied to yield an output a ∈ Rn . Multidimensional input is flattened by a
concatenating operation, which is a straightforward operation. The activation function
used for each fully-connected layer in all neural networks tested was the widely-used
Rectified Linear Unit (ReLU), defined as fReLU (z) = max(0, z). If we let w(i) denote the
54 M. T. Islam et al.
(i)
weight matrix of the ith dense layer, then we may write the activity ak of the kth neuron
of this layer as presented in Eq. (2).
m
(i) (i) (i) (i) (i)
ak = fReLU (zk ), zk = ∑ w jk x j (2)
j
For demonstration purposes, if we apply the chain rule for partial derivatives and
(l) (l)
assume a simple loss function of Ek = 0.5(ak − tk )2 , where ak is the activity of the
kth neuron in the last (l th ) layer and tk is the corresponding target value. We can obtain
the gradients for this layer as shown in Eq. (3).
(l) (l)
(l) (l) (l)
∂ Ek ∂ Ek ∂ ak ∂ zk (ak − tk )x j if zk > 0,
(l)
= (l) (l) (l)
= (l) (3)
∂ w jk ∂ ak ∂ zk ∂ w jk 0 if zk ≤ 0
A max-pooling layer refers to a layer with a pool size (u, v) specifying the dimen-
sions of a window over which the max function is applied to extract the most relevant
features. The activity of the entry in the jth row and kth column of the output feature
map (a jk ) resulting from the application of the max function over a pooling window at
index (l, m) of the two-dimensional input feature map x ∈ Rg×h is shown in Eq. (5).
a j,k = max xl,m , xl,m+1 , xl+1,m , . . . , xl+u−1,m+v−1 (5)
Both convolution and max-pooling layers have variable stride (s, the number of
pixels moved upon each application of a filter) and zero-padding (p, the number of
zeroes used to pad the borders of the image before filters are applied). They determine
the size and shape of the output layer depending on the input’s size and shape. Given
an input feature map x ∈ Rg×h , a kernel / pool size (u, v), we define the size (q, r) of the
resulting output feature map y ∈ Rq×r as shown in Eq. (6).
g − u + ple f t + pright h − v + ptop + pbottom
q= + 1, r= +1 (6)
shorizontal svertical
Resilient Internet of Space Things (ResIST) 55
We use a batch normalization layer, batch norm, to fix each scalar input’s means
and variances to the next layer independently via a normalization step across a given
mini-batch of training inputs. We accelerate and regularize the training of deep neural
networks by reducing internal covariate shift. During training, the batch norm step is
immediately followed by scaling and shifting operation with two additional parameters
γ and β for each normalized input to approximate the mean and variance, respectively.
FFNN can be distinguished from other neural networks like AlexNet and ConvNet,
as there is no cycle between the nodes. Values are straightly propagated forward through
a set of layers. FFNN’s input is 20-time steps over 200 channels, and it results in a single
prediction for each channel’s next bandwidth measurement. FFNN has three layers,
including output, with 4000, 500, and 200 neurons in order. There is a batch norm layer
between the second and third layers. Each layer uses a ReLU as its activation and zeros
as its bias initializer. ReLU, defined as fReLU (z) = max(0, z), is useful as it only gives
non-zero outputs if the input is greater than zero. Each layer’s weights use a He Normal
(He-et-al) initializer [3], which draws samples from a normal distribution centered on
zero. FFNN has trained three epochs at a time.
4 Evaluations
ML-based models, the FFNN in Fig. 3c, exhibits that the output signal pattern highly
matches the fundamental input signal level. AlexNet in Fig. 3d and ConvNet in Fig. 3e
show better matching patterns than the EWMA model but less accurate results than
FFNN with revealing different detection capabilities. For example, ConvNet can accu-
rately predict the signal level, whereas AlexNet can detect the edge of some interference
bursts but shows significant errors on the actual signal level.
Third, we conduct a Mean Absolute Error (MAE) calculation using the same gen-
erated test set. We calculate MAE by summing the absolute value of the difference
between each channel’s predicted bandwidth and the same channel’s actual bandwidth
at each point in time. The sum is then divided by the total number of predictions, which
can be calculated by the number of channels multiplied by the number of time-steps in
the test set. The models with a lower MAE are considered to have better performance.
Resilient Internet of Space Things (ResIST) 57
As summarized in Table 4, the simulation results show that FFNN has the least MAE
(3.12) followed by ConvNet’s MAE (5.08), which is much better than other approaches.
The MAE results align with the previous heatmap representations.
Table 4. Model performance for 2K consecutive time steps across 200 channels(20K training
examples, where applicable)
Forth, we also present each model’s prediction performance in MAE using Probabil-
ity Distribution Function (PDF) and Cumulative Distribution Function (CDF). The PDF
is generated by sampling each model’s predictions’ absolute error floor (integer-valued)
across the entire test set. Then, it counts the frequency of each integer value in each
model’s sampled errors. These results are L1-normalized (by dividing the frequency of
each integer-valued error by the total number of predictions) to produce proper proba-
bility distributions representing each error’s relative frequency. When a PDF of a model
concentrates on the lower error probability, it performs better than a model with a PDF
where most observations are either distributed evenly or focused away from zero error.
We also generate the CDF by iterating through the PDF (sorted from zero to max abso-
lute error) and replacing each relative frequency with the sum of relative frequencies.
The CDF visualizes and calculates the probability that the absolute error falls within
a specific range of values. A model with a CDF with a high positive slope towards
zero error and a vanishing slope further away from zero error is the better-performing
model. Additionally, if the slopes are compared between the two models, the higher
CDF values model is better. We have trained the prediction models with an extensive
training dataset (20K independent and identically distributed training examples) and
ran many sets of tests and validations. As shown in Fig. 4a, the MAE results show that
FFNN, followed by the ConvNet, performs 3 to 5 times better than other models and
the error is half less than the EWMA (with alpha = 0.125). Our test results show that
AlexNet performs the worst. Additionally, As presented in Fig. 4b, the CDF of both the
ConvNet and FFNN had the steepest slopes near zero, but FFNN has the higher values
overall. One significant difference between ConvNet and FFNN is that the latter had
a much smaller maximum absolute error than the former. It is consistently generated
predictions with a calculated MAE, which is larger than the EWMA by about 45%, and
over four times higher MAE than FFNN. Overall, our experimental results show that
AlexNet is the worst performer according to the PDF and CDF analysis.
Resilient Internet of Space Things (ResIST) 59
5 Conclusion
We have introduced the Resilient Internet of Space Things (ResIST) by building a
Machine Learning (ML)-based software-defined coordinator (ML-SDC), which accom-
modates a dynamic selection of the most reliable channel(s) from a set of divergent
frequency channels. We have conducted a feasibility study of using ML technologies
for wireless space communication in extreme or unreliable environments. We have
explored several channel prediction models (such as EWMA, FFNN, AlexNet, and Con-
vNet) and showed a comparative study for the effectiveness of those models. The exper-
imental results show that ML-prediction technologies can improve wireless communi-
60 M. T. Islam et al.
References
1. Keras Framework. https://keras.io/
2. Recap From the Expedition Lead Scientist. https://www.nasa.gov/mission_pages/station/
research
3. Xavier and He Normal (He-et-al) Initialization. https://bit.ly/3qkeyvv
4. Bai, L., et al.: Predicting wireless mmWave massive MIMO channel characteristics using
machine learning algorithms. Wirel. Commun. Mob. Comput. 2018 (2018)
5. Bitar, N., Muhammad, S., Refai, H.H.: Wireless technology identification using deep convo-
lutional neural networks. In: 2017 IEEE 28th Annual International Symposium on Personal,
Indoor, and Mobile Radio Communications (PIMRC), pp. 1–6. IEEE (2017)
6. Choi, B.Y., Boyd, D., Wilkerson, D.: Toward reliable and energy efficient wireless sensing
for space and extreme environments. In: 2017 IEEE International Conference on Wireless
for Space and Extreme Environments (WiSEE), pp. 156–157. IEEE (2017)
7. De Cola, T., Marchese, M.: Reliable data delivery over deep space networks: benefits of long
erasure codes over ARQ strategies. IEEE Wirel. Commun. 17(2), 57–65 (2010)
8. Drobczyk, M., Lübken, A.: Novel wireless protocol architecture for intra-spacecraft wireless
sensor networks (inspawsn). In: 2018 6th IEEE International Conference on Wireless for
Space and Extreme Environments (WiSEE), pp. 89–94. IEEE (2018)
9. Ferreira, P.V.R., Paffenroth, R., Wyglinski, A.M., Hackett, T.M., Bilén, S.G., Reinhart, R.C.,
Mortensen, D.J.: Multi-objective reinforcement learning-based deep neural networks for
cognitive space communications. In: 2017 Cognitive Communications for Aerospace Appli-
cations Workshop (CCAA), pp. 1–8. IEEE (2017)
10. Huang, J., et al.: A big data enabled channel model for 5G wireless communication systems.
IEEE Trans. Big Data (2018)
11. Huang, W., Song, G., Hong, H., Xie, K.: Deep architecture for traffic flow prediction: deep
belief networks with multitask learning. IEEE Trans. Intell. Transp. Syst. 15(5), 2191–2201
(2014)
12. Kang, J.M., Chun, C.J., Kim, I.M.: Deep-learning-based channel estimation for wireless
energy transfer. IEEE Commun. Lett. 22(11), 2310–2313 (2018)
13. Kato, N., et al.: Optimizing space-air-ground integrated networks by artificial intelligence.
IEEE Wirel. Commun. 26(4), 140–147 (2019)
Resilient Internet of Space Things (ResIST) 61
14. Li, J., Zhang, X., Zhang, J., Wu, J., Sun, Q., Xie, Y.: Deep reinforcement learning-based
mobility-aware robust proactive resource allocation in heterogeneous networks. IEEE Trans.
Cogn. Commun. Network. 6(1), 408–421 (2019)
15. Luo, C., Ji, J., Wang, Q., Chen, X., Li, P.: Channel state information prediction for 5G wire-
less communications: a deep learning approach. IEEE Trans. Netw. Sci. Eng. (2018)
16. Mroue, M., Prevotct, J.C., Nouvel, F., Mohanna, Y., et al.: A neural network based handover
for multi-rat heterogeneous networks with learning agent. In: 2018 13th International Sym-
posium on Reconfigurable Communication-Centric Systems-on-Chip (ReCoSoC), pp. 1–6.
IEEE (2018)
17. Northon, K.: Nasa opens international space station to new commercial opportunities, private
astronauts, June 2020. https://go.nasa.gov/2CUqekw
18. Pelissou, P.: Building blocks for an intra-spacecraft wireless communication. In: 2015 IEEE
International Conference on Wireless for Space and Extreme Environments (WiSEE), pp.
1–6. IEEE (2015)
19. Rehan, W., Fischer, S., Rehan, M.: Machine-learning based channel quality and stability esti-
mation for stream-based multichannel wireless sensor networks. Sensors 16(9), 1476 (2016)
20. Sumathi, D., Manivannan, S.: Machine learning-based algorithm for channel selection utiliz-
ing preemptive resume priority in cognitive radio networks validated by ns-2. Circuits Syst.
Sig. Process. 39(2), 1038–1058 (2020)
21. Tang, F., Fadlullah, Z.M., Mao, B., Kato, N.: An intelligent traffic load prediction-based
adaptive channel assignment algorithm in SDN-IoT: a deep learning approach. IEEE Internet
Things J. 5(6), 5141–5154 (2018)
22. Tang, F., Mao, B., Fadlullah, Z.M., Kato, N.: On a novel deep-learning-based intelligent
partially overlapping channel assignment in SDN-IoT. IEEE Commun. Mag. 56(9), 80–86
(2018)
23. Wang, T., Wen, C.K., Wang, H., Gao, F., Jiang, T., Jin, S.: Deep learning for wireless physical
layer: opportunities and challenges. China Commun. 14(11), 92–111 (2017)
24. Ye, H., Li, G.Y., Juang, B.H.: Power of deep learning for channel estimation and signal
detection in OFDM systems. IEEE Wirel. Commun. Lett. 7(1), 114–117 (2017)
25. Yu, L., Wang, Q., Guo, Y., Li, P.: Spectrum availability prediction in cognitive aerospace
communications: a deep learning perspective. In: 2017 Cognitive Communications for
Aerospace Applications Workshop (CCAA), pp. 1–4. IEEE (2017)
26. Yu, Y., Wang, T., Liew, S.C.: Deep-reinforcement learning multiple access for heterogeneous
wireless networks. IEEE J. Sel. Areas Commun. 37(6), 1277–1290 (2019)
27. Yuan, J., Ngo, H.Q., Matthaiou, M.: Machine learning-based channel prediction in massive
MIMO with channel aging. IEEE Trans. Wirel. Commun. 19(5), 2960–2973 (2020)
A Schedule of Duties in the Cloud Space Using
a Modified Salp Swarm Algorithm
Hossein Jamali, Ponkoj Chandra Shill, David Feil-Seifer, Frederick C. Harris Jr.,
and Sergiu M. Dascalu(B)
Department of Computer Science and Engineering, University of Nevada, Reno, Reno, NV, USA
{hossein.jamali,ponkoj}@nevada.unr.edu, {dave,fred.harris,
dascalus}@cse.unr.edu
1 Introduction
Today, modern computing methods have attracted the attention of researchers in many
fields such as cloud computing, artificial intelligence, and machine learning by using
techniques including artificial neural networks in building air quality prediction models
that can estimate the impact of climate change on future summer trends [1]. A compu-
tational science algorithm is used in this article to determine the schedule of duties in
the cloud.
Cloud computing has brought about the availability of tools that provide extensive
computing resources on the internet platform. Users can submit their requests for various
resources, such as CPU, memory, disk, and applications, to the cloud provider. The
© IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 62–75, 2023.
https://doi.org/10.1007/978-3-031-45878-1_5
A Schedule of Duties in the Cloud Space 63
provider then offers the most suitable resources, which meet the user’s requirements and
offer benefits to the resource owners, based on the price that they can afford to pay [2].
In cloud computing, the main entities are users, resource providers, and a scheduling
system whose main body has been proposed for the users’ tasks and timeline strategy
[3].
Cloud computing consumers rent infrastructure from third-party providers instead of
owning it. They opt for this to avoid extra costs. Providers typically use a “pay-as-you-
go” model, allowing customers to meet short-term needs without long-term contracts,
thus reducing costs [4].
Behind the numerous benefits of cloud computing, there are many challenges too.
The most important is the task scheduling problem or resource allocation to the users’
requests. The targets of task scheduling in cloud computing are to provide operating
power, the optimal timeline for users, and service quality simultaneously. The specific
targets related to scheduling are load balance, service quality, economic principles, the
best execution time, and the operating power of the system [5]. Cloud computing has
three timelines: resources, workflow, and tasks. Resource scheduling involves mapping
virtual resources to physical machines. Workflow scheduling ensures the orderly flow
of work. Task scheduling assigns tasks to virtual resources. Task scheduling methods
can be concentrated or distributed, homogeneous or heterogeneous, and performed on
dependent or independent tasks.
Task scheduling in cloud computing has two types based on the characteristic of the
tasks:
• Static: In static scheduling, the tasks reach the processor simultaneously and are
scheduled on accessible resources. The scheduling decisions are made before reaching
the tasks and the processing time after doing the entire run of duty is updated. This
type of scheduling is mostly employed for tasks that are sent continuously [6]; and
• Dynamic: In dynamic scheduling, the number of tasks, the location of the virtual
machines, and the method for resource allocation are not constant, and the input time
of tasks before sending them is unknown [6].
Scheduling the mechanism of dynamic algorithms compared to static algorithms
is better but the overhead of the dynamic algorithm is quite significant [7]. Dynamic
scheduling can be done in two ways; in batch and online modes. In batch mode, the
tasks are lying in a line, gathered in a set, and after a certain time, scheduled. In the
online mode, when the tasks reach the system, they are scheduled [6].
The task scheduling problem in cloud computing focuses on efficiently distribut-
ing tasks among machines to minimize completion time [8]. Proper task arrangement
has numerous benefits, including reduced energy consumption, increased productiv-
ity, improved distribution across machines, shorter task waiting times, decreased delay
penalties, and overall faster task completion [9].
The task scheduler plays a crucial role in efficiently scheduling computing actions
and logically allocating computing resources in IaaS cloud computing. Its objective is to
assign tasks to the most suitable resources to achieve specific goals. Selecting an appro-
priate scheduling algorithm is essential to enhance resource productivity while maintain-
ing a high quality of service (QoS). Task scheduling involves optimizing the allocation
64 H. Jamali et al.
of subtasks to virtual servers in order to accomplish the task schedule’s objective. This
area of research continues to receive significant attention [10].
Efficient task planning in cloud computing is essential to minimize fetch time, wait-
ing time, computing time, and resource usage. Task scheduling is crucial for maxi-
mizing cloud productivity, meeting user needs, and enhancing overall performance. Its
primary goal is to manage and prioritize tasks, reducing time and preventing work fail-
ures while meeting deadlines. Task scheduling optimizes the cloud computing system
for improved calculation benefits, high performance, and optimal machine output. The
scheduling algorithm distributes work among processors to maximize efficiency and
minimize workflow time [11].
The rest of this paper is organized as follows: Sect. 2 covers related work; Sect. 3 pro-
vides details of the SDSA optimization algorithm; Sect. 4 describes our proposed method,
including the expansion and improvement of the salp algorithm; Sect. 5 focuses on the
algorithm’s target, the fitness function; Sect. 6 presents the results of our simulation;
and, finally, Sect. 7 contains the conclusions of our work.
2 Related Works
Ghazipour et al. [12] have proposed a task scheduling algorithm so the tasks existing
in the grid are allocated to accessible resources. This algorithm is based on the ACO
algorithm, which is mixed with the scheduling algorithm right to choose so that its results
are used in the ACO algorithm. The main goal of their article is to minimize the total
finish time (makespan) for setting up tasks that have been given [12].
In their research on task scheduling in cloud computing, Sharma and Tyagi [13]
examined nine heuristic algorithms. They conducted comparative analyses based on
scheduling parameters, simulation tools, observation domain, and limitations. The results
indicated the existence of a heuristic approach that satisfies all the required parameters.
However, considering specific parameters such as waiting time, resource utilization, or
makespan for each task or workflow individually can lead to improved performance.
[13].
In 2019, Mapetu et al. [14] researched the “binary PSO algorithm for scheduling
the tasks and load power in cloud computing”. They introduced a binary version of the
PSO algorithm named BPSO with lower complexity and cost for scheduling the tasks
and load power in cloud computing, to minimize waiting time, and imbalance degree
while minimizing resource use. The results showed that the proposed algorithm presents
greater task scheduling and load power than existing heuristic algorithms [14].
Saeedi et al. [15] studied the development of the multi-target model of PSO for
scheduling the workflow in the cloud areas. They proposed an approach for solving
the scheduling problem considering four contrasting goals (i.e., minimizing the cost,
waiting time, energy consumption, and maximizing reliability). The results showed
that the proposed approach had a better performance compared to LEAF and EMS-C
algorithms [15].
Zubair et al. [10] presented an optimal task scheduling method using the modified
symbiotic organisms search algorithm (G_SOS) and aimed to minimize the makespan
of the tasks, costs, response time, and imbalance degree, and improve the convergence
A Schedule of Duties in the Cloud Space 65
speed. The performance of the proposed method using CloudSim (a simulator tool) was
evaluated and according to the simulation results, the proposed technique has better
performance than the SOS and PSO-Simulated Annealing (PSO-SA) in terms of the
convergence speed, cost, imbalance degree, response time, and makespan. The findings
confirm the suggested G_SOS approach [10].
Rajagopalan et al. [16] introduced an optimal task-scheduling method that combines
the firefly optimization algorithm with a genetics-based evolutionary algorithm. This
hybrid algorithm creates a powerful collective intelligence search algorithm. The pro-
posed method excels in minimizing the makespan for all tasks and quickly converges
to near-optimal solutions. The results demonstrated that this hybrid algorithm outper-
formed traditional algorithms like First in, First Out (FIFO) and genetics. However, a
potential drawback of this method is the increased overload resulting from the sequential
use of two algorithms [16].
This section briefly describes the SSA optimization algorithm proposed by Mirjalini Al
which is an extension of the standard SSA algorithm [17]. salps are a type of Salpidae
family and have a transparent and barrel-shaped body. Their bodies are very similar to
jellyfish. They still move the same as jellyfish, and water is pumped from the middle of
the body as a motive force to move forward [17]. The shape of salp is shown in Fig. 1(a).
The biological study of these animals is just starting because it is so difficult to
capture them and maintain them in laboratories. One of the most intriguing habits of
salps is their tendency to swarm. The salps commonly form a chain in the deep oceans.
This chain is shown in Fig. 1(b). Although the primary cause of this behavior is unknown,
some researchers think that it is carried out through quick coordinated movements and
searches to achieve improved movement [17].
To model mathematically the salp chains, first, the population is divided into two
groups: leaders and followers. The leader is in front of the chain, while the remaining
66 H. Jamali et al.
are considered the followers. As seen from their names, the leader guides the group and
the followers follow each other [17].
Like other techniques based on the swarm, the location of salps in a search space is
n-dimensional, where n is the number of variables in a problem and known; therefore,
the location of all salps is stored in the two-dimensional matrix x. Also, it is assumed
that a food source, F, exists in the search space as a swarm target [17].
Equation 1 has been proposed for updating the location of the leader as follows:
Fj + c1 ubj − lbj c2 + lbj c3 ≥ 0
xj1 = (1)
Fj − c1 ubj − lbj c2 + lbj c3 < 0
where xj1 shows the location of the first salp (leader) in the j dimension, F j is the location
of the food source in the j dimension, ubj identifies the upper boundary of the j dimension,
lbj identifies the lower boundary of the j dimension, and c1 , c2 and c3 are random numbers
(between 0,1) [17].
Equation 1 shows that the leader just updates its location according to the food
source. The c1 the constant is the most important parameter in the SSA because it
creates a balance between exploration and detection and is defined as Eq. 2:
2
− 4l
c1 = 2e L
(2)
The pseudo-code for the SSA algorithm is shown in Fig. 2 [17]. This figure shows
that the SSA algorithm begins the global optimum by starting several salps at random
locations. Then, each fitting related to the salps are calculated and the location where
they have acquired the best fitting is allocated to the variable F as a food source followed
by the salp chain. Meanwhile, the value of c1 constant is updated by Eq. 2. For every
dimension, the location of the leader is updated by relation 1 and that of the followers
by Eq. 4. If each salp goes out of the search space, they are returned to the border again.
All the mentioned stages except for the initial value are iterated till consent is obtained.
The computing complexity of the SSA algorithm is considered
as O(t(d ∗ n + Cof ∗ n)) where t shows the number of iterations, d is that of variables
(dimension), n is that of solutions, and Cof is a target cost of the function.
4 Proposed Method
Our proposed method for scheduling the tasks of the virtual machines in the cloud
computing area uses an optimized SSA based on the fitness function. First, a set of
random answers created is assigned as the initial population. Each member of this set
is called a salp. In the first stage, the fitness of salps produced randomly is calculated
by the target function and the best slap is chosen among all salps and its location is
determined by the location of the food source. In the following, the salps move towards
the food source until they achieve the best food source (i.e., solution). In this algorithm,
each salp is represented as a solution that moves for searching based on a mechanism in
the problem space. In the suggested method, the salps are divided into two groups, the
leaders and the followers. One group of salps named leader salps updates its location
according to the food source and tries to move towards the existing food source and
discover a better solution. If they find a better solution than the existing food source, the
location of the leader salp is considered as its new location. The group salps follow each
other, and if they discover a better solution for the food source location, the location of
the salp follower is considered the new location of the food source.
68 H. Jamali et al.
The task scheduling problem in the cloud is allocating the settings of tasks to a set of
sources. We have assumed a set of n tasks, T = (T1 .T2 .T3 . · · · .Tn ), and of m sources,
which are virtual machines in targeted source research, V = (V1 .V2 .V3 . · · · .Vm ). The
set of T includes the tasks which should be scheduled. Each task should be processed
by virtual machines so that the completion time of all tasks is minimized as much as
possible.
The main goal of task scheduling is to allocate optimally to the sources so that the
lowest completion time of the tasks (i.e., makespan) and the lowest cost is obtained. The
makespan shows the total required time for implementing all the tasks. The main goal
of our research is to minimize the makespan using the modified SSA.
Assume that an array of 200 tasks exists and each task has a value between 1–15. For
example, if the second value of this array is 4, it shows that task 2 has been allocated to
the virtual machine 4 and, if the seventh value of the array is 14, it means that, task 7 has
been allocated to the virtual machine 14. Similarly, all the tasks T1 to T200 are allocated
to virtual machines V1 –V15 . In Fig. 3, an example of allocating tasks to virtual machines
is depicted.
In the suggested algorithm, solutions are shown by a salp chains. Each solution of
the suggested algorithm is shown by an array of natural numbers. The locations of all
salps are stored in a 2-dimensional matrix named x. For instance, in a problem with n
tasks and m virtual machines, the rows of a two-dimension matrix are considered as the
number of the salp population. It means that the location of each salp is restored in a
row of a matrix. The columns of the matrix are equal to n. Also, the content of each cell
of the array shows the virtual machine number, which can be a number between 1 to m.
Figure 4 shows an example of a salp.
To begin the work, this salp can be produced as a desired number where this number is
the same as the primary population of the algorithm that is adjusted. First, the population
A Schedule of Duties in the Cloud Space 69
is randomly generated and stored in a two-dimensional matrix where its rows are identical
to the number of salps and its columns equal to those of tasks identified for the scheduling.
After generating the primary population of salps in the range of the problem answer,
the fitness of all salps is assessed by all salps and the salp with the best fitness is
determined. In this algorithm, it is assumed that a food source named F exits in a search
space as a swarm target that all salps try to move towards it.
In the first stage of this algorithm, the location of the best generated salp (the best
solution) is considered as the food source.
In the next stage of this algorithm, the salps are divided into two groups of leaders and
followers. The number of salps is considered as the leader salp group and the remaining
as the follower one. In the proposed algorithm, 50% of salps are considered as the leader
group and the remaining 50% as followers. The location of the leader group is updated
by Eq. 5.
where xji is the location of the leader salp, i, Fj the location of the food source in the j
dimension, α the constant of the random moving step in the range of [0, 1] that is adjusted
by the targeted problem, and Randn( ) a random number with a normal distribution and
determines a random step with a normal distribution for the leader group. Equation 6
updates the location of the follower group.
1 i
xji = xj + xji−1 + c1 Randn( ) (6)
2
where xji is the location of the follower salp i in the j dimension. The constant c1 creates
a balance between the exploration and discovery by generating an adaptive step, and this
constant decrease consistently during the iterations; so, it leads to higher discovery in
the first iterations and higher exploration in the end iterations if the algorithm, Randn( )
is a random number with a normal distribution and determines a random step with this
distribution for the leader group. The parameter c1 is defined in Eq. 7 and is updated in
each iteration.
2
− 4l
c1 = 2e L
(7)
diversifying a search space, releasing from the local optimum, as well as discovering
accurately surrounding the existing food source.
Based on this, the population of the leader group updated its location using the
location of the food source. When the location of each leader salp group is substituted
for that of the food source, the latter group has updated its location using the new location
of the food source. Figure 5 depicts the algorithm’s pseudo-code of the optimized SSA.
The stages of the algorithm until reaching the end are continued. In the proposed
algorithm, the condition for finishing the algorithm is the number of iterations.
5 Fitness Function
The main goal of this research is to minimize the makespan, one of the most important
targets for the task scheduling problem in the cloud areas. An example of task samples
and task sizes is given in Table 1 and another is shown in Table 2 for virtual machines
and the processor speed as individual values.
We aim to reduce the completion time of tasks in this research. This time duration is
the longest completion time among virtual machines. If we consider Ti as the task size
A Schedule of Duties in the Cloud Space 71
of i and Cj as the processor speed of the virtual machine j, we can obtain the makespan
i from Eq. 8.
According to the allocated tasks for each resource and the length of desired tasks,
there has been a completion time for tasks relative to the processor speed of the virtual
machine for each of them.
For instance, assume that the tasks T3 , T6 , T10 , T8 are allocated to virtual machine 2,
the makespan of each task delivered to virtual machine 2 can be calculated as follows:
19 41
texe (3.2) = = 7.9 texe (6.2) = = 17.1
2.4 2.4
16 12
texe (10.2) = = 6.7 texe (8.2) = =5
2.4 2.4
So, the completion time of tasks calculated on virtual machine 2 is:
Similarly, the times for all virtual machines can be computed from the assigned tasks.
The longest completion time of tasks amongst that for all virtual machines is calculated
by Eq. 9:
Makespan = Max tcomplete (j) 1 ≤ j ≤ m (9)
In Eq. 9, tcomplete (j) shows the completion time of tasks allocated to the virtual
machine j. Minimizing Eq. 9 (i.e., the completion time of all tasks (makespan)) is the
main target of this research.
Fig. 6. Performance output for the four scenarios, comparing MSSA with other algorithms; MSSA
shows lower calculation amount, which is desirable as lower values indicate improved efficiency
in minimizing makespan for cloud computing task scheduling.
that the MSSA algorithm can be a useful tool for scheduling cloud computing tasks in
real-world scenarios.
7 Conclusion
The results from the stated scenarios show that the proposed algorithm had better per-
formance compared to the other algorithms to solve the task scheduling problem in all
four scenarios of cloud computing.
The results show that the makespan is reduced by increasing the number of virtual
machines and vice versa. They also indicate that the optimized salp swarm algorithm
has increased performance compared to the basic one. The outputs of all scenarios were
similar and the MSSA is better in all case. As a result, the suggested method has shown
better performance in all scenarios to solve the task scheduling problem in the cloud
computing domain.
In addition, the findings of this study provide valuable insights into the efficiency
of different optimization algorithms for scheduling cloud computing tasks. The MSSA
algorithm has shown substantial potential in reducing task completion time and improv-
ing the overall performance of cloud computing systems. Therefore, it can be concluded
that the MSSA algorithm can be a useful tool for scheduling cloud computing tasks in
real-world scenarios.
Overall, while the study’s results demonstrate the effectiveness of the MSSA algo-
rithm in reducing task completion time and improving the overall performance of cloud
computing systems, it is important to consider the limitations and scope of the study’s
findings. Future work could explore alternative performance metrics, evaluate the algo-
rithm’s robustness and scalability, and investigate its suitability for different cloud
computing scenarios.
Acknowledgment. This material is based in part upon work supported by the National Science
Foundation under grant #DUE-2142360. Any opinions, findings, and conclusions or recommen-
dations expressed in this material are those of the authors and do not necessarily reflect the views
of the National Science Foundation.
References
1. Mosadegh, E., Ashrafi, K., Motlagh, M.S., Babaeian, I.: Modeling the regional effects of
climate change on future urban ozone air quality in Tehran, Iran. arXiv: abs/2109.04644
(2021)
2. Jamali, H., Karimi, A., Haghighizadeh, M.: A new method of cloud-based computation
model for mobile devices: energy consumption optimization in mobile-to-mobile compu-
tation offloading. In: Proceedings of the 6th International Conference on Communications
and Broadband Networking, pp. 32–37. Presented at Singapore (2018). https://doi.org/10.
1145/3193092.3193103
3. Chen, H., Wang, F.Z., Helian, N., Akanmu, G.: User-priority guided Min-Min scheduling
algorithm for load balancing in cloud computing. In: 2013 National Conference on Parallel
Computing Technologies (PARCOMPTECH), pp. 1–8 (2013)
A Schedule of Duties in the Cloud Space 75
4. Sehgal, N.K., Bhatt, P.C.P.: Cloud Computing: Concepts and Practices. Springer, Cham
(2018). https://doi.org/10.1007/978-3-319-77839-6
5. Sun, H., Chen, S.-P., Jin, C., Guo, K.: Research and simulation of task scheduling algorithm in
cloud computing. TELKOMNIKA Indonesian J. Electr. Eng. 11, 6664–6672 (2013). https://
doi.org/10.11591/telkomnika.v11i11.3513
6. Akilandeswari, P., Srimathi, H.: Survey and analysis on task scheduling in cloud environment.
Indian J. Sci. Technol. 9(37), 1–6 (2016). https://doi.org/10.17485/ijst/2016/v9i37/102058
7. Singh, A.B., Bhat, S., Raju, R., D’Souza, R.: A comparative study of various scheduling
algorithms in cloud computing. Am. J. Intell. Syst. 7(3), 68–72 (2017). https://doi.org/10.
5923/j.ajis.20170703.06
8. Lavanya, M., Shanthi, B., Saravanan, S.: Multi objective task scheduling algorithm based on
SLA and processing time suitable for cloud environment. Comput. Commun. 151, 183–195
(2020)
9. Mansouri, N., Javidi, M.M.: Cost-based job scheduling strategy in cloud computing environ-
ments. Distrib. Parallel Databases 38, 365–400 (2020). https://doi.org/10.1007/s10619-019-
07273-y
10. Zubair, A.A., et al.: A cloud computing-based modified symbiotic organisms search algorithm
(AI) for optimal task scheduling. Sensors 22(4), 1674 (2022). https://doi.org/10.3390/s22
041674
11. Rajakumari, K., Kumar, M.V., Verma, G., Balu, S., Sharma, D.K., Sengan, S.: Fuzzy based
Ant Colony Optimization scheduling in cloud computing. Comput. Syst. Sci. Eng. 40(2),
581–592 (2022)
12. Ghazipour, F., Mirabedini, S.J., Harounabadi, A.: Proposing a new job scheduling algorithm
in grid environment using a combination of Ant Colony Optimization Algorithm (ACO) and
Suffrage. Int. J. Comput. Appl. Technol. Res. 5(1), 20–25 (2016)
13. Sharma, S., Tyagi, S.: A survey on heuristic approach for task scheduling in cloud computing.
Int. J. Adv. Res. Comput. Sci. 8, 1089–1092 (2017)
14. Mapetu, J.P., Chen, Z., Kong, L.: Low-time complexity and low-cost binary particle swarm
optimization algorithm for task scheduling and load balancing in cloud computing. Appl.
Intell. 49, 3308–3330 (2019)
15. Saeedi, S., Khorsand, R., Ghandi Bidgoli, S., Ramezanpour, M.: Improved many-objective
particle swarm optimization algorithm for scientific workflow scheduling in cloud computing.
Comput. Ind. Eng. 147, 159–187 (2020)
16. Rajagopalan, A., Modale, D.R., Senthilkumar, R.: Optimal scheduling of tasks in cloud com-
puting using hybrid firefly-genetic algorithm. In: Satapathy, S.C., Raju, K.S., Shyamala, K.,
Krishna, D.R., Favorskaya, M.N. (eds.) ICETE 2019. LAIS, vol. 4, pp. 678–687. Springer,
Cham (2020). https://doi.org/10.1007/978-3-030-24318-0_77
17. Mirjalili, S., Gandomi, A.H., Mirjalili, S.Z., Saremi, S., Faris, H., Mirjalili, S.M.: Salp Swarm
Algorithm: a bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 114,
163–191 (2017). https://doi.org/10.1016/j.advengsoft.2017.07.002
Layer-Wise Filter Thresholding Based
CNN Pruning for Efficient IoT Edge
Implementations
1 Introduction
In recent years, the field of artificial intelligence (AI) has made remarkable
strides, with deep learning techniques, particularly convolutional neural net-
works (CNNs), leading the way. CNNs have played a pivotal role in achieving
state-of-the-art results in image classification, object detection, and computer
vision tasks. However, their computational complexity and memory requirements
present challenges for deploying them on resource-constrained devices like mobile
phones and embedded systems.
Optimizing CNNs for hardware has become increasingly crucial as these net-
works grow in complexity and size. The substantial computational and mem-
ory demands of CNNs make it difficult to deploy them on devices with limited
resources such as smartphones, embedded systems, and IoT edge devices. Hard-
ware optimization techniques can significantly reduce the computational require-
ments and memory usage of CNNs, making them more viable for deployment on
resource-constrained devices.
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 76–93, 2023.
https://doi.org/10.1007/978-3-031-45878-1_6
Layer-Wise Filter Thresholding Based CNN Pruning 77
The benefits of optimizing CNNs for hardware are manifold, including faster
inference times and lower energy consumption. Faster inference times are particu-
larly important for real-time applications like autonomous vehicles and robotics,
where rapid decision-making is critical. Similarly, lower energy consumption is
desirable for such applications as it can extend battery life and reduce overall
ownership costs.
Optimizing CNNs for hardware is crucial for making deep learning more
accessible and feasible for a wider range of applications and devices. By reduc-
ing computational requirements, improving energy efficiency, and reducing costs,
optimized CNNs have the potential to enable new possibilities and benefits for a
range of fields from healthcare to education to entertainment. Ongoing research
endeavors are required to devise inventive techniques that maximize CNN per-
formance on resource-constrained devices, driving advancements in hardware
optimization and facilitating a broader deployment of CNNs in real-world sce-
narios [8,11].
Another important aspect of optimizing CNNs for hardware is the need for
efficient memory management. In addition to reducing the size of the network,
memory management techniques can be applied to optimize memory usage dur-
ing the training and inference process. For example, techniques such as data
augmentation and batch normalization can reduce the memory usage of the
CNN during training. During inference, techniques such as model compression
can be used to further reduce the memory requirements of the network [4].
Our proposed approach presents a novel technique called range-based thresh-
old pruning for optimizing convolutional neural networks (CNNs). Traditional
pruning methods rely on fixed threshold values, which may not effectively cap-
ture the weight distribution within each layer. In contrast, our approach lever-
ages the weight ranges present in each layer to determine the pruning threshold
dynamically. By profiling the weight matrix of a pre-trained CNN, we calculate
the maximum weight value for each layer. This maximum weight serves as a
reference for normalizing the pruning threshold. We then iteratively adjust the
threshold for each layer based on its weight range, selectively pruning weights
below the threshold. We have used profiling on multilayer perceptron (MLP)
networks to find the least sensitive neurons to prune in [7]. The key advantage
of range-based threshold pruning is its fine-grained control over the pruning
process. By considering the weight ranges, we can preserve important network
information while removing unnecessary weights. This leads to improved network
efficiency and reduced memory footprint, making CNN models more suitable for
deployment on resource-constrained devices. Our experimental results demon-
strate the effectiveness of range-based threshold pruning. We evaluate our app-
roach on the widely used LeNet5 network, pruning each layer individually. The
results show that our method achieves significant weight reduction while main-
taining high accuracy on three benchmark datasets, MNIST, Fashion-MNIST,
and SVHN datasets.
Fine-grained pruning refers to the ability to selectively prune individual
weights based on their importance or contribution to the network. Instead of
78 L. K. Kalyanam et al.
using a fixed threshold that applies uniformly to all weights, our approach takes
into account the weight ranges to determine the optimal pruning threshold for
each layer. This allows us to remove less important or redundant weights while
preserving the critical ones. This makes the network more efficient and suitable
for deployment on devices with limited resources, such as mobile phones or IoT
edge devices, where computational power and memory capacity are constrained.
between predicted outputs and actual labels. This training process typically uti-
lizes massive image datasets with labels. CNNs have applications in domains such
as natural language processing, speech recognition, recommendation systems,
and biomedical image analysis, in addition to computer vision. In tasks such
as sentiment analysis, language translation, speech recognition, and biomedical
image classification, CNNs have demonstrated promising findings. CNNs’ abil-
ity to automatically acquire hierarchical representations from input data makes
them applicable to tasks and domains beyond computer vision. In addition to
computer vision, CNNs have applications in disciplines such as natural language
processing and drug discovery. In Sect. 1, we discussed the significance of prun-
ing CNNs as a hardware optimization technique. However, additional techniques
exist to improve the hardware performance of CNNs, and we will examine these
techniques in the following sections.
– Quantization: This involves reducing the precision of the weights and acti-
vations of the network. For example, using 8-bit integers instead of 32-bit
floating point numbers can significantly reduce the memory requirements and
improve the speed of the network [2,16,21].
– Pruning: Pruning convolutional neural networks (CNNs) involves removing
redundant weights from the neural network, which can significantly reduce
the network’s size and computational requirements. Pruning is one of several
techniques used to optimize CNNs for hardware implementation, and it has
been shown to be effective in reducing the hardware requirements of CNNs
while maintaining their accuracy and performance [6].
– Weight sharing: This method reduces the memory requirements of convolu-
tional neural networks (CNNs) by sharing the weights of multiple neurons. By
reducing the number of unique weights stored in memory, weight sharing can
considerably reduce the memory bandwidth requirements of CNNs, allowing
them to be implemented more efficiently on resource-constrained devices [3].
– Hardware-specific optimizations: The network’s architecture and algorithms
require being configured to the hardware platform it will be implemented on.
For instance, the network’s speed and efficiency can be substantially increased
by employing specifically designed hardware for convolutional operations [5].
Efficient memory management is also crucial when optimizing CNNs for hard-
ware. Techniques such as data augmentation and batch normalization can be
used to reduce memory usage during training, and model compression can be
utilized during inference to further reduce memory requirements [4]. Overall,
these methods can significantly improve the efficiency and performance of CNNs
when implemented on hardware devices with limited resources, such as embed-
ded systems or mobile devices.
Neuron pruning is a technique in deep learning that aims to enhance the effi-
ciency of a neural network by reducing the number of neurons. The aim is to
80 L. K. Kalyanam et al.
the network has a large number of redundant weights that can be removed with-
out affecting the network’s performance. This is because unstructured pruning
allows for greater flexibility in removing individual weights, while structured
pruning involves removing entire neurons or groups of neurons from the net-
work. Structured pruning is better suited for cases where the network has a
large number of redundant structures, such as filters or channels [1,10].
2.3 LeNet-5
The LeNet-5 architecture has been the subject of extensive research and exper-
imentation over the years, with many researchers exploring new and innova-
tive ways to optimize its performance for various tasks and hardware platforms.
In addition, researchers have explored the use of different activation functions,
such as ReLU and sigmoid, to improve the performance of the network. Some
researchers have even experimented with the use of hybrid activation functions,
which combine the strengths of multiple activation functions to achieve bet-
ter results. We use the ReLU activation function in our network. LeNet-5 is
extensively utilized in numerous applications, including handwritten digit recog-
nition, object detection, and facial recognition. A notable application of LeNet-5
is the recognition of handwritten characters, such as postal codes on letters and
checks dataset [14]. The architecture of LeNet-5 served as the basis for many
other CNNs, including AlexNet, VGGNet, and ResNet, and has had a signif-
icant impact on the evolution of modern CNNs. LeNet-5 can be implemented
in the Internet of Things applications requiring efficient and accurate image
recognition.
LeNet-5 can accurately classify and analyze medical images, such as X-rays
and MRI scans, for quicker and more precise diagnosis. By using the power of
deep learning, LeNet-5 can solve difficult image recognition tasks quickly and
accurately [9].
In this section, we provide an overview of the related work in the field of opti-
mizing convolutional neural networks (CNNs). Numerous techniques, including
model compression, weight pruning, and quantization, have been proposed to
boost the performance of CNNs. ADMM-NN is a framework for joint weight
pruning and quantization of deep neural networks that use ADMM, a technique
to solve non-convex optimization problems with combinatorial constraints. It
achieves significantly higher pruning ratios than the state-of-the-art, with 85×
and 24× pruning on LeNet-5 and AlexNet models, respectively, and 1,910× and
231× reductions in overall model size on these two benchmarks when combining
weight pruning and quantization. The framework also prioritizes convolutional
layer compression and accounts for hardware performance overhead. Addition-
ally, ADMM-NN performs hardware-aware DNN optimizations, which take into
account computation reduction, energy efficiency improvement, and hardware
performance overhead due to irregular sparsity. This is achieved through a con-
cept of break-even pruning ratio, which is the minimum pruning ratio of a spe-
cific layer that results in no hardware performance degradation. Furthermore,
ADMM-NN achieves these results without accuracy loss and has shown highly
promising results on other representative DNNs such as VGGNet and ResNet-
50. Overall, ADMM-NN presents a robust and efficient solution for DNN model
compression for hardware implementation [17].
You et al. [20] propose a Reconfigurable Sparse convolutional Neural Network
accelerator design (RSNN) that combines software and hardware optimizations
Layer-Wise Filter Thresholding Based CNN Pruning 83
for sparse CNN computation on FPGAs. The proposed design includes an effi-
cient sparse dataflow for convolution, a load balance-aware pruning method,
a kernel merging technique, and an efficient reconfigurable hardware accelera-
tor design. RSNN achieves high performance on Xilinx Zynq ZC706, with 87.7
GOPS for AlexNet and 112.8 GOPS for VGG16. It is significantly more efficient
than previous dense FPGA accelerator designs and even more efficient than
sparse accelerators like NullHop. Salama et al. [18] proposed a pruning method
to compress neural networks by removing entire filters and neurons according to
a global threshold across the network without pre-calculation of layer sensitiv-
ity. The method has been proven viable by producing highly compressed models,
including VGG-16, ResNet-56, and ResNet-110 on CIFAR10 and ResNet-34, and
ResNet-50 on ImageNet. Additionally, the method compresses more than 56%
and 97% of AlexNet and LeNet-5 respectively. They prune the network and
retrain iteratively to reach the pruned model, and they find that it results in a
network with more activated neurons.
3 Proposed Method
On a pre-trained CNN network, we propose a threshold-based optimization.
The seven layers of the LeNet network are described in Sect. 2. The methodol-
ogy involves a systematic process for determining the optimal pruning threshold
and selectively pruning weights based on their magnitude. The pre-trained net-
work evaluates the precision of the pruned model on the test set and adjusts
the pruning threshold. We profile the weight matrix of a pre-trained CNN to
identify the maximum weight value for each layer. This maximum weight value
serves as a reference point for normalizing the pruning threshold. We calculate
the normalized threshold by dividing a threshold value by the layer’s weights
maximal weight value as shown in Eq. 1.
T hreshold
T hresholdnormalized = (1)
M ax(weight matrixlayer )
The novelty of the proposed method lies in the approach of range-based thresh-
old pruning for optimizing convolutional neural networks (CNNs). Traditional
pruning approaches often use fixed threshold values, which may not represent
the weight distribution inside each layer properly [12]. The proposed method,
on the other hand, dynamically sets the pruning threshold based on the weight
ranges inside each layer. By profiling the weight matrix and calculating the
optimal threshold for each layer, the method achieves fine-grained pruning and
maintains important network information while removing unnecessary weights.
In addition, the iterative adjustment of the pruning threshold and the use of
a stopping condition ensure that the pruning process is optimized to strike a
balance between model size reduction and preservation of accuracy. Overall, the
novelty of the proposed method lies in its adaptive and data-driven approach to
Layer-Wise Filter Thresholding Based CNN Pruning 85
threshold-based pruning, which effectively exploits the weight ranges within each
layer and provides a systematic method for optimizing the efficiency of CNNs
while keeping the loss in accuracy minimal. The goal of profiling is to understand
86 L. K. Kalyanam et al.
the characteristics of the weights in each layer, which can aid in determining an
appropriate pruning threshold. By analyzing the weight range and distribution,
it becomes possible to set a threshold that effectively prunes unnecessary or less
important weights while preserving important network information. The profil-
ing step helps in making informed decisions about the pruning threshold and
contributes to the overall effectiveness of the pruning process.
4 Experimental Results
In this section, we present the experimental setup and results of our study. Our
experimentation focuses on evaluating the effectiveness of the threshold-based
optimization technique on the LeNet-5 network using three different datasets:
MNIST, CIFAR10, and SVHN. The pretrained network is implemented and
trained using the PyTorch framework [15] on a cloud computing platform. The
pretrained network is the baseline model we make comparisons against.
To assess the performance of the proposed approach, we conduct multiple
experiments where the network is trained independently on each dataset. The
threshold-based optimization process is applied to each baseline model, and the
resulting models are evaluated based on their accuracy and network sparsity.
Sparsity represents the ratio of pruned weights to the total number of weights
in the network. Higher sparsity values indicate a more substantial reduction in
network size and computational demands.
The MNIST dataset is a widely used benchmark for evaluating machine learning
algorithms, particularly those focused on image recognition and classification. It
consists of 60,000 training images and 10,000 testing images of handwritten digits
from 0 to 9. Each image is grayscale and has a resolution of 28×28 pixels. CNNs
have achieved remarkable performance on the MNIST dataset, with some models
achieving error rates as low as 0.23%. The MNIST dataset is a useful benchmark
for evaluating machine learning algorithms new deep-learning architectures and
optimization techniques [5,9].
Table 1, shows the results of optimizing the LeNet-5 network on the MNIST
dataset for five different models trained. The network has 61,706 trainable
parameters. We observe that for the average optimal threshold, we can prune
at least 60% of the prunable parameters, which are weights for an accuracy loss
of 1–5%. Table 2, shows the optimization results for one model in detail, we can
see the number of weights pruned for each layer and compare them to the num-
ber in the baseline model. The results highlight the importance of finding an
optimal pruning threshold that strikes a balance between model size reduction
and maintaining acceptable accuracy. Table 1 provides information about the
pruning threshold, the number of weights in the baseline model, and the num-
ber of weights pruned for each layer of a neural network. In the fully connected
Layer-Wise Filter Thresholding Based CNN Pruning 87
layer 1 (FC1) layer, 32,160 weights are pruned of 48,000, resulting in a sparsity
of approximately 67%. This means that around 67% of the weights in the FC1
layer have been pruned, leading to a sparse network representation. We use the
optimal pruning thresholds to calculate the average optimal threshold, and a
total of 39,973 weights are pruned out of the initial 61,706 weights, demonstrat-
ing a significantly sparser matrix. Figure 4 showcases the impact of pruning on
accuracy for different layers of the MNIST dataset. It compares the accuracy of
the original model with the pruned model as each layer undergoes pruning. The
plot demonstrates that applying dynamic pruning to the CNN model leads to
accuracy changes in each layer. It is evident that the pruning thresholds have
a more pronounced impact on the fully-connected layers (FC1, FC2, and FC3)
compared to the convolutional layers (Conv1 and Conv2). The plot depicts a
steeper decline in accuracy for the fully-connected layers as the pruning thresh-
old increases, signifying their higher sensitivity to pruning. Conversely, the con-
volutional layers demonstrate a more gradual decrease in accuracy, indicating
their relative resilience to changes in the pruning threshold.
Table 2. Optimization results for Lenet with MNIST dataset for one model.
layer name optimal pruning threshold # weights in baseline model # weights pruned
conv1 0.67 150 100
conv2 0.67 2,400 1,608
fc1 0.67 48,000 32,160
fc2 0.30 10,080 5,541
fc3 0.30 840 564
Total avg = 0.48 61,706 39,973
4.2 FASHION-MNIST
The Fashion MNIST dataset is a collection of images of clothing items, designed
as a replacement for the traditional MNIST dataset. It consists of 70,000
88 L. K. Kalyanam et al.
Fig. 4. Change in accuracy for each layer during pruning: comparison between the
original model and the pruned model on the MNIST dataset.
Exp # # weights % weights accuracy (%) threshold accuracy (%) in loss (%)
pruned in pruned regular model proposed model accuracy
model
1 34,160 70.90 89.63 0.44 87.39 2.24
2 28,907 65.07 89.67 0.41 87.24 2.43
3 28,237 63.56 89.15 0.38 87.11 2.04
4 30,218 68.02 89.98 0.41 86.76 3.22
5 25,976 58.47 89.02 0.32 85.34 3.68
Table 3 presents the results of optimizing the LeNet-5 network on the Fashion-
MNIST dataset for five different models. The network has a total of 62006 train-
able parameters, and we can see that the optimal pruning thresholds for each
model differ. We observe that, for the average optimal threshold, we can prune an
average of 65% of parameters, which are the network weights. The correspond-
ing loss percentages range from 2.04% to 3.68%, suggesting a slight trade-off
between accuracy and loss.
Layer-Wise Filter Thresholding Based CNN Pruning 89
In Table 4, we present the optimization results for one model in detail, includ-
ing the number of weights pruned for each layer and a comparison to the baseline
model. We observe that the weights pruned for each layer vary, indicating that
the optimal pruning threshold for each layer is different. This observation is sup-
ported by the results presented in Fig. 5, where we see the change in accuracy
for each layer of the model for different average optimal threshold values.
To calculate the average optimal pruning thresholds, we use the values
obtained for all models, and the results are presented in Table 3. Figure 5 illus-
trates the change in accuracy for each layer during the pruning process of the
proposed model on the Fashion-MNIST dataset. It is worth noting that, with the
highest average optimal threshold value of 0.48, the accuracy drops drastically at
the first convolution layer. The analysis of the graph shows that as the pruning
threshold increases, there is a gradual decrease in accuracy for each layer.
Table 4. Optimization results for Lenet with FASHION-MNIST dataset for one model.
layer name optimal pruning threshold # weights pruned # weights baseline model
conv1 0.4 79 150
conv2 0.4 1,272 2,400
fc1 0.4 16,282 30,720
fc2 0.35 5,147 10,080
fc3 0.15 150 840
total 34,160 44,426
average threshold 0.48
Fig. 5. Change in accuracy for each layer during pruning: comparison between the
original model and the pruned model on the Fashion-MNIST dataset.
Exp # # weights % weights accuracy (%) avg threshold accuracy (%) loss (%)
pruned in pruned model in proposed accuracy
model model
1 23,106 52.01 88.37 0.4 87.0 1.37
2 33,783 54.48 88.34 0.4 85.46 2.88
3 31,351 50.56 88.69 0.4 86.72 1.97
4 33,779 54.48 88.37 0.35 87.08 1.29
5 36,166 58.33 87.54 0.4 86.34 1.2
We use the format in 32×32 image sizes because LeNet-5 has an input size of
32×32.
Table 5 shows the results of optimizing the LeNet-5 network on the SVHN
dataset for five different models. The network has 62006 trainable parameters.
We observe that, for the average optimal threshold, we can prune at least 50%
of the prunable parameters (which are weights) for an accuracy loss of 1–3%.
Table 6 shows the optimization results for one model in detail, including the
number of weights pruned for each layer and a comparison to the baseline model.
We use the optimal pruning thresholds of all models to calculate the average,
which is presented in Table 5. In Fig. 6, we see the change in accuracy for each
layer of the model for different average optimal threshold values. We observe that,
for the highest average optimal threshold of 0.42, accuracy drops drastically at
the first convolution layer.
Figure 6 illustrates the relationship between the pruning threshold and the
accuracy of the pruned model for different layers in the SVHN dataset. Some
Layer-Wise Filter Thresholding Based CNN Pruning 91
layers exhibit a more gradual decline in accuracy, while others show a steeper
decline. This suggests that different layers contribute differently to the overall
model performance and have varying degrees of sensitivity to weight pruning.
Fig. 6. Change in accuracy for each layer during pruning: comparison between the
original model and the pruned model on the SVHN dataset.
Table 6. Optimization results for Lenet with SVHN dataset for one model.
layer name optimal pruning threshold # weights pruned # weights baseline model
conv1 0.4 180 450
conv2 0.4 960 2400
fc1 0.4 19,200 48000
fc2 0.35 7,308 10080
fc3 0.15 166 840
total 36,166 62006
average threshold 0.34
References
1. Anwar, S., Hwang, K., Sung, W.: Structured pruning of deep convolutional neural
networks. CoRR abs/1512.08571 (2015). http://arxiv.org/abs/1512.08571
2. Gholami, A., Kim, S., Dong, Z., Yao, Z., Mahoney, M.W., Keutzer, K.: A survey of
quantization methods for efficient neural network inference. CoRR abs/2103.13630
(2021). https://arxiv.org/abs/2103.13630
3. Han, S., Mao, H., Dally, W.J.: Deep compression: compressing deep neural network
with pruning, trained quantization and Huffman coding. In: Bengio, Y., LeCun, Y.
(eds.) 4th International Conference on Learning Representations, ICLR 2016, San
Juan, Puerto Rico, 2–4 May 2016, Conference Track Proceedings (2016). http://
arxiv.org/abs/1510.00149
4. Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for
efficient neural network. In: Advances in Neural Information Processing Systems,
pp. 1135–1143 (2015)
5. Howard, A.G., et al.: Mobilenets: efficient convolutional neural networks for
mobile vision applications. CoRR abs/1704.04861 (2017). http://arxiv.org/abs/
1704.04861
6. Ide, H., Kobayashi, T., Watanabe, K., Kurita, T.: Robust pruning for efficient
CNNs. Pattern Recogn. Lett. 135, 90–98 (2020). https://doi.org/10.1016/j.
patrec.2020.03.034, https://www.sciencedirect.com/science/article/pii/S01678
65520301185
7. Kalyanam, L.K., Joshi, R., Katkoori, S.: Range based hardware optimization of
multilayer perceptrons with ReLUs. In: 2022 IEEE International Symposium on
Smart Electronic Systems (iSES), pp. 421–426 (2022). https://doi.org/10.1109/
iSES54909.2022.00092
Layer-Wise Filter Thresholding Based CNN Pruning 93
8. Kalyanam, L.K., Ramnath, V.L., Katkoori, S., Zheng, H.: A distributed framework
for real time object detection at low frame rates with IoT edge nodes. In: 2020 IEEE
International Symposium on Smart Electronic Systems (iSES) (Formerly iNiS), pp.
285–290 (2020). https://doi.org/10.1109/iSES50453.2020.00070
9. Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to
document recognition. Proc. IEEE 86(11), 2278–2324 (1998). https://doi.org/10.
1109/5.726791
10. Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P.: Pruning filters for efficient
convnets. CoRR abs/1608.08710 (2016). http://arxiv.org/abs/1608.08710
11. Li, J., Zhang, S., Liu, Y., Cheng, H., Wang, X.: Efficient convolutional neural
network pruning based on batch normalization. IEEE Access 8, 24010–24018 (2020)
12. Liu, J., Tripathi, S., Kurup, U., Shah, M.: Pruning algorithms to accelerate con-
volutional neural networks for edge applications: a survey. CoRR abs/2005.04275
(2020). https://arxiv.org/abs/2005.04275
13. Liu, X., Zhang, X., Wang, H., Lu, J., Liu, Y.: Pruning convolutional neural net-
works for efficient inference on IoT edge devices. IEEE Internet Things J. 5(5),
3664–3675 (2018)
14. Netzer, Y., Wang, T., Coates, A., Bissacco, A., Wu, B., Ng, A.Y.: Reading digits
in natural images with unsupervised feature learning. In: NIPS Workshop on Deep
Learning and Unsupervised Feature Learning (2011)
15. Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learn-
ing library. In: Advances in Neural Information Processing Systems, vol. 32, pp.
8024–8035. Curran Associates, Inc. (2019). http://papers.neurips.cc/paper/9015-
pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
16. Rastegari, M., Ordonez, V., Redmon, J., Farhadi, A.: Xnor-net: imagenet classifi-
cation using binary convolutional neural networks. CoRR abs/1603.05279 (2016).
http://arxiv.org/abs/1603.05279
17. Ren, A., et al.: Admm-nn: an algorithm-hardware co-design framework of DNNs
using alternating direction methods of multipliers. In: Proceedings of the Twenty-
Fourth International Conference on Architectural Support for Programming Lan-
guages and Operating Systems. ASPLOS ’19, pp. 925–938. Association for Com-
puting Machinery, New York (2019). https://doi.org/10.1145/3297858.3304076
18. Salama, A., Ostapenko, O., Klein, T., Nabi, M.: Pruning at a glance: global neural
pruning for model compression. CoRR abs/1912.00200 (2019). http://arxiv.org/
abs/1912.00200
19. Xiao, H., Rasul, K., Vollgraf, R.: Fashion-mnist: a novel image dataset for bench-
marking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017)
20. You, W., Wu, C.: RSNN: a software/hardware co-optimized framework for sparse
convolutional neural networks on FPGAS. IEEE Access 9, 949–960 (2021). https://
doi.org/10.1109/ACCESS.2020.3047144
21. Zhou, A., Yao, A., Guo, Y., Xu, L., Chen, Y.: Incremental network quantization:
Towards lossless CNNs with low-precision weights. CoRR abs/1702.03044 (2017).
http://arxiv.org/abs/1702.03044
Energy-Aware Security for IoT (EAS)
Shrew Distributed Denial-of-Service
(DDoS) Attack in IoT Applications: A
Survey
Wright State University, 3640 Colonel Glenn Hwy, Dayton, OH 45435, USA
{singh.276,baligodugula.2,fathi.amsaad}@wright.edu
Abstract. With the rise of IoT and cloud computing, DDoS attacks
have become increasingly harmful. This paper presents a survey of tech-
niques for detecting and preventing DDoS attacks, specifically focus-
ing on Shrew DDoS or low-rate DDoS attacks. We explore the use of
machine learning for DDoS detection and prevention and introduce a
new potential technique that simplifies the process of detecting and pre-
venting DDoS attacks originating from multiple infected machines, typi-
cally known as zombie machines. As a future direction, we discuss a new
technique to simplify the detection and prevention of shrew DoS attacks
originating from multiple infected machines, commonly known as bot-
nets. The insights presented in this paper will be valuable for researchers
and practitioners in cybersecurity.
1 Introduction
The Internet of Things (IoT) represents the physical thing (i.e.devices, objects,
etc.) connected through the internet. It utilizes embedded systems, smart sen-
sors, software, and other technologies to exchange data with each other over the
internet. This survey explores Shrew distributed denial-of-service attacks to sum-
marize the recent studies in secure communication of IoT applications. First, the
survey explains denial-of-service attacks and the different prevention methods.
It includes papers using the NS-2 simulator to generate a data set and detect
the impact on communication devices. We also explore the security constraints
of the victim network. For that, DDoS is classified into different parameters.
We also discuss novel research that can protect IoT applications against DDoS
attacks. For that, we explain the future work with inference from a real-world
example by detecting the number of machines performing the DoS attack. The
article is organized into different parts which are: a) Literature review: Review
of previously published works b) Research challenges: Identification and exami-
nation of difficulties encountered in research c) Related Research: Investigation
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 97–103, 2023.
https://doi.org/10.1007/978-3-031-45878-1_7
98 H. Singh et al.
of studies that are related to the topic d) Discussion and Improvement: Con-
versation and enhancement of findings e) Future work: Plans for future research
endeavors (Fig. 1).
2 Literature Review
Distributed denial-of-service (DDoS) attacks are designed to disrupt the nor-
mal traffic of a server or network by flooding it with malicious packets. These
attacks are initiated by single or multiple infected systems within a network,
making them robust. To address this issue, researchers have developed various
preventive measures, and we discuss how to detect and counter multiple machines
performing a DDoS attack on a single system. Many researchers examined DDoS
attacks, including UDP flood, SYN flood, HTTP flood, protocol-exploitation
flooding, reflection-based flooding, and amplification-based flooding. Using IoT
as an example, we better understand how these different flooding and exploita-
tion attacks are performed. The perception, network, and application layers are
the three standard levels of IoT architecture.
Shrew DDoS attacks are a specific type of DDoS attack that involves contin-
uously sending packets to a network at a low rate. Unlike flooding DDoS attacks,
shrew attacks occur periodically, which makes them unique. However, they can
still significantly reduce the quality of service on the victim’s system. To counter
this type of attack, researchers have proposed various approaches, including the
spectral template matching approach [1]. Researchers are utilizing the NS sim-
ulator to create a large-scale network simulation test-bed to simulate an attack
flow over background traffic that exhibits self-similarity. TCL scripts are used
to create simulations, and the GREP command-line interface is a helpful tool
for writing the code. The OTCL configuration file, known as the “TCL script,”
contains information on topology creation, node creation, link setup, and other
parameters [2].
3 Research Challenges
When smart home appliances are connected to the internet, the likelihood of
malicious attacks increases, posing significant security concerns for the home sys-
tems if any critical security area is breached. Additionally, the need for firmware
updates for older IoT devices is a challenge. New devices are introduced to the
market without considering legacy devices already paired with users’ networks,
leading to vulnerabilities that attackers can exploit.
Furthermore, privacy concerns arise with the developing new technology,
including smart home appliances. It is essential to ensure that consumers’ privacy
is protected and their personal data is not compromised.
Distributed denial-of-service (DDoS) attacks can cause significant damage
to companies by disrupting their services and resulting in financial losses [3].
The authors propose a machine-learning solution that involves building a Z-
wave network on a Raspberry Pi gateway and using testbed equipment to carry
out a DDoS attack by flooding the gateway with packets to address this issue.
Wireshark is used to capture network traffic during the attack and employs
various machine learning models such as logistic regression, decision trees, ran-
dom forests, support vector machines, and deep learning models to assess the
effectiveness of the proposed intrusion detection systems.
Chen Yu discusses shrew DDoS attacks, which exploit a system’s temporary
behavior to gradually reduce it’s capacity or service quality. These attacks are
also known as “pulsing DDoS attacks” or “reduction of quality attacks.” The
paper proposes a distributed system for collaborative detection and filtering
(CDF) that distinguishes shrew attack flow from legitimate TCP and UDP traffic
flows by detecting a traffic stream with higher energy in the low-frequency band.
Agarwal studied using support vector machines to detect zombies carrying
out DDoS attacks [5]. The paper used radial basis functions and polynomial
functions for the training set and employed the structural risk minimization
principle to generalize performance under small datasets. The training dataset
was obtained from the ns-2 simulator and generated using the Gt-IMT topol-
ogy generator. The study found that with an increase in zombie machines, the
deviation in entropy increased using both the training and test data.
A new approach based on IoT sensors is proposed to create resistance against
DDoS attacks [6]. The proposed solution involves using an Apache server as a
load balancer, a Python script for traffic filtering, and the Wireshark network
protocol analyzer for analyzing traffic.
4 Related Research
There are only a few researchers have explored utilizing machine learning algo-
rithms to detect the number of machines involved in a DoS attack. Therefore, it
is imperative to investigate and develop effective methods to detect the number
of machines participating in a DoS attack using machine learning techniques.
Practical implementation has also highlighted the importance of identifying the
100 H. Singh et al.
Fig. 3. Methodology on tackling screw DOS attack coming from multiple sources.
6 Future Work
As a further development to the research above, we aim to conduct a tech-
nical investigation focused on generating a single wavelet using multiple pulse
rate frequencies obtained from various zombie machines during a low-rate DDoS
attack. This wavelet will encompass all relevant pulse rate data, including the
time intervals of maximum and minimum rates and the duration of the pulse
rate during which the infected packets congest the network line.
References
1. Chen, Yu., Hwang, K.: Collaborative detection and filtering of shrew DDoS attacks
using spectral analysis. J. Parallel Distrib. Comput. 66(9), 1137–1151 (2006)
2. Li, L., Lee, G.: DDoS attack detection and wavelets. Telecommun. Syst. 28, 435–
451 (2005)
3. Toutsop, O., Das, S., Kornegay, K.: Exploring the security issues in home-based
IoT devices through denial of service attacks. In: 2021 IEEE SmartWorld, Ubiq-
uitous Intelligence and Computing, Advanced and Trusted Computing, Scalable
Computing and Communications, Internet of People and Smart City Innovation
(SmartWorld/SCALCOM/UIC/ATC/IOP/SCI). IEEE (2021)
4. Takahashi, Y., Inamura, H., Nakamura, Y.: A low-rate DDoS strategy for unknown
bottleneck link characteristics. In: 2021 IEEE International Conference on Per-
vasive Computing and Communications Workshops and other Affiliated Events
(PerCom Workshops). IEEE (2021)
Shrew DDoS Attack in IoT Applications 103
5. Agrawal, P.K., Gupta, B.B., Jain, S.: SVM based scheme for predicting number of
zombies in a DDoS attack. In: 2011 European Intelligence and Security Informatics
Conference. IEEE (2011)
6. Huraj, L., Šimon, M., Horák, T.: Resistance of IoT sensors against DDoS attack
in smart home environment. Sensors 20(18), 5298 (2020)
7. Sinha, M., et al.: Securing an accelerator-rich system from flooding-based denial-
of-service attacks. IEEE Trans. Emerg. Top. Comput. 10(2), 855–869 (2021)
8. Araki, R., Hsu, Y.-F., Matsuoka, M.: Early detection of campus network DDoS
attacks using predictive models. In: GLOBECOM 2022–2022 IEEE Global Com-
munications Conference. IEEE (2022)
9. Ali, J., et al.: A machine learning framework for prevention of software-defined
networking controller from DDoS attacks and dimensionality reduction of big data.
In: 2020 International Conference on Information and Communication Technology
Convergence (ICTC). IEEE (2020)
10. Ibrahim, R.F., Al-Haija, Q.A., Ahmad, A.: DDoS attack prevention for internet of
thing devices using ethereum blockchain technology. Sensors 22(18), 6806 (2022)
11. Rani, S.V.J., et al.: Detection of DDoS attacks in D2D communications using
machine learning approach. Comput. Commun. 198,32–51 (2023)
12. Shen, Y., Fei, M., Du, D., Zhang, W., Stanković, S., Rakić, A.: Cyber security
against denial of service of attacks on load frequency control of multi-area power
systems. In: Li, K., Xue, Y., Cui, S., Niu, Q., Yang, Z., Luk, P. (eds.) LSMS/ICSEE
-2017. CCIS, vol. 763, pp. 439–449. Springer, Singapore (2017). https://doi.org/
10.1007/978-981-10-6364-0_44
13. Al-Maani, M., et al.: A classifier to detect number of machines performing DoS
attack against arduino Oplà device in IoT environment. In: 2022 5th International
Conference on Advanced Communication Technologies and Networking (Comm-
Net). IEEE (2022)
14. Rustam, F., et al.: Denial of service attack classification using machine learning
with multi-features. Electronics 11(22), 3817 (2022)
15. Biron, Z.A., Dey, S., Pisu, P.: Real-time detection and estimation of denial of
service attack in connected vehicle systems. IEEE Trans. Intelli. Transport. Syst.
19(12), 3893–3902 (2018)
16. Idhammad, M., Afdel, K., Belouch, M.: Detection system of HTTP DDoS attacks
in a cloud environment based on information theoretic entropy and random forest.
In: Security and Communication Networks 2018 (2018)
Honeypot Detection and Classification
Using Xgboost Algorithm for Hyper
Tuning System Performance
1 Introduction
A smart contract is a simple computer program designed to automate certain
things or events under an agreement or contract. The first public blockchain to
enable smart contracts is Ethereum. However, the contract design in Ethereum
exhibits obvious flaws such as when a user sends some funds to a contract,
then that same user can withdraw the funds that were just sent. Due to these
flaws, many attempts are to poach the vulnerabilities and target naive users.
To prevent this from happening such a detection system’s existence is crucial.
When someone attempts to exploit this flaw, they are unable to recover the
money initially sent. In this case, the malicious attacker is the one who creates
the contract. On the other hand, the one who falls for the trap is known as the
victim. Honeypots take advantage of the naivete of the victim by preying on
the victim’s ignorance of hidden attributes. There are hundreds of honeypots in
Ethereum that have been discovered, and depending on the various techniques
used, they are divided into many types. Malicious users creating smart contracts
on the blockchain mainly benefit from such honeypots. Thus, there is a higher
emphasis on security when public smart contracts are concerned.
Ethereum smart contracts have gained immense popularity from many
sources, such as the media and industry, in the past few years. With the increase
in popularity, there has been a rise in malicious users capitalizing on this sit-
uation, trying to find new opportunities to deceive newcomers and profit from
them. Consequently, such lucrative individuals started luring others into con-
tracts that seem exploitable, but they are a honeypot with a hidden trap that
in turn benefits the one who creates the contract. The detection of this honey-
pot is a challenging task. Hence, this system is developed to detect and classify
honeypots based on their features in the Ethereum blockchain platform.
2 Related Work
This section focuses on the important work done and research carried out rel-
evant to this topic. It also provides comprehensive information such as the
authors, the methodology adopted, and the research gap identified. Research on
honeypots focuses on measurement and detection. Torres et al. [3] first defined
the honeypot in The Art of Scam. They used a variety of methods to measure the
number of smart contracts and honeypots. They found that there were 150,000
smart contracts and 282 honeypots and that these numbers are growing each
day. In the past, Honeypot systems were used neither detecting intrusion detec-
tion systems nor the firewall for a direct specific problem. Honeypots now are
used as a part of security systems and what kind of problem they will offer a
solution is depends on the design and usage purposes. Hence contrary to other
information security equipment is not to be able to mention a honeypot that
can give a general answer to every problem solution [3].
106 V. Musale et al.
The dataset is split into 10 parts. In each iteration nine-tenth of the dataset
is used to train the model and the remaining one-tenth part is used for testing
the accuracy. In every iteration, the one-tenth testing portion keeps on changing.
With the help of such parameters and many others being hyper-tuned, the model
is able to increase its accuracy by a combination of iterations mentioned in the
results section. This makes the model more robust than what we have seen in the
past in the form of literature review and gap identification (Fig. 3). A detailed
explanation of the same is presented in the results and discussions section.
Additional data cleaning process is carried out before applying machine learning
algorithms to conduct experiments. Mainly, the missing values are managed in
our dataset by doing feature extraction. Because all of the transactions in the
contract have errors, it is impossible to compute the aggregated characteristics
Honeypot Detection and Classification Using Xgboost Algorithm 109
of any type of transaction. When a contract has only one normal transaction, it
is difficult to measure the difference in time or block number between subsequent
normal transactions. Since all transactions have at least one normal outcome, any
remaining characteristics of a normal transaction are always defined. Sometimes,
we cannot find the right answer to a question. But even when we cannot find
the answer, still used zeros to represent it.
The system uses a machine learning model to predict whether a contract is
a honeypot or not. We use k-fold cross-validation to make sure that the model
is effective at predicting data that has not been seen before. The data is divided
into 10 different groups, each of which is used to test the algorithm. Each group
of ten people will be used for testing once. We used a stratified k-fold cross-
validation procedure to ensure that the data is representative of the population.
There are more non-honeypot contracts in our dataset than honeypot contracts.
This makes it hard for computer algorithms to learn how to classify contracts.
The XGBoost algorithm is set to use a scaling factor for the positive class
so it would learn better. This model also measures how much power each fold
has, the AUROC technology. AUROC stands for the Area Under the Receiver
Operating Characteristics graph. The ROC curve shows how well a model can
correctly identify different categories. The area under the curve shows how many
times the model correctly classified a particular item as belonging to one of the
categories. The performance of the XGBoost classifier on the selected dataset is
shown below along with the results of the Honeypot classification model.
train ROC AUC 0.996 TN 141700 FP 1012 FN 0 TP 264
Iteration 1: test ROC AUC 0.980 TN 15729 FP 127 FN 1 TP 30
train score - test score = 0.017
train ROC AUC 0.996 TN 141666 FP 1034 FN 0 TP 276
Iteration 2: test ROC AUC 0.997 TN 15757 FP 111 FN 0 TP 19
train score - test score = -0.000
110 V. Musale et al.
This section describes the readings of a particular traffic signal at one of the
chosen squares. From Table 2, Signal No. (i) depicts the iteration number (1 to
Honeypot Detection and Classification Using Xgboost Algorithm 111
10). N and Traffic Level used in Table 2 represents the count of vehicles and
traffic density at that particular traffic signal respectively.
7 Conclusion
To detect honeypots, the implemented system offered a step-by-step technique
in for obtaining, processing, and analyzing Ethereum contracts. With the use of
real data, we demonstrated how assumptions and theories regarding honeypot
behavior may be compared, and we developed features for classification models.
Even after removing all the contracts associated with a single honeypot approach
from training, the machine learning models continued to perform well in terms
of generalization. Most crucially, we demonstrated that our method discovered
honeypots from two novel methodologies, which would not have been achievable
using byte code analysis without the creation of new detection criteria manu-
ally. Of particular interests would be to assess honeypot detection systems with
regards security (e.g. [7,8,10,15–18]) and improvements in the sustainability
(e.g. energy consumption [4–6,19,20]) of such detection systems.
References
1. Bashir, U., Chachoo, M.: Intrusion detection and prevention system: challenges
& opportunities. In: 2014 International Conference on Computing for Sustainable
Global Development (INDIACom), pp. 806–809. IEEE (2014)
2. Benmoussa, H., Abou El Kalam, A., Ouahman, A.A.: Towards a new intelligent
generation of intrusion detection system. In: Proceedings of the 4th Edition of
National Security Days (JNS4), pp. 1–5. IEEE (2014)
3. Camino, R., Torres, C.F., Baden, M., State, R.: A data science approach for detect-
ing honeypots in ethereum. In: 2020 IEEE International Conference on Blockchain
and Cryptocurrency (ICBC). pp. 1–9. IEEE (2020)
4. Castellon, C., Roy, S., Kreidl, P., Dutta, A., Bölöni, L.: Energy efficient Merkle trees
for blockchains. In: 2021 IEEE 20th International Conference on Trust, Security
and Privacy in Computing and Communications (TrustCom), pp. 1093–1099. IEEE
(2021)
5. Castellon, C.E., Roy, S., Kreidl, O.P., Dutta, A., Bölöni, L.: Towards an energy-
efficient hash-based message authentication code (hmac). In: 2022 IEEE 13th Inter-
national Green and Sustainable Computing Conference (IGSC), pp. 1–7. IEEE
(2022)
6. Escobar, C.C., Roy, S., Kreidl, O.P., Dutta, A., Bölöni, L.: Toward a green
blockchain: Engineering Merkle tree and proof of work for energy optimization.
IEEE Trans. Netw. Serv. Manag. 19(4), 3847–3857 (2022)
7. Garrett, K., Talluri, S.R., Roy, S.: On vulnerability analysis of several password
authentication protocols. Innov. Syst. Softw. Eng. 11, 167–176 (2015)
8. Gouge, J., Seetharam, A., Roy, S.: On the scalability and effectiveness of a cache
pollution based dos attack in information centric networks. In: 2016 International
Conference on Computing, Networking and Communications (ICNC), pp. 1–5.
IEEE (2016)
9. Haltaş, F., Uzun, E., Şişeci, N., Poşul, A., Emre, B.: An automated bot detection
system through honeypots for large-scale. In: 2014 6th International Conference
On Cyber Conflict (CyCon 2014), pp. 255–270. IEEE (2014)
10. Khatwani, C., Roy, S.: Security analysis of ECC based authentication protocols. In:
2015 International Conference on Computational Intelligence and Communication
Networks (CICN), pp. 1167–1172. IEEE (2015)
11. Koniaris, I., Papadimitriou, G., Nicopolitidis, P., Obaidat, M.: Honeypots deploy-
ment for the analysis and visualization of malware activity and malicious con-
nections. In: 2014 IEEE International Conference on Communications (ICC), pp.
1819–1824. IEEE (2014)
12. Markert, J., Massoth, M.: Honeypot effectiveness in different categories of attacks
on wireless sensor networks. In: 2014 25th International Workshop on Database
and Expert Systems Applications, pp. 331–335. IEEE (2014)
13. Musca, C., Mirica, E., Deaconescu, R.: Detecting and analyzing zero-day attacks
using honeypots. In: 2013 19th International Conference on Control Systems and
Computer Science, pp. 543–548. IEEE (2013)
14. Paul, S., Mishra, B.K.: Honeypot based signature generation for defense against
polymorphic worm attacks in networks. In: 2013 3rd IEEE International Advance
Computing Conference (IACC), pp. 159–163. IEEE (2013)
15. Roy, S.: Denial of service attack on protocols for smart grid communications. In:
Research Anthology on Combating Denial-of-Service Attacks, pp. 560–578. IGI
Global (2021)
Honeypot Detection and Classification Using Xgboost Algorithm 113
16. Roy, S., Das, A.K., Li, Y.: Cryptanalysis and security enhancement of an advanced
authentication scheme using smart cards, and a key agreement scheme for two-
party communication. In: 30th IEEE International Performance Computing and
Communications Conference, pp. 1–7. IEEE (2011)
17. Roy, S., Khatwani, C.: Cryptanalysis and improvement of ECC based authentica-
tion and key exchanging protocols. Cryptography 1(1), 9 (2017)
18. Roy, S., Morais, F.J.A., Salimitari, M., Chatterjee, M.: Cache attacks on blockchain
based information centric networks: an experimental evaluation. In: Proceedings
of the 20th International Conference on Distributed Computing and Networking,
pp. 134–142 (2019)
19. Roy, S., Rudra, A., Verma, A.: An energy complexity model for algorithms. In:
Proceedings of the 4th Conference on Innovations in Theoretical Computer Science,
pp. 283–304 (2013)
20. Roy, S., Rudra, A., Verma, A.: Energy aware algorithmic engineering. In: 2014
IEEE 22nd International Symposium on Modelling, Analysis and Simulation of
Computer and Telecommunication Systems, pp. 321–330. IEEE (2014)
21. Shukla, R., Singh, M.: Pythonhoneymonkey: detecting malicious web URLs on
client side honeypot systems. In: Proceedings of 3rd International Conference on
Reliability, Infocom Technologies and Optimization, pp. 1–5. IEEE (2014)
22. da Silva Vargas, I.R.J., Kleinschmidt, J.H.: Capture and analysis of malicious traffic
in voip environments using a low interaction honeypot. IEEE Latin Am. Trans.
13(3), 777–783 (2015)
23. Suo, X., Han, X., Gao, Y.: Research on the application of honeypot technology in
intrusion detection system. In: 2014 IEEE Workshop on Advanced Research and
Technology in Industry Applications (WARTIA), pp. 1030–1032. IEEE (2014)
Electromagnetic Fault Injection Attack
on ASCON Using ChipShouter
Centre for Cyber Security Systems and Networks, Amrita Vishwa Vidyapeetham,
Amritapuri, Kollam 690525, Kerala, India
[email protected], [email protected]
1 Introduction
procedures are employed to ensure that hardware meets strict security require-
ments and can withstand various attacks. These measures help establish trust
with customers and stakeholders by ensuring that devices are trustworthy and
secure [4,16]. Hardware security is a fundamental aspect of overall security in
industries dealing with sensitive data or critical equipment. It requires a multi-
layered approach that combines various hardware security controls, certification
and testing procedures, and supply chain security controls.
We use EMFI as an attack technique in our field it involves intentionally
causing faults or glitches in a device by exposing it to electromagnetic interfer-
ence. This technique is commonly employed in security evaluations to assess the
resistance of electronic devices, including micro-controllers and smart cards, to
attacks. By exposing a device to electromagnetic pulses of various frequencies
and intensities, researchers can simulate real-world electromagnetic interference
and study how the device responds. Through EMFI, security vulnerabilities and
flaws in a device can be identified and used to develop defenses against electro-
magnetic attacks. Nevertheless, EMFI is also a possible attack vector that mali-
cious actors could exploit to breach a device’s security [12]. Thus, safeguarding
the security of electronic devices necessitates awareness and protection against
EMFI attacks.
In addition to the definition of EMFI, researchers have employed several
methods to carry out this attack. Some common approaches [10,11] include
using a pulse generator with handmade injection probes and an oscilloscope to
monitor the EMPs and target devices’ status. AES algorithm implementation
on FPGA and EMI attack injection onto the target device is another approach
taken by researchers. Varying diameters of injection tips comprising ferrite core
and copper wire have also been utilized by researchers. Furthermore, optical
radiations such as a laser or vibrant white light have been applied in some
instances. The ChipShouter, which is used in our work, was also used by the
author in [7] for penetrating hardware wallets. This simple EMP generator sold
by NewAE Technologies has various functions such as voltage, pulse width, pulse
dead-time, pulse repeat, and other status functions. It costs roughly 3.3k USD
and comes with four probes of two types: Counter Clockwise and Clockwise,
each having 1mm and 4mm diameters.
The ChipShouter is used as an EMP (Electromagnetic Pulses) source to
execute an EMFI attack on the M5STACK ESP32 Timer Camera (OV3660)
module, which belongs to the ESP32 family. The module features an 8 MB
PSRAM and 4 MB flash memory, as well as a camera with a maximum reso-
lution of 2048 × 1536 pixels and an OV3660 sensor with a resolution of 3 MP.
It is equipped with a reset button and an LED status indicator on the board,
and its ultra-low power consumption, timing, sleep, and wake-up functions are
enabled by the use of RTC (BM8563). To program the module, the Arduino IDE
platform is used, and it can be connected to a computer through a USB type C
cable. The ASCON for Arduino is programmed and implemented on the module
to store input values such as plaintext, key, nonce, associated data, etc., on the
PSRAM, where the ChipShouter’s injection probe can be targeted. The RS232
116 V. Narayanan and S. Sankaran
Our study intends to evaluate how EMPs affect the PSRAM of the
M5STACK ESP32 module. We are particularly interested in identifying the
faulty bits and the locations of the subsequent faults, as well as how the behavior
of the module changes when it is subjected to EMPs. This module was chosen
due to its large flash and RAM capacities to support the ASCON library and
its 8 MB PSRAM, which can hold variables of various lengths, including plain-
text, keys, nonces, and related data. This allows direct modification of ASCON
program values by injecting EMPs. To perform the research, a variable string of
plaintext values with sizes of 250 Kb, 500 Kb, and 1 MB is used. The resulting
proportion of fault injection is as follows: 24% for a 250 Kb string, 54% for a
500 Kb string, and 90% for a 1 MB string.
2 Background
2.1 Electromagnetic Fault Injection
EMFI assaults can affect any conventional Integrated Chip (IC), including CPUs,
SRAM, voltage regulators, microcontrollers, and processors. An EMFI has the
power to change an IC’s behavior by applying electromagnetic pulses to it; the
resulting electromagnetic interference can create voltage spikes or transient volt-
age dips, which can interrupt the current flow and lead to an IC’s failure. This
interference may affect the circuit’s clock, data, or power supply connections,
which might lead to errors or unexpected behavior [1]. By using specialized
equipment an attacker can inject EMPs onto the target device/module, This
specialized equipment commonly consists of a pulse generator and a magnetic
coil. This equipment can be available commercially at decent rates and can sup-
port multiple features like varying voltage, pulse width, pulse dead time, number
of pulses, etc. via programming. One such device is called the ChipShouter, as
shown in Fig. 1 which is used in our work.
2.2 PSRAM
read-and-write operations but don’t require the high capacity or high power
consumption of traditional DRAM or flash memory would particularly benefit
from it. Like all other integrated circuits (ICs), the PSRAM is vulnerable to
electromagnetic interference (EMI), a disturbance that can cause data loss, bit
flipping, and data corruption. Knowing these details, we made the decision to
launch an EMFI attack on the module’s PSRAM because the ASCON algorithm
is programmed in such a way that the code stores data like plaintext, associated
data, nonce, key, etc. are all stored in the PSRAM. To support the hypothe-
sis that ASCON is vulnerable to EMFI the EMPs from the ChipShouter are
injected into the PSRAM.
Figure 2 shows an 8 MB PSRAM.
2.3 ASCON
The Austrian Graz University of Technology’s research team created ASCON
in 2014 [15]. Its design aims to make it lightweight effective, and safe against
side-channel attacks. Several variations of the ASCON family are available, each
with a different security level and block size, and they may be chosen according
to the particular security needs of the application. The permutation function in
the method is used to convert the input data into an encrypted output. It is
based on a sponge structure. The final encrypted output is created by apply-
ing a set of round keys that are created using the encryption key to the input
data throughout several rounds. Applying a different key to the input data and
the encrypted output results in the generation of the authentication tag. The
118 V. Narayanan and S. Sankaran
National Institute of Standards and Technology (NIST) advises ASCON for use
in lightweight cryptography, and the European Telecommunications Standards
Institute (ETSI) has standardized it. It is utilized in several applications, includ-
ing embedded systems, wireless communication protocols, and the Internet of
Things (IoT). The ASCON encryption process is divided into 4 parts:
1. Initialization: Starts the state with the key K and the nonce N.
2. Associated Data Processing: This method changes the state with related data
blocks. Ai .
3. Plaintext Processing: Introduces plaintext blocks Pi into the state and
retrieves ciphertext blocks Ci .
4. Finalization: Adds the key K and subsequently retrieves the tag T for authen-
tication.
After injecting each block except the last plaintext block, the entire state is
subjected to the core permutation pb . During initialization and finalization, a
stronger permutation pa with more rounds is used. The number of rounds for a
and b and the speed and capacity of the sponge are determined by the specific
version of ASCON (Fig. 3).
3 Related Work
3.1 Electromagnetic Fault Injection
A 32-bit microcontroller built-in CMOS 130 nm technology underwent an EMFI
by the authors of [4] a few years ago. The microcontroller is based on an ARM
Cortex-M3 CPU with a 56 MHz operating frequency and no cache memory built
in. They apply pulses with an amplitude ranging from -200V-+200V and a pulse
Electromagnetic Fault Injection Attack on ASCON Using ChipShouter 119
width of 10ns-200ns to the target device using an EMP generator with a mag-
netic coil. In order to retrieve data from the device they use a Serial Wire Debug
(SWD) which is an alternative to the Joint Test Action Group (JTAG). SWD is
a 2-pin interface that can be used to communicate with the microcontroller and
has the same protocol as that of a JTAG, It employs the bi-directional wire pro-
tocol that is a standard for ARM CPUs and was developed by the ARM Debug
programmer. It is the author’s intention to provide readers with a basic grasp of
the vulnerabilities that an EMFI may introduce into embedded software. They
were able to show that an EMFI can lead to timing constraint violations during
bus transfers from flash memory, which would enable an attacker to circumvent
some safeguards against more traditional timing fault-injection techniques like
clocks or voltage glitches.
Injecting EMPs into a Solo Key open-source FIDO2 authentication key as
well as a Trezor bitcoin wallet using the ChipShouter was done by Colin O’Flynn
[7]. The vulnerable code, which also contains the EMFI, is a part of the USB
stack. The author was able to obtain the device’s private info using this method.
The attack used a standard logical flow that almost all USB stacks provide to
allow reading up to 64 Kb of data from the device. The author further notes that
to successfully execute this attack, the exact timing of the fault injection in refer-
ence to the location of the USB transactions is necessary. As a result, the experi-
ment can be supported by an open-source application called PhyWhisperer-USB,
which triggers a fault injection platform from a USB message with incredibly
high temporal precision.
Beckers, Arthur, et al. [10] explored the impact of EM pulses on the flash
memory of an ATmega328p 8-bit microcontroller. Their study aimed to develop
a fault model that accurately depicts the properties and consequences of the
injected errors. To conduct their EM fault injection experiments, they employed
the Langer EM fault injection setup, which includes various probes generating
magnetic fields, electric fields, and current pulses, as well as a power generator
and a magnetic field pulse source capable of supplying up to 500V to the mag-
netic probe. In their experiment, the authors tried two different approaches: one
included filling the flash memory entirely with zeroes (0 × 00), while the other
involved filling it entirely with ones (0xff). They discovered that only the data
read from the flash memory can be altered and that the modification only lasts
as long as the flash memory values are set to 0xff. This indicates that the real
values in the flash memory are not changed only the process of reading from the
flash memory was impacted.
Dehbaoui, Amine, et al. [2] explain the usage of an EM channel to execute
active attacks against a hardware AES built in an FPGA and a software AES
operating on a CPU. The experiment is carried out on two different platforms,
one of which has a smart card emulation board made up of an 8-bit AVR Atmega
128 micro-controller coupled with a 128 Kb flash program memory, 4KBEEP-
ROM, and 4KBSRAM. The working voltage and frequency of this microcon-
troller are 4.5–5.5V and 3.57 MHz, respectively. The second platform uses an
FPGA from the Xilinx Spartan 3 series. Both platforms use the same EMP sys-
120 V. Narayanan and S. Sankaran
majority of the study has focused on fault injection on FPGA, Raspberry Pi,
and other commercially accessible microcontrollers, where they rely on either an
EMP generator or equipment capable of producing EMPs to inject faults. The
study conducted by [2,4,12] aided in situating the injection probe and selecting
the range of pulse widths to experiment with. The author of [4] assists in compre-
hending the basic problems that an EMFI can create in an embedded program.
Colin O’Flynn [7] uses the ChipShouter to access data from a Trezor Bitcoin
wallet, which aided in understanding the ChipShouter’s possible functionality.
The study done in [10]discusses the injection of faults into a target module’s
flash memory, which motivated us to perform the same.
4 Proposed System
The security and reliability of IoT devices needing protection might suffer signifi-
cantly from ASCON’s susceptibility to Electromagnetic Fault Injection (EMFI).
Previous attempts [3,14,17] used the AES algorithm on FPGAs and microcon-
trollers, thus the researchers had to precisely point the EM probe onto the tar-
get’s memory ICs. This strategy is challenging because it necessitates pinpointing
the configuration memory, power supply lines, and input/output pins, which are
all connected to the microcontroller. This microcontroller might be damaged
by introducing faults, which would produce incorrect results. The goal of this
part is to provide a thorough understanding of how the experiment was carried
out and how the resulting data was analyzed. The proposed system describes
the technique used in our research, including test scenarios, data collecting, and
experiment analysis which are mentioned below:
4.3 Design
We analyze the research done on EMFI [4,10,12] and came up with a different
methodology from them which is mentioned in the previous section. The study’s
overall concept entailed using an experimental setup to cause electromagnetic
faults in the PSRAM of an M5STACK ESP32 module as shown in Fig. 4. In
order to conduct the study, data on the kinds and locations of faults caused by
the ChipShouter were gathered and examined. Additionally, consideration was
given to the study’s limitations. The design of the study is given in the following:
1. Experiment Setup: The study made use of an M5STACK ESP32 Timer
Camera module with a 4 MB flash memory and an 8 MB PSRAM. The encryp-
tion and decryption keys and data were stored in the PSRAM as part of the
Electromagnetic Fault Injection Attack on ASCON Using ChipShouter 123
device’s ASCON algorithm written. The EMPs are injected using a Chip-
Shouter placed close to the module’s PSRAM. Both the Chipshouter and the
target module are connected to the PC.
2. EMFI Parameters: The faults were injected into the PSRAM of the module
using a ChipShouter connected to a PC by adjusting settings such as pulse
width, pulse dead time, number of pulses, and pulse duration. The position
and orientation of the ChipShouter probe were also taken into account.
3. Data Collection: The suggested method gathers data by displaying the
output on the serial monitor of an Arduino IDE installed in the experiment
computer. The output consists of encryption and decryption time and prints
faulty bits when a fault occurs.
4. Data Analysis: The collected data is evaluated by noting all of the errors
that occur throughout the encryption and decryption processes. Each error
is examined separately to see if bits were added or flipped.
5. Limitations: The investigation is constrained by the size of the experimen-
tal set-up and the difficulty of the employed encryption and decryption algo-
rithms. Because most countermeasures in this field of work are standard, the
research did not examine potential defenses against EMFI.
The block diagram in Fig. 6 depicts the implementation of the ASCON algo-
rithm on the M5STACK ESP32 module using the Arduino C language. The code
starts by initializing setup procedures, which principally include a PSRAM ini-
tialization function to test the module’s PSRAM functioning. If the initialization
fails, an error message indicating “PSRAM initialization failed” is printed. In
the event of a successful startup, the allocation of plaintext values, related data,
encrypted text, and decrypted text is performed in the PSRAM rather than the
124 V. Narayanan and S. Sankaran
module’s main memory. If the allocation fails, the message “Memory allocation
failed” is displayed (see Fig. 5).
Following these preliminary steps, the algorithm starts a loop in which data
is placed into the plaintext variable based on the string size defined in the setup
function. Letters from ‘A’-‘Z’ are added repeatedly until the necessary string size
is reached. Following that, the encryption and decryption operations are carried
out, and the timings for each operation are printed.
A condition is introduced after printing the decrypted text to see if it matches
the original plaintext. If they are not equal, it signals a problem. In such circum-
stances, the program is designed to output the glitched values and their related
index numbers, allowing the location at which the glitch occurred.
5 Evaluation
This section covers the results and evaluations of the EMFI attack on the
ASCON cipher experimentation. The ChipShouter is put to use to target the
M5STACK ESP32 module and inject EMPs into its PSRAM. Faults are intro-
duced into the cipher, resulting in outputs that support our contention that the
Electromagnetic Fault Injection Attack on ASCON Using ChipShouter 125
ASCON is susceptible to EMFI attacks. The ASCON could only handle a plain-
text string size of up to 1 MB; if the amount was increased, the code would not
function properly, hence a maximum value of 1 MB was chosen.
We experiment with the insertion of faults using all four injection probes
offered by the ChipShouter and uncovered a common fault. The first register
of the PSRAM memory, where plaintext information is placed, generates an
identical result with all four injection probes. This output sequence implies that
the first register of PSRAM memory is impacted in this way (Table 1).
Glitch Plaintext Around Plaintext Binary Decrypted text Decrypted text Found in
Found at Glitch Values at Around Glitch Binary Values at
Index Glitched Glitched
Position Position
0 ABCDEFGHIJK 01000001 Unrecognized 00000000 250 Kb,
Unicode 500 Kb,
Values 1 MB
The following tables show the outcomes of injecting pulses into the PSRAM
using various string values as code input. We perform fault injection on strings
of 250 Kb, 500 Kb, and 1 Mb, The tables identify where glitches happen in the
plaintext, and the algorithm offers both the places and the accompanying binary
values to determine whether bits were flipped or added. The experiment focuses
on changing the ChipShouter’s pulse width and pulse repeat settings. The pulse
width ranges from 80ns-540ns for all injection tips, while the pulse repeat reflects
the number of pulses per trigger, which is mainly set between 3 and 5. The
ChipShouter supports pulse widths of up to 960ns and pulse repeat values of
up to 1000, we avoid using them to avoid potential harm to the target module.
We could not use string sizes less than 250 Kb and greater than 1 MB because
injection o faults on values less than 250 Kb is difficult since the memory size
of PSRAM is 8 MB and string sizes above 1 MB are not compilable because the
algorithm breaks.
Table 2 shows faults found for a 250 Kb string. Certain values, however,
are missing because they exist in Unicode format, where they show unknown
symbols. The faults found are completely random, involving flipped, added, or
swapped bits. These faults are most noticeable after the encryption process. The
glitches in Table 3 involve bits being changed into lowercase characters, partic-
ularly special characters being added. In Table 4, certain plaintext values are
impacted by glitches prior to the encryption process, thus affecting the decrypted
values.
The graphical data shows how long the ASCON algorithm takes to encrypt
and decrypt plaintext strings and how many faults arise. The horizontal axis
shows the number of algorithm cycles, whereas the vertical axis shows encryption
and decryption durations in microseconds because seconds would not effectively
reflect the fluctuations. Figures 7, 8 and 9 shows the encryption and decryption
Electromagnetic Fault Injection Attack on ASCON Using ChipShouter 127
times for a 250 Kb, 500 Kb, and 1 Mb plaintext string, as well as the faults that
occurred during the operation.
We can see in the graphs that number of faults increases with the increase in
plaintext size. This also affects the encryption and decryption time. A plaintext
string size of 1 MB takes a longer duration than that of 250 Kb and 500 Kb
respectively as the longer the plaintext size the easier to inject faults and the
lesser the size the more difficult to inject faults.
The fault injection percentages are calculated by counting the number of
faults that occur in plaintext string sizes of 250 Kb, 500 Kb, and 1 Mb. The
obtained count is then divided by 50, representing the total number of cycles
taken for each observation of string size. Finally, the result is multiplied by 100
to obtain the fault injection percentages. The equation for the calculation of
fault percentage is given below:
Number of faults
× 100
Total number of cycles
The calculated fault injection percentages according to the equation are as
follows: 24% for the 250 Kb string size, 54% for the 500 Kb size, and 90% for the
1 MB size.
Glitch Plaintext Around Plaintext Binary Decrypted text Decrypted text Binary
Found at Glitch Values at Around Glitch Values at Glitched
Index Glitched Position
Position
132632 WXYZABCDEF 10000111 WXYZABCDEF 01000100
GHIJKLMNOPQ DtKLMNOPQ
216092 WXYZABCDEF 01000111 WXYZABCD 00000000
GHIJKLMNOPQ EF x001D
194704 GHIJKLMNOPQ 01010001 GHIJK 00000000
RSTUVWXYZA LMNOP
59836 ABCDEF 00000000 ABCDEFGHIJ 01001011
GHIJ KLMNOPQRSTU
63170 GHIJKLMNOPQ 01010001 GHIJKLMNOPU 01010101
RSTUVWXYZA UUUUUUUUUU
120872 OPQRSTUVWX 01011001 OPQRSTUVW 10010101
YZABCDEFGHI X$4DTdt
During the initial stage of the experiment, where EMPs were injected into the
module’s PSRAM, the device frequently became unresponsive and occasionally
resulted in memory corruption, and the device obtained responsiveness only after
many resets. Memory corruption happened as a result of the injected EMPs
influencing the target device’s core and, as a result, affecting the algorithm’s
execution.
128 V. Narayanan and S. Sankaran
Glitch Plaintext Around Plaintext Binary Decrypted text Around Decrypted text Binary
Found at Glitch Values at Glitched Glitch Values at Glitched
Index Position Position
80571 NOPQRSTUVW 1011000 NOPQRSTUVWe 1100101
XYZABCDEFGH YZABCDEFGH
208414 OPQRSTUVW 01011001 OPQRSTUVWX 10000101
XYZABCDEFGHI $4DTdt
477393 XYZABCDEF 01001000 XYZABCDEFGD 01000100
GHIJKLMNOPQR IJKLMNOPQR
227959 HIJKLMNOPQ 00000101 HIJKLMNOPQR 01010010
%5EUWXYZAB STUVWXYZAB
154981 LMNOPQRSTU 01010110 LMNOPQRSTU 01000110
VWXYZABCDEF FWXYZABCDEF
204184 WXYZABCDE 01000111 WXYZABCDEF 01000110
FGHIJKLMNOPQ FHIJKLMNOPQ
Glitch Plaintext Around Plaintext Binary Decrypted text Around Decrypted text Binary
Found at Glitch Values at Glitched Glitch Values at Glitched
Index Position Position
195330 IJKLMNOPQR 1010011 IJKLMNOPQRCT 1000011
STUVWXYZABC UVWXYZABC
70726 WXYZABCDE 1100100 WXYZABCDEF 1000111
Fdt GHIJKLMNOPQ
928828 UVWXYZABC 1000101 UVWXYZABCD 1000110
DEFGHIJKLMNO FGHIJKLMNOP
923933 NOPQRSTUVW 1011000 NOPQRSTUVW 1111000
XYZABCDEFGH xYZABCDEFGH
36 x0019 00011001 x001D 00011101
Fig. 7. Fault Occurrence for Encryption & Decryption Processes for a 250 Kb String
Value
Electromagnetic Fault Injection Attack on ASCON Using ChipShouter 129
Fig. 8. Fault Occurrence for Encryption & Decryption Processes for a 500 Kb String
Value
Fig. 9. Fault Occurrence for Encryption & Decryption Processes for a 1 Mb String
Value
6 Conclusion
fault injection studies. This tool allowed us to precisely control the timing
and severity of fault injections, facilitating comprehensive testing of ASCON’s
resilience against EMFI assaults. Through iterative experimentation and analy-
sis, ASCON’s resilience can be continually evaluated.
The nature of the injected faults varied depending on the injection tips pro-
vided by the ChipShouter. Some tips introduced new bits, while others modified
existing ones. Interestingly, we observed a consistent result in the first register of
the memory upon fault injection using all injection tips. To evaluate the impact,
we tested different string sizes for input plaintext, namely 250 Kb, 500 Kb, and
1 MB. The results revealed that the fault injection percentages were as follows:
24% for the 250 Kb string size, 54% for the 500 Kb size, and 90% for the 1 MB
size.
Further work can expand our research by developing mitigation strategies
to defend against such attacks on ASCON. Previous works [8,9] have explored
dynamic key generation for the AES algorithm, employing additional logical and
bitwise procedures to generate all keys. This complexity makes it challenging to
inject faults into the key, offering potential avenues for enhancing ASCON’s
resistance to EMFI attacks.
References
1. Ordas, S., Guillaume-Sage, L., Maurine, P.: EM injection: fault model and locality.
In: 2015 Workshop on Fault Diagnosis and Tolerance in Cryptography (FDTC).
IEEE (2015)
2. Dehbaoui, A., et al.: Electromagnetic transient faults injection on a hardware and
a software implementations of AES. In: 2012 Workshop on Fault Diagnosis and
Tolerance in Cryptography. IEEE (2012)
3. Majéric, F., Bourbao, E., Bossuet, L.: Electromagnetic security tests for SoC.
In: 2016 IEEE International Conference on Electronics, Circuits and Systems
(ICECS). IEEE (2016)
4. Moro, N., et al.: Electromagnetic fault injection: towards a fault model on a 32-bit
microcontroller. In: 2013 Workshop on Fault Diagnosis and Tolerance in Cryptog-
raphy. IEEE (2013)
5. Beckers, A., et al.: (Adversarial) electromagnetic disturbance in the industry. IEEE
Trans. Comput. 72, 414–422 (2022)
6. Schellenberg, F., et al.: On the complexity reduction of laser fault injection cam-
paigns using OBIC measurements. In: 2015 Workshop on Fault Diagnosis and
Tolerance in Cryptography (FDTC). IEEE (2015)
7. O’Flynn, C.: MIN () imum failure: EMFI attacks against USB stacks. In: WOOT@
USENIX Security Symposium (2019)
8. Pilla, R., Jain, K.: A new authentication protocol for hardware-based authenti-
cation systems in an IoT environment. In: Smys, S., Kamel, K.A., Palanisamy,
R. (eds.) Inventive Computation and Information Technologies. Lecture Notes in
Networks and Systems, vol. 563, pp. 629–640. Springer, Singapore (2023). https://
doi.org/10.1007/978-981-19-7402-1 44
9. Saran, D.V.G., Jain, K.: An improvised algorithm for a dynamic key genera-
tion model. In: Smys, S., Kamel, K.A., Palanisamy, R. (eds.) Inventive Compu-
tation and Information Technologies. Lecture Notes in Networks and Systems, vol.
Electromagnetic Fault Injection Attack on ASCON Using ChipShouter 131
1 Introduction
Cardiovascular diseases (CVDs) are critical and common heart diseases that can
be detected using electrocardiogram (ECG or EKG) signals. The ECG signals are
used to diagnose different types of heart diseases such as heart failure, myocardial
infarction (MI), premature ventricular contractions (PVCs), etc. Analyzing the
bio-electrical signals of each heartbeat, cardiologists can detect abnormalities in
the heart, such as irregular heartbeats or abnormal rhythms. However, manual
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 135–150, 2023.
https://doi.org/10.1007/978-3-031-45878-1_10
136 U. T. Utsha et al.
scrutiny of continuous ECG signals for long durations for each patient is not
practical or feasible [1]. Thus, automated detection using machine learning (ML)
models is essential for accurate and efficient diagnosis of heart disease.
With the integration of IoT technology into heart disease monitoring, wear-
able devices and sHealth applications are gaining popularity [2]. These devices
and apps use sensors and algorithms to collect and analyze ECG signals in real
time, providing patients with immediate feedback and promoting better man-
agement of heart disease. In addition to ML algorithms, the integration of IoT
technology and sHealth applications has revolutionized the way we approach
cardiovascular health. KardiaMobile is a compact, portable electrocardiogram
(ECG) gadget that allows people to monitor their heart health and detect poten-
tial cardiac problems [3]. It works by recording a single-channel ECG through
two electrodes on the device’s back. It has been shown to be successful in clinical
investigations for identifying atrial fibrillation (AF), with a sensitivity of 96.6%
and a specificity of 94.1% [4]. An ECG check, like KardiaMobile, is a portable
electrocardiogram (ECG) equipment that employs two or more electrodes to
record the electrical activity of the heart [5]. The ECG check app transfers the
recorded ECG data to a server for processing. Apple Watch is another device
that can monitor the ECG signal and measure heart rate. The ECG feature,
which is available on Apple Watch Series 4 and later, enables users to record
an electrocardiogram, a test that examines the electrical activity of the heart.
The watch can detect aberrant cardiac rhythms like AF and alert the user if one
is identified [6]. Previously, our research team developed another smart health
framework using body-worn flexible Inkjet-printed (IJP) sensors, commercial
wearables such as smart wristbands, a scanner on a printed circuit board, and
customized smartphone software [7]. The technology to collect and analyze ECG
signals, providing patients with real-time feedback and enabling them to take
control of their heart health using wearable devices and smart health apps has
the potential to greatly improve the prevention, management, and treatment of
heart diseases.
The use of multi-stage classification has shown significant potential in tack-
ling the complexities of adjusting Artificial Intelligence (AI) models to novel
sensor data or in the evolution of decision-making methodologies in smart sys-
tems [8]. Segmenting the AI model into various stages enhances its scalability
and upgradability, offering a more flexible alternative compared to single-stage
classification that could potentially struggle in adapting to changes. Multi-stage
classification allows for independent modifications, enabling a more flexible and
adaptable approach. This approach has shown great potential in the detection
of cardiac diseases, demonstrating its scalability and upgradability for smart
health systems. With the integration of AI into cardiac disease detection, multi-
stage classification has emerged as a valuable tool in improving patient outcomes
and promoting heart health. Its benefits include increased accuracy, efficiency,
and the ability to adapt to changing data and circumstances. As such, multi-
stage classification has the potential to revolutionize the field of smart health
Real-time Cardiac Disease Detection by a Smart Health Application 137
and transform the way we approach cardiac disease detection and management
[9–15].
Fig. 1. The flowchart of the ECG signal processing, analysis, and classification.
2 Methodology
In this study, we utilized the MIT-BIH database, which follows the AAMI
criteria to classify heartbeat types [16]. The MIT-BIH database consists of
five categories of heartbeats, each containing multiple types of beats. Class N
Real-time Cardiac Disease Detection by a Smart Health Application 139
ECG (electrocardiogram) data is a sort of medical data that captures the electri-
cal activity of the heart. It is obtained by applying electrodes to the skin of the
chest, arms, and legs and connecting them to a device that records and amplifies
the electrical signals produced by the heart.
Signal processing and machine learning techniques are frequently used to ana-
lyze ECG data in order to extract diagnostic data and enhance clinical decision-
making. Arrhythmia, myocardial infarction, and heart failure are just a few of the
disorders that are frequently diagnosed and monitored using ECG data. Addi-
tionally, properties including heart rate variability, QT intervals, and P-wave
morphology can be extracted from it. We collected ECG data using an AD8232
chip (Analog Devices, Wilmington, MA) implemented on our custom ECG data
collection device [18]. Electrodes are attached to the wrists of the users and the
other part of the electrodes are connected to Sparkfun nRF 52840 mini that is
paired with the application via Bluetooth Low Energy (BLE) V5.3.
2.2 Application
Fig. 3. Snapshots from the application (a) Home (b) Dashboard (c) Available Devices
(d) Modules (e) Incoming Data (f) Real-time plot of ECG trace, Heart Rate and Signal
type.
1. Train and Save the model: First, we trained the model on a suitable
dataset and saved it in a format that can be loaded by TensorFlow Lite.
Here, we saved the model as a .h5 file using the Keras model.save() method.
2. Convert the model to TensorFlow Lite format: Next, we converted the
machine learning model to TensorFlow Lite format using the TensorFlow Lite
converter. This produced a .tflite file that we used in our Android application.
3. Add the model in the android application: We loaded the .tflite file in
the assets folder of the Android Studio project. Alternatively, we could go
to the File− >Other − >Tensorflow Lite Model and import the .tflite file. It
will be added in the ml folder on the project.
Real-time Cardiac Disease Detection by a Smart Health Application 141
Fig. 4. Real-time data collection (a) Normal Signals (b) Noisy Signals.
3 Results
3.1 Pre-trained Model
Accurately detecting R peaks is crucial in ECG heartbeat recognition. Figure 6
illustrates the P, Q, R, S, and T peaks, as well as a heartbeat. We utilized the
Pan-Tompkins algorithm to detect the precise position of the R peak, which
in turn affects the accurate positioning of the P, Q, S, T, and T’ peaks. The
algorithm identifies the R peak by employing a sliding window that spans two
heartbeats and advances one beat at a time until completion. The peaks detected
by the algorithm are then labeled as P, Q, R, S, and T, as demonstrated in Fig. 6,
which showcases the interval from 850 to 1100 at the apex.
Table 1. Evaluate the performance of machine learning (ML) and deep learning (DL)
methods on single-stage classification tasks by employing various classifier algorithms.
Single-stage Classification
Classifier Accuracy % Precision % Recall %
DT(10) 89 90 88
ANN(64) 94 92 93
SVM 78 78 79
Naive Bayes 81 83 82
KNN 75 75 74
Bagged tree 84 85 83
RNN 92 91 91
CNN 94 93 92
LSTM 90 89 87
In the machine learning domain, the Decision Tree (DT) classifier achieved
an accuracy of 89%, with precision and recall rates of 90% and 88%, respec-
tively. The Artificial Neural Network (ANN) exhibited an overall accuracy of
94%, accompanied by precision and recall scores of 92% and 93%. In the realm
of deep learning, the Convolutional Neural Network (CNN) demonstrated supe-
rior performance, obtaining an accuracy of 94% and precision and recall values
of 93% and 92%, respectively. Table 1 presents a comparative analysis of the per-
formance of various machine learning (ML) and deep learning (DL) techniques
in single-stage classification tasks.
Table 2 provides a summary of power consumption for single-stage classi-
fication using ANN and DT algorithms with varying parameters, such as the
Real-time Cardiac Disease Detection by a Smart Health Application 145
number of layers and maximum depth. For the ANN models, we explore perfor-
mance across a range of layer counts, from 1 to 256, with 64 layers as the stan-
dard configuration. Both the 128-layer and 256-layer ANN classifiers achieved
100% accuracy without significant changes in memory, CPU usage, or runtime.
Beyond 128 layers, the accuracy and power consumption decreased, while the
runtime became faster. For the DT models, we analyzed performance by varying
the maximum depth parameter from 1 to 25, with 10 as the standard setting.
The classifiers with a maximum depth of 25 and 24 achieved 100% accuracy,
again without noticeable changes in memory, CPU usage, or runtime. Beyond a
maximum depth of 24, the accuracy and memory usage decreased, while CPU
usage and runtime remained consistent.
Table 2. Power consumption for single-stage classification using ANN and DT algo-
rithms with varying parameters.
Single-stage Classification
Parameter Accuracy % Memory usage (MiB) CPU usage % Run time(s)
ANN(Layers)
256 100 431 4 20
128 100 431 4 18
64 94 429 3.5 12
32 82 427 3.5 9
16 67 424 3.0 9
8 52 423 3.0 7
4 49 420 2.0 7
2 33 417 2.0 5
1 23 417 2.0 3
DT(MaxDepth)
25 100 375 1.0 1
24 100 375 1.0 1
23 99 375 1.0 1
20 98 375 1.0 1
15 94 375 1.0 1
10 89 360 1.0 1
6 80 352 1.0 1
2 79 348 1.0 1
1 75 347 1.0 1
Table 3. Overview of the power consumption for multi-stage classification using ANN
and DT algorithms with different parameters and arrangements.
Multi-stage Classification
Parameter Accuracy% Memory usage (MiB) CPU usage % Run time(s)
First stage; Second stage First stage; Second stage
ANN(Layers)
256 100 ; 100 342 ; 389 17 42
128 100 ; 99 340 ; 387 16.5 35
64 100 ; 94 336 ; 387 7.5 20
32 98 ; 84 329 ; 375 14 15
16 95 ; 74 325 ; 368 13.7 11
8 87 ; 67 321 ; 365 13 11
4 74 ; 53 321 ; 357 12 10
2 68 ; 39 318 ; 357 11.5 7
1 67 ; 27 315 ; 351 16 5
DT(MaxDepth)
25 100 ; 100 384 ; 375 1.0 1
24 100 ; 100 383 ; 375 1.0 1
23 100 ; 99 383 ; 374 1.0 1
20 100 ; 94 382 ; 374 1.0 1
15 98 ; 88 383 ; 375 1.0 1
10 95 ; 79 382 ; 375 1.0 1
6 91 ; 53 383 ; 374 1.0 1
2 83 ; 42 383 ; 374 1.0 1
1 74 ; 36 382 ; 373 1.0 1
3.2 Application
The experiment involved validating the application’s functionality by simulating
real-time data reception from the embedded system. Instead of using actual
real-time data, we utilized an offline test dataset that contained labeled cardiac
disease data. This allowed us to assess the application’s ability to accurately
display the corresponding heart rhythms based on the provided labels.
As we don’t have any offline dataset containing the Noisy signal, we couldn’t
validate our pre-trained model 1 which is a binary classifier and classifies ECG
Real-time Cardiac Disease Detection by a Smart Health Application 147
signals into Noisy and Not Noisy signals (Fig. 1). But it can detect the Noisy
and Not Noisy signals in our application perfectly. We validated our pre-trained
model 2 using the MIT-BIH test dataset. It contains 21892 samples and 18118 of
them are Normal beats, 556 beats are Supraventricular Ectopic(SV), 1448 beats
are Ventricular Ectopic(V), 160 bears are Fusion(F) and the rest 1610 beats are
Unknown.
We have separated the MIT-BIH test dataset into five distinct sets. Each
set was passed to the Smart-Health application, which contained a pre-trained
model designed to classify heart diseases. We used Artificial Neural Network
(ANN) as a pre-trained model 2 because it exhibited an overall accuracy of 94%
which showed the best performance among ML models. The pre-trained model
classified each set based on the available ECG signals. After classification, the
labels obtained from the Smart-Health application were compared with the MIT-
BIH offline dataset. The accuracy was calculated for each set by comparing the
obtained labels from the Smart-Health application with the ground truth labels
from the offline dataset. The average accuracy across all five sets was computed to
evaluate the performance of the pre-trained model on the MIT-BIH test dataset.
Table 4 provides the accuracy of five distinct sets over the application. The
overall accuracy of the pre-trained ANN model was 91.178%.
The Smart-Health application detected 16955 Normal beats, 486 SV beats,
1247 V beats, 127 F beats and 1371 Unknown beats correctly. Figure 7 demon-
strates the beats of the MIT-BIH test dataset and the corresponding accurate
beats obtained from the Smart-Health application.
Arrhythmia refers to any abnormality in the rhythm of the heart’s electrical
activity. SVs, Vs, and fusion beats are all types of arrhythmias that can occur
in the heart. Figure 8 shows some snapshots of the Smart-Health application
after the detection of Normal, Noisy, and Arrhythmic beats. In some cases,
arrhythmias can be serious and lead to heart failure, stroke, or sudden cardiac
death.
148 U. T. Utsha et al.
4 Future Work
It is incredibly challenging to collect data when walking, jogging, or engaging
in any other action because of the complex system setup. There are some IoT
gadgets in the package that could be improved in the future, at which point we
could quickly attach those devices to the body and collect data. We are devel-
oping a custom wearable ECG data collection hardware, that can significantly
simplify the data collection process. Also, the accuracy of the ANN model on
the application can be improved. We will also try other algorithms which had a
lower accuracy on offline datasets but can perform well with real-time data.
Fig. 7. Comparison of beat classification between the MIT-BIH test dataset and Smart-
Health application.
Heart Rate. As we have passed the MIT-BIH test dataset through the applica-
tion to detect heart diseases and found quite a good accuracy, our next goals
are:
5 Conclusion
In this study, we presented Smart-Health, a smartphone application that can
continuously monitor ECG data, display Heart Rate and detect cardiac dis-
eases using a pre-trained machine-learning model. The MIT-BIH test dataset
was used to evaluate our model, and the findings suggest that our application
can accurately detect various heart conditions. Patients can use this application
to check their heart health in real time and take appropriate steps depending
on the results. Overall, the Smart-Health application has the potential to be a
valuable tool for the early detection and monitoring of cardiac problems. The
Smart-Health application can aid in the prevention and control of cardiovas-
cular illnesses, resulting in better health outcomes and a higher quality of life
for patients by empowering them to actively participate in their own health
management.
Acknowledgment. This material is based upon work supported by the National Sci-
ence Foundation under Grant No. 2105766. The development of the ECG device was
performed by Mahfuzur Rahman, Robert Hewitt, and Bashir I. Morshed.
References
1. Glovaci, D., Fan, W., Wong, N.D.: Epidemiology of diabetes mellitus and cardio-
vascular disease. Curr. Cardiol. Rep. 21, 1–8 (2019)
2. Walker, A., Muhlestein, J.: Smartphone electrocardiogram monitoring: current per-
spectives. Adv. Health Care Technol. 8(4), 15–24 (2018)
3. Koltowski, L., et al.: Kardia mobile applicability in clinical practice: a comparison
of Kardia mobile and standard 12-lead electrocardiogram records in 100 consecutive
patients of a tertiary cardiovascular care center. Cardiol. J. 28(4), 543–8 (2021)
4. William, A.D., et al.: Assessing the accuracy of an automated atrial fibrilla-
tion detection algorithm using smartphone technology: the iREAD study. Heart
Rhythm 15(10), 1561–5 (2018)
5. Haverkamp, H.T., Fosse, S.O., Schuster, P.: Accuracy and usability of single-lead
ECG from smartphones-a clinical study. Indian Pacing Electrophysiol. J. 19(4),
145–9 (2019)
6. Raja, J.M., et al.: Apple watch, wearables, and heart rhythm: where do we stand?
Ann. Transl. Med. 7(17) (2019)
7. Rahman, M.J., Morshed, B.I., Harmon, B., Rahman, M.: A pilot study towards
a smart-health framework to collect and analyze biomarkers with low-cost and
flexible wearables. Smart Health 1(23), 100249 (2022)
150 U. T. Utsha et al.
1 Introduction
measured the accuracy of detected AoA and convergence time, by tuning both
the learning rate α and exploration policy . Further, the algorithms take a cer-
tain number of unknown iterations to converge. To tackle this problem, we use
a threshold on the coefficient of variation (CoV) of RSS data samples as the
criteria to detect convergence. We compare the performance of our algorithms
using Horn antennas controlled by a Pyhton-based SDR setup in connection with
GNU radio [15].
Our main contributions are as follows:
The rest of the paper is organized as follows: Sect. 2 surveys the related
literature on AoA detection and experimental mmWave SDR efforts. Section 3
presents our experimental platform and how the AoA detection algorithms are
implemented in that platform. Next, Sect. 4 provides a detailed description of
our Q-learning and Double Q-learning algorithms for AoA detection. Section 5
details experimental setup and discusses results from our experiments. Finally,
Sect. 6 summarizes our work and outlines directions of future work.
2 Related Work
Angle (or direction) of arrival (AoA) detection/estimation has been an exten-
sively studied problem within the context of wireless localization [16]. With the
recent advent of directional beam-forming capabilities in super-6 GHz systems,
AoA detection, in particular, has gained a renewed interest due to emerging
applications using such systems [17].
Experimental demonstration and evaluation of AoA detection in super-6 GHz
bands such as mmWave bands has been lacking. The main reason for this has
been the limited availability of mmWave experimental testbeds due to the lack
and high cost of mmWave hardware [18]. The U.S. National Science Founda-
tion (NSF) is currently funding wireless communication testbed platforms to
enable such experimentation. The COSMOS platform [19], for example, includes
a 28 GHz phased array antenna, designed by IBM. The front end uses a custom
software for steering the antenna beam with respect to azimuth and elevation
angles. The AERPAW platform uses drones for 5G experimentation [20], which
is the first of its kind. These platforms enable users to perform a variety of wire-
less communication experiments, such as, AoA detection. However, they are still
being adapted by researchers. Unlike these high-end testbeds, we use a cheap
154 M. Jean and M. Yuksel
SDR platform and mmWave hardware to evaluate our AoA detection mech-
anisms. Further, the application programming interface (API) used by these
testbed platforms can limit user experimentation. For example, the AERPAW
API restricts users from running on the fly experiments. As a result, users aren’t
able to collect or train radio frequency (RF) data on the fly. This can restrict
the types of algorithms users can use on the platform.
Researchers have relied on virtual environments and simulations to perform
mmWave experiments. These virtual environments have gotten more sophisti-
cated with the usage of 3D ray tracing. In [21], 3D ray tracing is used to sim-
ulate mmWave signals in virtual environments. Users can use the open source
software to design large intelligent reflective surfaces and determine AoA using
compressive sensing algorithms. Although using simulation-based approaches is
cost effective, they do not render the physical world and fall short of precisely
modeling complicated physical communication channel dynamics in mmWave or
other super-6 GHz bands.
Recently, cheaper off-the-shelf SDRs have been used to setup testbed plat-
forms for AoA detection. The testbed platform [9] uses a Kerberos radio with
four whip antennas at the receiving end. At the transmitting end a long range
(LoRa) radio is used to transmit a signal at 826 MHz. LoRa is beneficial for long
range communication and uses low transmit power. The transmitter includes
a GPS and compass unit used to label the direction of the transmitted signal.
The labeled data set is the ground truth that is trained in the machine learning
(ML) algorithm. The data is trained using a deep learning convolutional neural
network (CNN) model [9].
Multiple Signal Classification (MUSIC) is a widely used AoA detection algo-
rithm and assumes that the received signal is orthogonal to the noise signal [9].
MUSIC uses this assumption to decompose the noise from the received signal into
separate sub-spaces. The power spectral density (PSD) of the signal is taken as a
function of angle [22]. The angular value which results in the maximum power is
estimated to be the AoA. The assumption that the received signal is orthogonal
to the noise signal is not valid in real world scenarios. Therefore, MUSIC does
poorly in environments that involve NLoS propagation. Since mmWave signals
can experience severe environmental path loss and involve multiple NLoS signals,
MUSIC may not be a good choice for mmWave AoA detection.
Support Vector Regression (SVR) has also been used to estimate AoA. SVR
is a supervised ML algorithm. Regression does poorly in estimating AoA from
impinging signals at multiple time steps [9]. The algorithm cannot be used to
determine AoA since the number of impinging signals is unknown [23]. As a
result, the algorithm can be used for detecting AoA for a single source at a time.
This makes SVR less robust for AoA detection in environments with multiple
signal sources. Therefore, SVR is not a good choice for mmWave AoA detection.
The CNN model used in [9] adapts a hybrid configuration. A classification
method is used to determine the number of impinging receive signals and two
regressive heads are used to determine the AoA. The study showed that CNN
RL-Based AoA Detection for MmWave SDR Systems 155
outperformed the other classical ML methods, MUSIC and SVR. Further, the
CNN model was able to estimate AoA within 2o ’s of accuracy.
Our approach does not use a deep learning approach or supervised learning.
These approaches are not the most suitable for many hardware-constrained IoT
devices as the former requires large memory hardware to perform well and the
latter requires availability of ground truth. Resources-constrained IoT devices
like wearables do not have sufficient memory to keep trained models nor the
extensive sensing or coordination capability to obtain the ground truth in AoA.
To make it more relevant to IoT devices with high resource constraints, we
design RL-based AoA detection methods that do not require the ground truth
and determine the AoA based only on the RSS observations at the receiver.
3 mmWave Testbed
To perform a thorough evaluation of our reinforcement learning (RL)-based AoA
detection methods, we use our mmWave testbed [24] that allows beam-steering
capability from Python.
LO Servo
Arduino
USRP N210
PC
USRP daughter-board. The receive signal is mixed down to 2 GHz and injected
into the daughter-board. The I and Q base-band data samples are streamed from
the USRP source block. The samples are used to determine the RSS using the
complex to mag-square block. The RSS samples are streamed into a socket using
the ZMQ pub sink block. Python socket libraries are used to connect and receive
the RSS data from the socket.
Acon:
Turn Right or Le
State:
(0 -180)
Fig. 3. RL Configuration
Q-learning uses the Bellman equation to populate the Q-table. The Bellman
equation is
the action that results in the largest Q value. Increasing increases the likelihood
that the algorithm will explore the environment, by selecting random actions.
Our software approach is presented in Algorithm 1. At the start of each
run the hyper-parameters γ, α, and are initialized. The current state st is
initialized to a random angular value, based off our experimental scenario. For
the 90◦ experimental setup the random angle can be any value in [80,100] degrees
and [120,140] degrees for the 130o . The Qtable used to store the Q values from
Eq. 1 is initialized to zero. An outer while loop comparing the threshold value
and CoV is used to decide the stopping condition for the algorithm until the
threshold value is met. If the value of is greater than a random number in
(0,1), the agent takes a random action, either turn the antenna using the servo
left or right with one degree resolution. If the value of is greater than the
random value, then the algorithm will exploit the Qtable by selecting the action
that results in the largest Q value. A positive reward value is given if ΔRSS
is greater than zero. If ΔRSS is less than zero, then the action resulted in a
decrease in RSS. Therefore, the action yields a negative reward value of -5. The
Bellman equation is updated with the reward at every iteration.
Q-learning based algorithms have to train for a certain number of iterations.
Based on the number of iterations, the algorithm may or may not converge to the
solution. This makes selecting the number of iterations trivial. To address this
issue, we take CoV of a certain number of RSS samples (defined by windowSize)
to detect convergence. CoV is a statistical measure of how dispersed data samples
are from the mean of the sample space. It is the ratio of standard deviation
and mean of a certain number of samples. In our algorithm, a window size is
initialized to windowSize ∈ (5, 30). The array RSS is used to store RSS samples
as the algorithm is training on the fly. If the counter sampleCount is greater or
equal to the windowSize, then enough samples have been collected to calculate
the CoV. While the CoV is greater or equal to the selected threshold value, the
algorithm will continue to train. When CoV is less then or equal to the threshold,
the convergence criteria is met and the algorithm breaks out of the while loop.
The smaller CoV means that the RSS samples have stabilized and it is safer to
stop the algorithm and return the last steering angle as the detected AoA.
Q-learning uses a single estimator maxa Q(st+1 , a), the maximum next state
Qt+1 value for all possible actions. As shown in [12], this causes the algorithm
to overestimate the desired solution, by causing a positive bias. As a result,
standard Q-learning can perform poorly in certain stochastic environments. To
improve the performance of standard Q-learning other variants, such as, Dou-
ble Q-learning [12], Delayed Q-learning [28], Fitted Q-iteration [29], and Phased
Q-learning [30] were developed to improve convergence time. To reduce over-
estimation, the double Q-learning variant uses two Q functions QA (st , at ) and
QB (st , at ) seen in Eqs. 2 and 3 below. QA (st , at ) is able to learn from
QA (st , at ) = QA (st , at ) + α ∗ (rt + γQB (st+1 , arg max QA (st+1 , a)) − QA (st , at )) (2)
a
QB (st , at ) = QB (st , at ) + α ∗ (rt + γQA (st+1 , arg max QB (st+1 , a)) − QB (st , at )) (3)
a
160 M. Jean and M. Yuksel
the experiences of QB (st , at ) and vice versa. This approach has been shown
in [12] to cause the algorithm to underestimate, rather than overestimate the
solution with a positive bias. In conditions where Q-learning performs poorly,
double Q-learning has been shown to converge to the optimum solution [12].
RL-Based AoA Detection for MmWave SDR Systems 161
Our AoA detection algorithm for Double Q-learning can be seen in Algorithm
2. Like Algorithm 1, the same parameters are initialized at the start of each run.
Two Q tables QA table and QB table are initialized to zero. The same greedy
and reward rt values are also used in Algorithm 2. The variable q is set equal to
a random uniform value q ∈ (0, 1). If q is larger than the threshold value of 0.5
then QA (st , at ). Otherwise, if q is less than the threshold of 0.5, then QB (st , at ) is
selected. The threshold value is set to 0.5 to give both QA (st , at ) and QB (st , at )
equal probability of being selected for training. The same convergence criteria is
used in Algorithm 1 is used in 2.
min < AoA Error > ∗ < Convergence Time > . (4)
,α
first experiment scenario, the transmit antenna is fixed to 90◦ , pointing towards
the receive antenna to compose an LoS path to the receiver. The receive horn
antenna is initialized to a random angular state value between 80 and 100◦ C.
For the second scenario, the transmit horn antenna is rotated 130◦ to the right,
composing a NLoS path to the receiver. The main lobe of the transmitted signal
is reflected off the wall. The initial angle is set to a random value between 120
and 140 ◦ C.
As previously mentioned, the hyper-parameters and α are tuned to 0.1,
0.2, 0.3, and 0.4. This was done to measure which combinations of (,α) resulted
in the best performance with respect to both AoA detection and convergence
time. The average AoA error and time of convergence was measured for thirteen
threshold values from 0.1 to 0.4 in increments of 0.025. This was done using Q
and double Q-learning for both the 90◦ and 130◦ experimental scenarios.
Figure 5 are graphs of Threshold vs. average AoA error and Threshold vs.
average convergence time. For each combination of (, α) the average convergence
time is reduced with respect to window size. A window size of 10 results in faster
AoA detection compared to the window size of 20.
RL-Based AoA Detection for MmWave SDR Systems 163
Both the 90◦ and 130◦ scenarios resulted in average AoA error less than 1%
for some combination of (, α). This occurs for both window sizes of 10 and 20.
With less than 1% AoA error, the detected AoA is within 1◦ of the correct AoA.
To understand which (,α) combinations give the best results, we look at
the heat map of (α, ) occurrences resulting in the minimum product in Eq. 4
as shown in Fig. 6. For the 90◦ case, of 0.4 with α of 0.1, 0.2, and 0.4 are the
dominate cases. Likewise, of 0.4 is also the dominant case for the 130◦ scenario.
The α values are also similar at 0.2 and 0.4 being the dominant cases. For Q-
learning the combination of (,α) that are most effective are (0.4,0.1), (0.4,0.2)
and (0.4,0.4).
Figure 8 shows the heat map of (α, ) occurrences resulting in the minimum
product in Eq. 4 for Double Q-learning. For the 90◦ scenario the (, α) combi-
nations that occur the most often are (0.1,0.1), (0.1,0.3) and (0.1,0.4) for both
window sizes. Like the 90◦ case, it can also be seen that an of 0.1 does occur
for the 130◦ scenario for a window size of 10. However, we also see that values
of 0.3 and 0.4 do occur for both window sizes. The dominant (, α) are (0.1,0.1),
(0.1,0.2), (0.2,0.3), (0.3,0.2), (0.4,0.2), and (0.4,0.3).
Acknowledgment. This was was supported in part by the U.S. National Science
Foundation award 1836741.
References
1. Seth, S., Yuksel, M., Vosoughi, A.: Forming coalition sets from directional radios.
In: MILCOM 2022–2022 IEEE Military Communications Conference (MILCOM),
pp. 507–514 (2022)
2. Jilani, S.F., Munoz, M.O., Abbasi, Q.H., Alomainy, A.: Millimeter-wave liquid crys-
tal polymer based conformal antenna array for 5G applications. IEEE Antennas
Wirel. Propag. Lett. 18(1), 84–88 (2019)
166 M. Jean and M. Yuksel
3. Dhruva, T., Rajesh, A., Abirami, K.: Design of conformal multi-band mmWave
wearable antenna for 5G applications. In: Proceedings of IEEE International Con-
ference on Smart Electronics and Communication (ICOSEC), pp. 573–577 (2022)
4. Hu, K., Zhou, Y., Sitaraman, S.K., Tentzeris, M.M.: Fully additively manufactured
flexible dual-band slotted patch antenna for 5G/mmWave wearable applications.
In: Proceedings of IEEE International Symposium on Antennas and Propagation
and USNC-URSI Radio Science Meeting (AP-S/URSI), pp. 878–879 (2022)
5. Bogale, T., Wang, X., Le, L.: Chapter 9 - mmwave communication enabling tech-
niques for 5g wireless systems: a link level perspective. In: Mumtaz, S., Rodriguez,
J., Dai, L. (eds.) mmWave Massive MIMO, pp. 195–225. Academic Press (2017).
https://www.sciencedirect.com/science/article/pii/B9780128044186000091
6. Jean, M., Velazquez, E., Gong, X., Yuksel, M.: A 30 ghz steerable patch array
antenna for software-defined radio platforms. SoutheastCon 2023, 856–860 (2023)
7. Fan, W., Xia, Y., Li, C., Huang, Y.: Channel tracking and aoa acquisition in
quantized millimeter wave MIMO systems. IEEE Trans. Vehicular Technol., 1–15
(2023)
8. Yazdani, H., Seth, S., Vosoughi, A., Yuksel, M.: Throughput-optimal d2d mmwave
communication: Joint coalition formation, power, and beam optimization. In: 2022
IEEE Wireless Communications and Networking Conference (WCNC), pp. 1539–
1544 (2022)
9. Dai, Z., He, Y., Vu, T., Trigoni, N., Markham, A.: Deepaoanet: learning angle of
arrival from software defined radios with deep neural networks (2021)
10. Chowdhury, M.M.U., Erden, F., Guvenc, I.: Rss-based q-learning for indoor uav
navigation. In: MILCOM 2019–2019 IEEE Military Communications Conference
(MILCOM), pp. 121–126 (2019)
11. Watkins, C., Dayan, P.: Technical note: Q-learning. Mach. Learn. 8, 279–292 (1992)
12. Van Hasselt, H.: Double q-learning, pp. 2613–2621 (Jan 2010)
13. Jaakkola, T., Jordan, M.I., Singh, S.P.: On the convergence of stochastic iterative
dynamic programming algorithms. Neural Comput. 6(6), 1185–1201 (1994)
14. Tsitsiklis, J.: Asynchronous stochastic approximation and q-learning. In: Proceed-
ings of 32nd IEEE Conference on Decision and Control, vol. 1, pp. 395–400 (1993)
15. GNU Radio - the free and open source radio ecosystem, GNU Radio. https://www.
gnuradio.org
16. Zafari, F., Gkelias, A., Leung, K.K.: A survey of indoor localization systems and
technologies. IEEE Commun. Surv. Tutorials 21(3), 2568–2599 (2019)
17. Wu, K., Ni, W., Su, T., Liu, R.P., Guo, Y.J.: Recent breakthroughs on angle-of-
arrival estimation for millimeter-wave high-speed railway communication. IEEE
Commun. Mag. 57(9), 57–63 (2019)
18. Şahin, A., Sichitiu, M.L., Guvenç, İ.: A Millimeter-Wave Software-Defined Radio
for Wireless Experimentation, arXiv:2302.08444, (Feb. 2023)
19. Gu, X.: A multilayer organic package with 64 dual-polarized antennas for 28ghz
5g communication. In: 2017 IEEE MTT-S International Microwave Symposium
(IMS), pp. 1899–1901 (2017)
20. Aerpaw: Aerial experimentation and research platform for advanced wireless.
https://aerpaw.org/
21. Alkhateeb, A.: Deepmimo: a generic deep learning dataset for millimeter wave and
massive MIMO applications. CoRR, vol. abs/ arXiv: 1902.06435, (2019)
22. Kaveh, M., Barabell, A.: The statistical performance of the music and the
minimum-norm algorithms in resolving plane waves in noise. IEEE Trans. Acoust.
Speech Signal Process. 34(2), 331–341 (1986)
RL-Based AoA Detection for MmWave SDR Systems 167
23. Pastorino, M., Randazzo, A.: A smart antenna system for direction of arrival esti-
mation based on a support vector regression. IEEE Trans. Antennas Propag. 53,
2161–2168 (2005)
24. Jean, M., Yuksel, M., Gong, X.: Millimeter-wave software-defined radio testbed
with programmable directionality. In: Proceedings IEEE INFOCOM Workshop
2023 (in press)
25. UBX 10–6000 MHz Rx/TX (40 MHz, N series and X series), Ettus Research, a
National Instruments Brand. https://www.ettus.com/all-products/ubx40
26. ADMV1013, Analog Devices. https://www.analog.com/en/products/admv1013.
html
27. ADMV1014, Analog Devices. https://www.analog.com/en/products/admv1014.
html
28. Kearns, M., Singh, S.: Finite-sample convergence rates for q-learning and indirect
algorithms. In: Proceedings of the 1998 Conference on Advances in Neural Infor-
mation Processing Systems II, pp. 996–1002. MIT Press, Cambridge, MA, USA
(1999)
29. Ernst, D., Geurts, P., Wehenkel, L.: Tree-based batch mode reinforcement learning.
J. Mach. Learn. Res. 6, 503–556 (2005)
30. Szepesvári, C.: The asymptotic convergence-rate of q-learning. In: Proceedings of
the 10th International Conference on Neural Information Processing Systems, ser.
NIPS 1997, pp. 1064–1070. MIT Press, Cambridge, MA, USA (1997)
Empowering Resource-Constrained IoT
Edge Devices: A Hybrid Approach
for Edge Data Analysis
1 Introduction
In recent years, IoT has witnessed tremendous growth and widespread adoption,
resulting in a substantial increase in data generation at the edge of the network.
This data surge is primarily attributed to the diverse array of sensors, actua-
tors, and embedded systems integrated into edge devices [12]. However, these
devices are often constrained in terms of computational resources. This constraint
presents a significant challenge when attempting to deploy resource-intensive ML
algorithms directly on these devices. The limitations in processing power and
memory capacity and the crucial requirement for energy efficiency have high-
lighted the pressing need for the development of hardware-friendly ML models
[8,9,11] specifically designed to facilitate effective decision-making at the edge.
The advent of IoT has brought about a significant transformation in con-
ventional computing approaches, emphasizing the transition from centralized
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 168–181, 2023.
https://doi.org/10.1007/978-3-031-45878-1_12
Empowering Resource-Constrained IoT Edge Devices: A Hybrid Approach 169
design and results along with its discussion. Finally, we present the conclusion
and potential future perspectives in Sect. 5.
sensitive to the scale of the features, and data normalization is typically per-
formed prior to applying PCA to ensure fair comparisons and accurate results.
Overall, PCA serves as a valuable tool for dimensionality reduction and feature
extraction to handle high-dimensional data efficiently and gain deeper insights
from complex datasets.
Decision Tree. DTs are ML techniques that are used for solving classification
and regression problems. They represent decision-making processes through a
hierarchical structure comprising nodes and branches. Figure 1 shows the clas-
sification of the Iris dataset using DT. At each node, the optimal split is deter-
mined using metrics like Gini impurity or entropy, aiding in the creation of
distinct branches [17]. To address the issue of overfitting, pruning techniques
can be employed to simplify the tree’s structure. One of the notable advan-
tages of decision trees is their ability to handle both categorical and numerical
data, making them versatile for a range of applications. Furthermore, ensemble
approaches such as random forest and gradient boosting can be employed to
enhance the performance of decision trees by combining multiple trees. While
decision trees offer interpretability, they can be sensitive to fluctuations in the
data. Moreover, there exist extensions and modifications to decision trees, such
as decision stumps (shallow trees) and advanced tree-based algorithms like XG-
Boost, which further augment their capabilities.
Ganaie et al. [5] proposed a method that utilizes twin-bounded support vector
machines (TBSVM) to create oblique decision trees. These trees employ ensem-
ble techniques and clustering hyperplanes to effectively separate data points into
distinct groups. Ajani et al. [2], conducted a comprehensive analysis on embed-
ded machine learning (EML), focusing on compute-intensive algorithms like k-
nearest neighbors (k-NNs), support vector machines (SVMs), and deep neural
networks (DNNs). They explored optimization techniques tailored for resource-
Empowering Resource-Constrained IoT Edge Devices: A Hybrid Approach 173
3 Proposed Work
The workflow for this hybrid approach incorporating PCA, DT, and SVM
for dimensionality reduction and classification for a resource-efficient inference
model hardware implementation for IoT edge devices consists of several key
steps:
4 Experimental Results
This section outlines the experimental design and results of our proposed
method. In order to gauge the efficacy of the proposed methodology, a compre-
hensive evaluation is conducted utilizing five well-known classification dataset
benchmarks, including Heart Disease [4], Breast Cancer Wisconsin (Diagnostic)
[20], Lung Cancer [1], Fetal Health [3], and Pima Indian Diabetes datasets [16].
The Heart Disease dataset comprises 1,025 instances and 10 input features. The
Breast Cancer Wisconsin (Diagnostic) dataset consists of 30 input features and
569 instances. The dataset pertaining to Lung Cancer comprises 309 instances
and 15 input features. Similarly, the dataset on Fetal Health comprises 21 input
features and 2,126 instances, and the Pima Indian Diabetes dataset consists of
11 input features and 767 instances. The experimental flow involves training
the classification datasets using our proposed hybrid workflow. The training and
testing data are randomly partitioned in an 80:20 ratio to generate the hybrid
model. Subsequently, the performance of the hybrid model is evaluated through
performance analysis in terms of accuracy.
hyperparameters of a hybrid model consisting of PCA, DT, and SVM. The objec-
tive is to identify the optimal combination of hyperparameters that produces the
best performance for each of these algorithms. Specifically, for PCA, we focus
on fine-tuning hyperparameters such as the selection of the optimal number of
orthogonal components that effectively capture the most significant variance in
the input features. This allows us to significantly reduce the dimensions of the
input features by condensing as much information as possible from the input
features into a smaller subset of transformed features, which are then utilized
as input for DT and SVM models. This reduction in dimensionality also leads
to more efficient memory utilization and computational resources, making it
suitable for resource-constrained IoT edge devices with limited processing power
and storage capacity. Moreover, it results in lower energy consumption during the
implementation of the hybrid inference model on IoT edge devices as well as helps
to minimize the computational workload and data movement, enabling real-time
and low-latency predictions, which are crucial for time-sensitive IoT applications.
Table 1 presents the optimized hybrid model input configurations obtained after
applying PCA. For the Heart Disease dataset, we observe a reduction in the
number of features from the original 10 to 3 transformed features, resulting in a
significant 70% reduction in the dimensionality of the input features. Similarly,
for the Breast Cancer dataset, there is a 60% reduction in dimensionality as the
number of input features decreases from 30 to 12 transformed features. The Lung
Cancer, Fetal Health, and Pima Indian Diabetes datasets also demonstrate sub-
stantial reductions in dimensionality, with percentage reductions of 53%, 52%,
and 55% respectively, in their respective input feature sets.
Subsequently, we proceed with fine-tuning the hyperparameters of DT and
SVM in our hybrid model. Given our objective of creating a robust hybrid model
specifically designed for IoT applications with hardware efficiency in mind, we
meticulously approach the process of selecting hyperparameters for DT and
SVM. In the case of DT, we pre-prune the DT prior to training by determin-
ing the optimal hyperparameters as follows: The criterion for DT is set as Gini
impurity, considering its computational efficiency compared to entropy and its
tendency to produce shorter and more cohesive branches. The maximum depth
of the decision tree is set to 10, limiting the depth to control overfitting and
model complexity. The minimum sample split is set to 2, ensuring that a split at
a node requires at least 2 samples, preventing further divisions with insufficient
data. The splitter strategy is selected as “best,” indicating that the best-split
strategy is chosen at each node during the tree construction process. Other
hyperparameters are left as default values. Similarly, we carefully determine the
hyperparameters for SVM as follows: To optimize hardware implementation and
efficiency, we select the sigmoid kernel for SVM, which offers simpler mathemat-
ical operations compared to more complex kernels like Gaussian (RBF). This
choice facilitates easier implementation and optimization in hardware architec-
tures, resulting in reduced computational complexity and memory requirements.
Consequently, our approach reduces hardware resource utilization, energy effi-
ciency, and execution time. Specifically, we set the regularization parameter (C)
178 R. Joshi et al.
Table 4. Performance evaluation of DT and SVM models for training and testing
DT SVM
Training Testing Training Testing
Dataset Accuracy (%) Accuracy (%) Accuracy (%) Accuracy (%)
Heart Disease 99.14 98.50 98.00 96.08
Breast Cancer 99.99 98.79 99.58 94.26
Lung Cancer 99.71 97.74 98.00 96.45
Fetal Health 99.99 97.00 99.99 97.00
Pima Indian Diabetes 99.95 99.00 96.00 93.53
5 Conclusions
In this work, we present a hybrid approach that integrates PCA, DT, and SVM
for deployment in IoT edge devices. The experimental findings demonstrate the
effectiveness of this approach in enhancing the performance of ML inference mod-
els for IoT edge applications. By integrating these contemporary ML techniques,
the proposed approach achieves improved accuracy while effectively addressing
computational and memory limitations of resource-constrained IoT edge devices.
In the future, we are interested to focus on further optimizing the hybrid model
for specific IoT edge device architectures and exploring additional feature selec-
tion and classification techniques to improve performance and efficiency.
180 R. Joshi et al.
References
1. Ahmad, A.S., Mayya, A.M.: A new tool to predict lung cancer based on risk factors.
Heliyon 6(2), e03402 (2020)
2. Ajani, T.S., Imoize, A.L., Atayero, A.A.: An overview of machine learning within
embedded and mobile devices-optimizations and applications. Sensors 21(13), 4412
(2021)
3. Ayres-de Campos, D., Bernardes, J., Garrido, A., Marques-de Sa, J., Pereira-Leite,
L.: SisPorto 2.0: a program for automated analysis of cardiotocograms. J. Maternal-
Fetal Med. 9(5), 311–318 (2000)
4. Detrano, R.: UCI Machine Learning Repository: Heart Disease Data Set (2019)
5. Ganaie, M., Tanveer, M., Suganthan, P.N.: Oblique decision tree ensemble via twin
bounded SVM. Expert Syst. Appl. 143, 113072 (2020)
6. Hearst, M.A., Dumais, S.T., Osuna, E., Platt, J., Scholkopf, B.: Support vector
machines. IEEE Intell. Syst. Appl. 13(4), 18–28 (1998)
7. Joshi, R., Zaman, M.A., Katkoori, S.: Novel bit-sliced near-memory computing
based VLSI architecture for fast Sobel edge detection in IoT edge devices. In: 2020
IEEE International Symposium on Smart Electronic Systems (iSES) (Formerly
iNiS), pp. 291–296. IEEE (2020)
8. Joshi, R., Kalyanam, L.K., Katkoori, S.: Simulated annealing based integerization
of hidden weights for area-efficient IoT edge intelligence. In: 2022 IEEE Inter-
national Symposium on Smart Electronic Systems (iSES), pp. 427–432 (2022).
https://doi.org/10.1109/iSES54909.2022.00093
9. Joshi, R., Kalyanam, L.K., Katkoori, S.: Area efficient VLSI ASIC implementation
of multilayer perceptrons. In: 2023 International VLSI Symposium on Technology,
Systems and Applications (VLSI-TSA/VLSI-DAT), pp. 1–4. IEEE (2023)
10. Joshi, R., Zaman, M.A., Katkoori, S.: Fast Sobel edge detection for IoT edge
devices. SN Comput. Sci. 3(4), 302 (2022)
11. Kalyanam, L.K., Joshi, R., Katkoori, S.: Range based hardware optimization of
multilayer perceptrons with RELUs. In: 2022 IEEE International Symposium on
Smart Electronic Systems (iSES), pp. 421–426 (2022). https://doi.org/10.1109/
iSES54909.2022.00092
12. Laghari, A.A., Wu, K., Laghari, R.A., Ali, M., Khan, A.A.: A review and state of
art of internet of things (IoT). Arch. Comput. Methods Eng. 1–19 (2021)
13. Lee, L.C., Jemain, A.A.: On overview of PCA application strategy in processing
high dimensionality forensic data. Microchem. J. 169, 106608 (2021)
14. Shanthamallu, U.S., Spanias, A., Tepedelenlioglu, C., Stanley, M.: A brief survey
of machine learning methods and their sensor and IoT applications. In: 2017 8th
International Conference on Information, Intelligence, Systems and Applications
(IISA), pp. 1–8. IEEE (2017)
15. Shoaran, M., Haghi, B.A., Taghavi, M., Farivar, M., Emami-Neyestanak, A.:
Energy-efficient classification for resource-constrained biomedical applications.
IEEE J. Emerg. Sel. Top. Circuits Syst. 8(4), 693–707 (2018)
16. Smith, J.W., Everhart, J.E., Dickson, W., Knowler, W.C., Johannes, R.S.: Using
the ADAP learning algorithm to forecast the onset of diabetes mellitus. In: Pro-
ceedings of the Annual Symposium on Computer Application in Medical Care, pp.
261. American Medical Informatics Association (1988)
17. Somesula, R.S., Joshi, R., Katkoori, S.: On feasibility of decision trees for edge
intelligence in highly constrained internet-of-things (IoT). In: Proceedings of the
Great Lakes Symposium on VLSI 2023, pp. 217–218 (2023)
Empowering Resource-Constrained IoT Edge Devices: A Hybrid Approach 181
18. Struharik, R.: Decision tree ensemble hardware accelerators for embedded appli-
cations. In: 2015 IEEE 13th International Symposium on Intelligent Systems and
Informatics (SISY), pp. 101–106. IEEE (2015)
19. Wang, H., Barriga, L., Vahidi, A., Raza, S.: Machine learning for security at the IoT
edge-a feasibility study. In: 2019 IEEE 16th International Conference on Mobile
Ad Hoc and Sensor Systems Workshops (MASSW), pp. 7–12. IEEE (2019)
20. Wolberg, W., Street, W., Mangasarian, O.: Breast cancer wisconsin (diagnostic).
UCI Machine Learning Repository (1995)
21. Yong, B., et al.: Ensemble machine learning approaches for Webshell detection in
internet of things environments. Trans. Emerg. Telecommun. Technol. 33(6), e4085
(2022)
Energy-Efficient Access Point
Deployment for Industrial IoT Systems
1 Introduction
attention from both academia and industry. The large amount of energy con-
sumption also poses challenges to the environment, as renewable green energy is
typically not used as a power source for wireless networks [2].
The energy consumption attributable to Information and Communication
Technology (ICT) has exhibited large increases with the advent of new technolo-
gies, such as Fifth Generation (5G) and Multiple-Input and Multiple-Output
(MIMO), as such technologies require more power consumption to increase
response speed and accommodate more users. Therefore, innovation in green
communications technologies is in urgent need.
To help address this need, we propose and demonstrate a systematic Access
Point (AP) deployment strategy for energy-efficient IIoT systems. The remain-
der of this paper is organized as follows. Section 2 discusses background and
related literature about AP deployment strategies. Section 3 explains the pro-
posed energy-efficient AP deployment strategy. Section 4 reviews the factory sys-
tem flow model used in our simulation experiments. Section 5 presents the results
of the proposed strategy, which are obtained from the aforementioned simulation
experiments. Finally, the paper is concluded in Sect. 6.
3 Proposed Methods
In the problem formulation that is addressed in this work, the energy cost to
be optimized refers to the energy consumed by communication activities that
occur during normal operation of the IIoT system. Specifically, the problem
definition targeted in this work is the optimization of communication energy
given a placement of networked devices, which may be unevenly distributed,
and a characterization of the traffic demand for each device.
The communication energy considered in this paper refers to the transmis-
sion energy. Energy associated with communication reception for the devices is
not taken into account in the methods developed in the paper, as it is common
in related analysis contexts to focus on transmission power, and consideration
of the transmission energy provides an approximation of the overall energy con-
sumption due to communication. Incorporation of models for reception energy
into the developments of this paper is an interesting direction for future work.
To save energy, we consider optimal placement of the APs so that they can
deliver packets to all stations (STAs) in an area with an appropriate transmit
power according to their activity rates. The fitness function can be mathemati-
cally expressed as follows:
Energy-Efficient Access Point Deployment for Industrial IoT Systems 185
n
s
min 1i,j αi,j ti,j βi,j (1)
xi ,yi ,αi,j
i=1 j=1
s.t. C1 : (xi , yi ) ∈ Q
n
C2 : 1i,j = 1, ∀ j ∈ [1, s]
i=1
C3 : αmin ≤ αi,j ≤ αmax , ∀ i ∈ [1, n], ∀ j ∈ [1, s]
C4 : αi,j + Gi,j − Li,j ≥ α0 , ∃ i ∈ [1, n], ∀ j ∈ [1, s]
Here, n and s denote the number of used APs and STAs respectively, (xi , yi ) is
the position of AP i , 1i,j indicates whether STAj is associated to AP i , αi,j , ti,j ,
βi,j are the used transmit power, transmission time, and total communication
activity rate (including both downlink and uplink activity) occurring in the link
between AP i and STAj respectively, Q constrains the service area, αmin and
αmax set the lower and upper bound for the transmit power, Gi,j and Li,j are the
antenna gains and losses of the communication link between AP i and STAj , and
α0 refers to the receiver sensitivity of signal detection. For simplicity we assume
a single transmit power setting for both directions of a link; the framework can
readily extended to handle differing transmit power values.
The antenna gains add both the transmitter antenna gain and receiver
antenna gain. Similarly, the loss Li,j of each link contains three components:
cable and connector losses on both sides, path loss, and miscellaneous losses
such as fading margin. The propagation loss is estimated using the log-distance
path loss model:
d
L = L0 + 10γlog10 ( ), (2)
d0
where L0 is the path loss at the reference distance d0 , γ is the decay component,
and d is the distance between transmitter and receiver.
Figure 1 illustrates the communication activities in a simple network con-
sisting of two STAs and one AP. Different colors (i.e., blue and black) are
used to distinguish different directions of transmission. Dashed lines represent
expected/imagined communication paths, while solid lines represent the corre-
sponding actual occurring communication paths. Suppose STA1 needs to send
3 messages to STA2 . After receiving and analyzing the messages, STA2 sends
a message back to STA1 . The intermediate AP AP 1 acts as a relay node to
perform the above operations. In this case, β1,1 = 3(uplink ) + 1(downlink ) = 4
and β1,2 = 1(uplink ) + 3(downlink ) = 4. Note that they are equal because there
are only two links existing in this scenario. α1,1 is the transmit power used by
STA1 and AP 1 , and t1,1 is the transmission time of packets in the link between
AP 1 and STA1 . α1,2 and t1,2 have similar meaning but between AP 1 and STA2 .
If the constraints C1 , C3 , and C4 are jointly satisfied, then STAj is effi-
ciently covered by AP i in the given environment. Depending on the settings, it
186 X. Qi et al.
Fig. 3. The factory process flow model that we use in our experiments.
Five types of functional units (actors) are used in the system: part generator,
rail, machine, controller, and part sink, where they cooperatively model a basic
188 X. Qi et al.
5 Experiments
part source to the part sink. Three types of machines are used, and they have
activity rates of 8, 6, and 3 messages/cycle, representing high activity, medium
activity, and low activity, respectively. The activity rate of a DRSM controller
D is the sum of the activity rates of the two rails and one machine that are
connected to D, while the simple controller is only responsible for the last rail.
Considering the different volumes of machines and different lengths of convey-
ors that are typically found in practice, the spacing between actors is nonuniform
in our experiments.
We apply the same channel configuration across the entire system model.
Unless otherwise stated, the activity rate and placed location of each actor is
as listed in Table 1, and other aspects of the simulation setup are as listed in
Table 2.
Regarding the threshold for signal detection, two similar values are used in
related literature: –65 dBm [13], and –70 dBm [14]. We used the value of –65 dBm
to account for the severe multipath fading typical in industrial environments.
Table 1. Communication activity rate and position for each actor. The units for the
activity rate are messages/cycle.
Actor R1 M1 R2 M2 R3 M3
Activity 3 8 5 6 5 3
Position (0, 0) (0, 2) (0, 5) (0, 7) (0, 9) (0, 15)
Actor R4 C1 C2 C3 SC 1
Activity 5 14 11 8 2
Position (0, 18) (1, 2) (1, 6) (1, 14) (1, 17)
Parameter Value
Number of GA generations 1000
Population size 300
Crossover rate 0.5
Mutation rate 0.2
Tournament selection size 10
Number of bits in each variable 6
Maximum transmit power of AP (αmax ) 17 dBm
Minimum transmit power of AP (αmin ) 0 dBm
Path loss exponent (γ) 3
Reference distance (d) 1m
Threshold for signal detection (α0 ) −65 dBm
190 X. Qi et al.
Two scenarios are intensively considered in our experiments: (1) all devices
are cable-connected to their power supplies, and (2) all devices are powered by
batteries.
In this simulation, the goal of our deployment strategy is to minimize the com-
munication energy consumption of the network. The position of each actor is
listed in Table 1. The actors are non-uniformly distributed in this layout.
First, we do not take into account the non-uniform distribution of actors, nor
do we take into account unbalanced communication activity, and we place the AP
at the center position (0.50, 10.00) of the pipeline. This center position is a simple
and intuitive choice if we do not take into account non-uniform distribution and
unbalanced communication, as described above. We execute the simulator for
1000 cycles and record the obtained energy consumption.
Next, we execute AP2GA to take into account the non-uniform actor distri-
bution and unbalanced communication activity, and derive an optimized position
for a single AP. The resulting AP position is (0.70, 7.90). We move the AP to this
position in our simulation model, and again execute 1000 simulation cycles. We
compare the energy consumption brought by (1) deploying the AP in the center
position (“middle”), and (2) deploying the AP based on the result produced by
AP2GA. The results are shown in Fig. 4.
In Fig. 4, the total energy consumption is the summation of the energy con-
sumed by both the AP and the STAs. The AP column represents the over-
all downlink energy consumption, while the STAs column represents the over-
all uplink energy consumption. Since the maximum/minimum allowed transmit
power and available power levels are all set to be the same for every communi-
cation node in the simulation model, the differences in the energy consumption
between the downlink and uplink come from the uneven inflow and outflow the
actors.
It can be clearly seen from Fig. 4 that even for this relatively simple and
small-scale example, AP2GA results in a significant reduction in total energy
consumption compared to the simple/intuitive strategy of placing the AP at
the center position. The relative energy savings provided by AP2GA is about
10%, which will amount to significant absolute energy savings over long-term
operation. Since the uneven outgoing and incoming traffic makes the down-
link bear more long-distance workload, there is more significant reduction in
the energy consumed by the AP. For example, when Rail 2 sends packets to
DRSM Controller 1, the downlink transmission distance (from the AP to the
controller) is longer than the uplink transmission distance (from the rail to the
AP).
In Sect. 5.1, we optimized the total energy consumption supposing that all of
the STAs and the AP are cable connected to power supplies. However, due to
their low-price and easy installation, an increasing proportion of communication
devices in industrial environments are powered by batteries.
When battery-powered communication devices are employed, it is important
to consider the network durability when designing and configuring the network.
Intuitively, by network durability, we mean the length of time that the network
remains operational as batteries in the communication devices are drained. There
are various ways to measure the network durability depending on the particular
kinds of operational scenarios that are of interest. Since our scenario requires
the mutual work of all devices in the network, we use a measure of network
lifetime to assess durability, and we regard the time until one STA’s battery is
drained as the network lifetime. That is, the network lifetime is the time from
the beginning of operation until the time when the first STA stops operating
due to insufficient energy availability.
Assuming that all devices have the same battery capacity, maintaining a long
network lifetime requires that all devices consume energy at approximately the
same average rate — that is, the variance of energy consumption across the
battery-powered devices should be low. To assess energy consumption variance,
we plotted the energy consumption of each STA under both the center-position
and AP2GA-based AP deployment obtained from case 1 in Fig. 5a, and tabulated
their corresponding standard deviations (“Std.”) values in Table 3.
192 X. Qi et al.
Label center pos. AP2GA pos. 1 with cables AP2GA pos. 2 with battery
AP Position (0.50, 10.00) (0.70, 7.90) (–0.16, 8.57)
Std. 1.74 1.38 1.26
Fig. 5. Energy consumption levels of the different STAs under different AP deployment
configurations.
From Fig. 5a, we can see that the energy consumption of the STAs is unbal-
anced in both deployment scenarios — center-position and AP2GA-based. Peaks
appear on different devices depending on the combination of communication dis-
tance and activity rate. However, the distribution of STA energy consumption
under AP2GA-based deployment from case 1 has better performance in terms
of standard deviation.
To prolong network lifetime and ameliorate the imbalance described above,
the dispersion of STA energy consumption can be taken into account in AP2GA.
For this purpose, a maximum value for the standard deviation std max can be
imposed as another constraint:
s
s
j=1 (ej − μ)2 j=1 ej
C5 : ≤ std max , where ej = αi,j ti,j βj , μ = . (3)
s s
5.3 Summary
In summary, from the study and experimental results presented in this paper,
two main findings and implications emerge. First, in environments where users
are unevenly distributed and their communication traffic varies, proper deploy-
ment of APs can significantly reduce the transmission energy consumption of
the entire network. Second, the original formulation of AP2GA can be readily
extended to other energy-related scenarios by manipulating selected parameters
and introducing additional constraints. Averaging the transmission energy of
battery-powered devices is an example, and there are many additional possibil-
ities for performing other types of design space exploration.
6 Conclusion
References
1. Back, T., Hammel, U., Schwefel, H.: Evolutionary computation: comments on the
history and current state. IEEE Trans. Evol. Comput. 1(1), 3–17 (1997)
2. Björnson, E., Larsson, E.G.: How energy-efficient can a wireless communication
system become?. In: 2018 52nd Asilomar Conference on Signals, Systems, and
Computers (2018)
3. Candell, R., Kashef, M., Liu, Y., Lee, K.B., Foufou, S.: Industrial wireless systems
guidelines: practical considerations and deployment life cycle. IEEE Ind. Electron.
Mag. 12(4), 6–17 (2018)
4. Fortin, F.A., De Rainville, F.M., Gardner, M.A., Parizeau, M., Gagn’e, C.: DEAP:
evolutionary algorithms made easy. J. Mach. Learn. Res. 13, 2171–2175 (2012)
5. Geng, J., et al.: Model-based cosimulation for industrial wireless networks. In: Pro-
ceedings of the IEEE International Workshop on Factory Communication Systems,
Imperia, Italy, pp. 1–10 (June 2018)
6. Jaffres-Runser, K., Gorce, J., Ubeda, S.: QoS constrained wireless LAN optimiza-
tion within a multiobjective framework. IEEE Wirel. Commun. 13(6), 26–33 (2006)
7. Katoch, S., Chauhan, S.S., Kumar, V.: A review on genetric algorithm: Past,
present, and future. Multimed Tools Appl. 80(5), 8091–8126 (2021)
8. Kouhbor, S., Ugon, J., Rubinov, A., Kruger, A., Mammadov, M.: Coverage in
WLAN with minimum number of access points. In: 2006 IEEE 63rd Vehicular
Technology Conference (2006)
Energy-Efficient Access Point Deployment for Industrial IoT Systems 195
9. Li, H., Geng, J., Liu, Y., Kashef, M., Candell, R., Bhattacharyya, S.: Design space
exploration for wireless-integrated factory automation systems. In: Proceedings of
the IEEE International Workshop on Factory Communication Systems, Sundsvall,
Sweden (May 2019), 8 pages in online proceedings
10. Ling, X., Yeung, K.L.: Joint access point placement and channel assignment for
802.11 wireless LANs. IEEE Trans. Wireless Commun. 5(10), 2705–2711 (2006)
11. Project, n.: The network simulator ns–3 (2016). https://www.nsnam.org/
(Accessed 20 October 2016)
12. Tang, S., Ma, L., Xu, Y.: A novel AP placement algorithm based on user distribu-
tion for indoor WLAN system. China Commun. 13(10), 108–118 (2016)
13. Zhang, Z., Di, X., Tian, J., Zhu, Z.: A multi-objective WLAN planning method.
In: 2017 International Conference on Information Networking (ICOIN) (2017)
14. Zhi, Z., Wu, J., Meng, X., Yao, M., Hu, Q., Tang, Z.: AP deployment optimization
in non-uniform service areas: a genetic algorithm approach. In: 2019 IEEE 90th
Vehicular Technology Conference (VTC 2019-Fall) (2019)
Hardware/Software Solutions for IoT
and CPS (HSS)
FAMID: False Alarms Mitigation
in IoMT Devices
1 Introduction
The Internet of Medical Things (IoMT) refers to a group of medical devices and
software programs interconnected via computer networks that supports health-
care information technology systems to collect and exchange data. By 2030, the
global IoMT market is projected to reach USD 861.3 billion, expecting a com-
pound annual growth rate of 16.8% from 2023 to 2030 [41]. Physiological closed-
loop control systems (PCLCS), smart medical devices, wearable and implantable
medical devices (WIMDs), remote patient monitoring systems, and telemedicine
platforms are some examples of IoMT devices. These devices can gather health-
related information, including vital signs, medication dosage, blood glucose lev-
els, drug concentration, etc. Real-time transmission of this data to healthcare
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 199–217, 2023.
https://doi.org/10.1007/978-3-031-45878-1_14
200 S. Mahmud et al.
doctors and other systems makes it possible to monitor patients from a distance,
make more precise diagnoses, and develop better treatment strategies [47]. IoMT
devices have a wide range of security issues, such as vulnerabilities in software
and firmware, physical attacks, weak encryption/authentication, and a lack of
security patches or updates [35]. Additionally, system reliability is paramount
so that healthcare practitioners can not only access the data when required but
also trust that data is accurate, so patients can receive safe and effective care.
The addition of autonomous, closed-loop treatment adds additional complex-
ity to IoMT devices. Any vulnerabilities or potential reliability issues become
even more serious in this context. A fault in any system component can cause
the device to malfunction. A faulty device could result in inaccurate diagnoses,
improper treatment, or even patient injury if the device does not perform as
intended. Besides, IoMT devices must be able to communicate with other health-
care systems to provide continuity of treatment. The delivery of patient care may
be delayed or fail altogether if the devices are unstable or produce inaccurate
data. Additionally, regular maintenance, including updates and patching of a
defective IoMT device can be expensive. Therefore, a PCLCS must be resistant
to failure in order to ensure long-term in-field safety and reliability. The over-
all IoMT architecture is illustrated in Fig. 1, which shows the various scenarios
involving WIMDs and PCLCS. For example, a person can have WIMDs that are
IoMT-enabled but not part of a PCLCS. Alternatively, a WIMD may be utilized
in a PCLCS that is not IoMT-enabled. Additionally, a WIMD can be part of
both a PCLCS and an IoMT.
Smartglasses
Smartwatch
Fig. 1. Overall IoMT architecture, illustrating different WIMDs and cloud-based (A)
vs. local processing (B) for closed-loop systems.
drug or therapy [13]. System safety features are crucial for any PCLCS design
to enhance the safety and reliability of the device. For example, an alert system
can help the patient or the healthcare provider in a variety of ways, such as
indicating over- or under-medication, low battery, or any potential issue with the
device that needs to be addressed. While alarms can help to take prompt action,
false alarms can make it difficult for patients to distinguish between genuine
and fake incidents. In addition, false alarms impact the battery life and the
overall functionality of the device by unnecessarily consuming additional power.
Detecting and mitigating false alarms in real-time is essential for ensuring the
reliability of the device and improving the overall effectiveness of the system.
Several studies in the literature have explored various methods for identifying
and addressing false alarms in different fields, such as cyber-physical systems [26]
and medical systems or units [19,53]. However, there has been a lack of research
on detecting and correcting false alarms in closed-loop medical control systems.
This paper aims to address that research gap by exploring the following research
questions:
RQ-1: How to effectively identify false alarms in PCLCS that may occur due to
natural failures or malicious modifications?
RQ-2: How can we efficiently mitigate false alarms in PCLCS to ensure the
safety and reliability of the device?
1. We model different types of faults or attacks that can potentially disrupt the
alert mechanism system of a PCLCS and cause it to generate false alarms.
2. We propose an implementation-level redundancy approach for the PCLCS
controller that utilizes cloud-based parallel controllers to detect false alarms
in real-time.
3. We implement an efficient alarm mitigation technique, FAMID, that effec-
tively reduces false alarms arising from erroneous controller outputs to
decrease the number of false alarms.
202 S. Mahmud et al.
The rest of the paper is organized as follows: In Sect. 2, we explore the secu-
rity and reliability concerns that arise from the use of IoMT devices with PCLCS
components, describe the alert mechanism system and provide an overview of the
specific (APS) control algorithms. In Sect. 3, we discuss related works, emphasiz-
ing the various false alarm mitigation techniques found in the literature and how
they relate to IoMT devices. In Sect. 4, we delve into the details of the threat
model. We then present our methodology in detail in Sect. 5. Section 6 presents
and discusses the findings. Finally, we conclude in Sect. 7.
2 Background
Alert mechanism systems can be used in various domains to ensure the safety of
individuals. For example, the authors in [37] propose an SMS-based alert system
for detecting car accidents and notifying emergency services. The SMS-based
alert system in this approach takes advantage of the widespread use of mobile
devices to provide timely notifications to first responders. When the system
detects an accident, it uses GPS technology to pinpoint the location and sends a
text message to rescue services, providing them with the information they need
to dispatch assistance to the scene of the accident. The app-based alert system
in Cerberus leverages the widespread use of mobile devices to provide timely
notifications to users [12]. The mobile app can be easily installed on the user’s
device and communicate with the cloud server to receive and process alerts. The
app provides information about the alert’s location, type, and severity, allowing
users to take appropriate action to stay safe. For example, if there is a flood alert,
the app will notify the user of the unsafe water level at a particular location
and provide guidance on avoiding the area. Similarly, if there is a fire alert,
the app will provide information on the location and severity of the fire and
recommend evacuation routes. A device equipped with both visual and audio
alert mechanisms was created to ensure that miners are alerted in real-time of
hazardous conditions [33].
In medical devices, alert mechanism systems are used to alert healthcare
providers of potential issues or risks related to a patient’s medical condition or
the functionality of a medical device [22]. For example, an alert mechanism sys-
tem in a patient monitoring system may alert healthcare providers if a patient’s
vital signs indicate a potential medical emergency. In addition, medical device
manufacturers must implement alert mechanism systems to comply with regula-
tory requirements and ensure patient safety [21]. Alert mechanisms are essential
tools in various industries to enhance safety measures for workers and individu-
als. These mechanisms can come in various forms, such as mobile apps, wearable
devices, or SMS-based systems, and can provide timely and crucial notifica-
tions about potential hazards or emergencies. With the increasing importance
of safety measures, alert mechanism systems are becoming more widespread and
are now an integral part of regulatory requirements in many industries, includ-
ing healthcare. Furthermore, the use of alert mechanisms is expected to grow as
new technologies and innovations emerge, making workplaces and public spaces
safer for all.
The control algorithm determines the automatic operations of the PCLCS. The
primary purpose of the control algorithm is to apply modifications to the given
medicine or therapy to ensure that the PCLCS satisfies clinically related per-
formance requirements [13]. Characterizing the physiologic variable’s response
and the interaction of any elements that could impact these processes should
be the basis for the control algorithms employed in a PCLCS. For example, in
FAMID: False Alarms Mitigation in IoMT Devices 205
an automated blood glucose system, the blood glucose measurement and the
meals taken need to be considered by the controller to maintain the patient’s
insulin level safely. To meet Food and Drug Administration (FDA) recommen-
dations, control algorithms should be designed to operate under potential risks
and environmental interference.
In this paper, as a case study, we used an APS as a closed-loop system,
which includes a continuous glucose monitor (CGM). Computer simulation has
accelerated the development of AP control algorithms. For example, the authors
in [29] developed the UVA/PADOVA Type 1 diabetes simulator, which can
mimic meal challenges in virtual subjects, represented by a model parameter
vector that was randomly selected from a suitable joint parameter distribution.
The control algorithm implemented consists of a Ibo (bolus insulin) calculator
as described by Eqn 1,
CHO (Gp − Gt )
Ibo = + (1)
CR CF
with parameters Gp (amount of glucose in plasma), Gt (patient target
glucose), CF (patient correction factor), CR (carbohydrate ratio), and CHO
(ingested carbohydrate). The controller reads a patient’s glucose level and car-
bohydrate amount from the dataset [4] stored in a .csv file for 10 patients with
200 samples each. The algorithm examines the patient’s glucose readings and
identifies any values below 70 or above 180, known as threshold alert values,
as true alarms [5]. Furthermore, the algorithm computes the bolus amount by
considering the patient’s target glucose value and other relevant parameters.
3 Related Work
Detecting false alarms can help lower emotional or mental anxiety, reduce unnec-
essary computation power of the system, and ensure safety and reliability. Many
false alarms can increase the risk of poor responses from the system to actual
emergencies, which can harm people or cause financial loss. The primary goal of
false alarm detection is to ensure that appropriate actions are taken for authen-
tic situations instead of false ones. Researchers from various disciplines have
investigated several methods to detect and mitigate false alarms. For example,
several scientific research has been conducted to detect false alarms in a vari-
ety of application domains like the internet of connected vehicles [1], marine
environments [10], wind turbine [31], and medical cyber-physical systems [26].
However, as the WIMDs are evolving rapidly with sophisticated components, it
is required to develop new techniques that can effectively detect and mitigate
false alarms to enhance patient experiences.
The authors in [44] proposed an approach to identify sensor anomalies by
analyzing the historical data collected from various biosensors to detect and
reduce false alarms. Their methodology consists of four algorithms: the corre-
lation coefficient for physiological data preprocessing, random forest for sensor
value prediction, the dynamic threshold for error calculation, and the majority
206 S. Mahmud et al.
voting for alarm trigger. They used a large real-time healthcare dataset to evalu-
ate their methodology and found a high false alarm detection rate and a low false
positive rate. Furthermore, in [53], the authors proposed a robust methodology
to detect seizures for wearable devices and tested that with the CHB-MIT elec-
troencephalogram (EEG) dataset. To note, EEG acquisition is a time-consuming
and error-prone process, and many seizure detection techniques are associated
with unacceptably high false-alarm rates. Compared to the other studies on the
same issue, their approach resulted in a 34.70% reduction in false alarms and
demonstrated that their technique could extend the battery life of a cutting-edge
wearable device by reducing the frequency of false alarms. Decreasing the num-
ber of false alarms is essential in intensive care Unit (ICU) to improve patient
safety. In [52], the authors proposed a game-theoretic feature selection technique
that uses a genetic algorithm to find the most useful biomarkers from signals
obtained from various monitoring equipment. To reduce the false alarms in the
ICU, the authors presented this low-computational complexity approach to esti-
mate Shapley values for physiological signal characteristics. They reported that
their proposed technique captured the actual alarms with better sensitivity and
equivalent specificity compared to other feature selection methods and reduced
the false alarms considerably.
In recent years, neural networks (NN), such as deep neural networks (DNNs),
convolutional neural networks (CNNs), etc., have been used by researchers to
detect false alarms by identifying patterns in the dataset. For example, the
authors in [19] utilized the evolutionary and swarm algorithm to improve the
DNN performance in detecting false alarms in ICU. They reported reduced sup-
pressed true alarms by improving the accuracy compared to the benchmark Phy-
sionet challenge by utilizing dispersive flies optimization (DFO). In their study,
5-fold cross-validation was done using two models with different architectures.
The results showed that compared to other results, including the benchmark,
the DFO-modified version for both models performed better. Furthermore, the
authors in [54] used CNNs to learn the feature representations of physiological
waveforms to differentiate between authentic and false arrhythmia alarms. Their
method utilizes contrastive learning to reduce binary cross-entropy classification
and similarity loss. They tested their approach with a benchmark, the 2015 Phy-
sioNet Computing in Cardiology Challenge, and reported that their proposed
deep-learning framework performed better than the benchmark challenge.
Although various methods have been developed to detect and reduce false
alarms in multiple fields, there is a clear research gap in utilizing false alarm
detection and reduction techniques in closed-loop medical control systems. This
paper aims to fill this gap by introducing an innovative and effective false alarm
mitigation technique for PCLCS. A fault-tolerant system is required to ensure
reliability and effective treatment. One potential technique for fault tolerance
in control systems is hardware redundancy, where identical hardware modules
perform the same functions. In case of a fault in one module, the other modules
continue the process, maintaining the system’s functionality. The triple modu-
lar redundancy (TMR) approach is frequently used to improve hardware fault
FAMID: False Alarms Mitigation in IoMT Devices 207
tolerance. Our proposed approach benefits from the TMR method, particularly
N-version programming, where the same basic requirements are used to build
independent versions of many functionally identical programs. IoMT is then
leveraged to support cloud-based redundancy by offloading multiple versions or
implementations of the control algorithm in a trusted cloud environment. Diver-
sifying the alert mechanism in this way can help to improve reliability against
both natural faults and intentional attacks.
4 Threat Model
tampering with the PCLCS or changing any system component. For example,
the intentional malicious modification to any circuitry of an integrated circuit is
known as a hardware Trojan, and if that is undetected during the manufacturing
process, it can result in device malfunction [28]. From the perspective of IoMT
devices, protecting against such intentional attacks is crucial as software security
is sometimes inadequate [45].
PCLCS can malfunction as a result of faults in any of the system components.
This paper focuses on a PCLCS with a faulty controller that occasionally gener-
ates incorrect actuating values leading to false alarms. We are assuming that: 1)
the controller is defective and it can be a result of a natural fault or intentional
malicious attack, and 2) the other components (biosensor, transducer, signal
processing unit, analog to digital hardware, alert system, etc.) are error-free.
Our goal is to design and implement an effective reliability mechanism that can
reduce false alarms and maintain the safety and dependability of the PCLCS,
as any incorrect reading from the defective controller can cause the device to
malfunction and generate false alarms.
Comparison
Controller
(Faulty)
Actuator
No False alarm Yes
detected?
Continue with
HW Controller
Biosensor
Fig. 2. An overview of our proposed approach illustrating how cloud controllers can
help detect and mitigate false alarms in a PCLCS.
5 Methodology
This section presents a detailed description of the methodology and the exper-
imental setup. We divide our approach into four main parts: 1) implementing
cloud-based controllers, 2) implementing WebSocket API for two-way interactive
communication, 3) generating false alarms to simulate accidental or intentional
faults with the controller, and 4) detecting and mitigating the false alarms in
real-time. An overview of our proposed approach is illustrated in Fig. 2.
FAMID: False Alarms Mitigation in IoMT Devices 209
Our proposed approach is built upon the TMR method, a fault-tolerant tech-
nique commonly used in safety-critical systems. Specifically, we utilize the N-
version programming aspect of TMR, where multiple independent versions of
the same program are created based on the same requirements. Each version is
designed to be functionally identical but implemented using different algorithms
and/or programming languages. We implemented three versions of the same con-
trol algorithm, written in different programming languages, which include C++,
Rust, and Python. These languages were chosen based on their memory efficiency
and speed, with Python being used for the local physical device (Raspberry Pi)
and C++ and Rust for their speed and memory efficiency. It is important to
note that our approach to implementing cloud-based controllers is not limited
to specific programming languages and is not limited to implementation diver-
sity.
We analyzed the compiled code of the Rust and C++ algorithms and found
significant differences in the number of instructions and clock cycles. Specifi-
cally, the Rust compiled algorithm had nearly 48% more instructions than the
C++ version. The version of the Rust compiler used was rustc v.1.69, while the
version of the C++ compiler used was gcc v.17. We wrote a script to convert
the Python-interpreted code into C to find the differences between all three pro-
grams. We observed the execution time to be not significantly different among
the three languages. However, the number of main memory access in the Rust
version was the highest, followed by C++ and C. Therefore, it is important
to consider the programming language and compiler version when implement-
ing control algorithms, as these factors can significantly impact the resulting
program’s performance and efficiency.
of the PC and the port number specified during server configuration. Once a con-
nection was established, the client awaited a predefined symbol to arrive in a
WebSocket message to indicate that all clients on the network should run an iter-
ation of the simulation. We also tested the system using multiple virtual machine
clients running on the same PC as the WebSocket server. Each virtual machine
was configured with a unique IP address and ran a simple Python script that
connected to the WebSocket server as a client and printed incoming messages to
the console. This allowed us to simulate multiple clients and test the scalability
and resiliency of the system. We collected data by storing the returned result
values received by the WebSocket server in a .csv file for later analysis. Figure 3
shows our experimental communication setup between cloud and local devices.
Fig. 3. Diagram of methods and technologies used in our experimental setup to com-
municate between cloud and local devices.
our proposed approach is effective in detecting and mitigating false alarms gen-
erated by a faulty controller. It is important to note that the chosen scenarios
and parameters are worse than what would typically occur in the real world.
However, they are still comparable to real-world scenarios but with a higher
probability of faults due to the limited dataset.
Fig. 4. Bar charts illustrating the true alarms, total alarms including false alarms,
and the number of alarms after applying FAMID technique for random nonsequential
dropped values (RNDV).
Fig. 5. Bar charts illustrating the true alarms, total alarms including false alarms, and
the number of alarms after applying FAMID technique for single-bit error (SBE).
212 S. Mahmud et al.
the output signals of the local and cloud controllers to detect false alarms and
then chooses to continue operation with either the local or cloud controller.
By detecting and mitigating false alarms generated by the local controller, the
system’s reliability is improved.
We present the results and discussions in this section divided into two main
parts: 1) the accuracy of cloud-based controllers and 2) false alarm detection
and mitigation using FAMID. We recorded the number of true alarms triggered
by the alarm-triggering condition to observe the number of occurrences in the
local and cloud controllers. The alarm-triggering condition for true alarms was
set when the blood glucose value was greater than 180 or less than 70. For
all samples in the test dataset, we confirmed that the cloud-based algorithms
generated the same alarms as the unmodified local physical controller.
Our approach for detecting and mitigating false alarms addresses two crit-
ical issues with the local controller: single-bit errors (SBE) and random non-
sequential dropped values (RNDV). To validate the effectiveness of our pro-
posed approach, we conducted evaluations using data from ten different patients.
In Fig. 4, we showcase the results of our technique in reducing the number of
additional false alarms caused by RNDV in the controller’s output signal. For
instance, for patient 1, there were initially 53 true alarms and 7 false alarms,
totaling 60 alarms before implementing our technique. However, after applying
our proposed approach, the number of alarms was significantly reduced to 53,
matching the number of true alarms. This significant reduction demonstrates
our approach’s high accuracy and efficiency in mitigating false alarms. Simi-
larly, Fig. 5 presents compelling results when an SBE was introduced into the
local controller’s output values. Our proposed technique also successfully mit-
igated false alarms in this scenario, showcasing its versatility and robustness.
To provide a comprehensive overview of our results, Table 2 summarizes each
patient’s true, false, and total alarms before and after implementing our FAMID
technique. The data further confirms the effectiveness of our approach across
various patient cases.
Our proposed technique has proved its effectiveness in substantially reducing
false alarms caused by both RNDV and SBE, presenting a promising solution
to enhance alarm accuracy and ensure patient safety in medical systems. The
evaluation results clearly illustrate the potential advantages of implementing
our technique in real-world applications, significantly contributing to creating a
more reliable and efficient healthcare environment. By addressing false alarms,
our approach can minimize unnecessary alerts and enable medical professionals
to focus on critical cases promptly, improving patient outcomes and a more
streamlined healthcare system. The positive results from this study highlight
the importance of further integrating our approach in various medical devices
and settings to optimize alarm management and overall patient care.
FAMID: False Alarms Mitigation in IoMT Devices 213
Table 2. Comparing Total Alarms (A), True Alarms, and False Alarms for RNDV and
SBE using FAMID
7 Conclusion
In this paper, we have presented a comprehensive overview of the security
and reliability challenges associated with WIMDs and PCLCS in the IoMT.
These devices are globally significant in improving individual well-being, but
their growing complexity, driven by hardware and software advancements, intro-
duces crucial concerns regarding their safety and effectiveness. Securing WIMDs
and PCLCS is essential to ensure patient safety and reliability. Several secu-
rity issues have been identified in these devices, and false alarms in PCLCS
are particularly concerning, as they can lead to alarm desensitization, alarm
fatigue, and distress for the patient. The technique proposed in this work aims
to mitigate false alarms in real-time using implementation-level redundancy by
implementing cloud-based controllers to improve the reliability of the PCLCS.
Our approach demonstrated a complete success rate in detecting and mitigating
false alarms, thereby ensuring the reliability of the PCLCS. Future research on
this work will focus on enhancing the robustness of our approach by integrating
algorithmic-level diversity with implementation-level redundancy and testing the
proposed technique on additional datasets to evaluate its effectiveness in different
scenarios.
References
1. Al Zamil, M.G., et al.: False-alarm detection in the fog-based internet of connected
vehicles. IEEE Trans. Veh. Technol. 68(7), 7035–7044 (2019)
2. Alshalalfah, A.L., Hamad, G.B., Mohamed, O.A.: Towards system level security
analysis of artificial pancreas via uppaal-smc. In: 2019 EEE International Sympo-
sium on Circuits and Systems (ISCAS), pp. 1–5. IEEE (2019)
214 S. Mahmud et al.
20. Hovorka, R.: Closed-loop insulin delivery: from bench to clinical practice. Nat. Rev.
Endocrinol. 7(7), 385–395 (2011)
21. International Organization for Standardization: ISO 13485:2016 - Medical devices
- Quality management systems - Requirements for regulatory purposes (2018)
22. Joshi, S., Joshi, S.: A sensor based secured health monitoring and alert technique
using iomt. In: 2019 2nd International Conference on Intelligent Communication
and Computational Techniques (ICCT), pp. 152–156 (2019)
23. Kim, D., Park, S., Choi, K., Kim, Y.: BurnFit: analyzing and exploiting wearable
devices. In: Kim, H., Choi, D. (eds.) WISA 2015. LNCS, vol. 9503, pp. 227–239.
Springer, Cham (2016). https://doi.org/10.1007/978-3-319-31875-2 19
24. Kune, D.F., et al.: Ghost talk: mitigating emi signal injection attacks against analog
sensors. In: 2013 IEEE Symposium on Security and Privacy, pp. 145–159. IEEE
(2013)
25. Li, Q., Ding, D., Conti, M.: Brain-computer interface applications: security and pri-
vacy challenges. In: 2015 IEEE conference on communications and network security
(CNS), 663–666. IEEE (2015)
26. Li, W., Meng, W., Su, C., Kwok, L.F.: Towards false alarm reduction using fuzzy
if-then rules for medical cyber physical systems. IEEE Access 6, 6530–6539 (2018)
27. Mahmud, S., Majerus, S.J., Damaser, M.S., Karam, R.: Design tradeoffs in bioim-
plantable devices: a case study with bladder pressure monitoring. In: 2018 IEEE
24th International Symposium on On-Line Testing and Robust System Design
(IOLTS), pp. 69–72. IEEE (2018)
28. Mahmud, S., Zareen, F., Olney, B., Karam, R., et al.: Trojan resilience in
implantable and wearable medical devices with virtual biosensing. In: 2022 IEEE
40th International Conference on Computer Design (ICCD), pp. 577–584. IEEE
(2022)
29. Man, C.D., Micheletto, F., Lv, D., Breton, M., Kovatchev, B., Cobelli, C.: The
uva/padova type 1 diabetes simulator: new features. J. Diabetes Sci. Technol. 8(1),
26–34 (2014)
30. Marin, E., Singelée, D., Garcia, F.D., Chothia, T., Willems, R., Preneel, B.: On the
(in) security of the latest generation implantable cardiac defibrillators and how to
secure them. In: Proceedings of the 32nd Annual Conference on Computer Security
Applications, pp. 226–236 (2016)
31. Marugán, A.P., Chacón, A.M.P., Márquez, F.P.G.: Reliability analysis of detecting
false alarms that employ neural networks: a real case study on wind turbines.
Reliability Eng. Syst. Safety 191, 106574 (2019)
32. Newaz, A.I., Sikder, A.K., Rahman, M.A., Uluagac, A.S.: A survey on security and
privacy issues in modern healthcare systems: attacks and defenses. ACM Trans.
Comput. Healthcare 2(3), 1–44 (2021)
33. Noorin, M., Suma, K.: Iot based wearable device using wsn technology for min-
ers. In: 2018 3rd IEEE International Conference on Recent Trends in Electronics,
Information & Communication Technology (RTEICT), pp. 992–996 (2018)
34. Olney, B., Mahmud, S., Karam, R.: Evaluating edge processing requirements in
next generation iot network architectures. In: Casaca, A., Katkoori, S., Ray, S.,
Strous, L. (eds.) IFIPIoT 2019. IAICT, vol. 574, pp. 252–269. Springer, Cham
(2020). https://doi.org/10.1007/978-3-030-43605-6 15
35. Papaioannou, M., et al.: A survey on security threats and countermeasures in
internet of medical things (iomt). Trans. Emerging Telecommun. Technol. 33(6),
e4049 (2022)
216 S. Mahmud et al.
36. Papp, D., Ma, Z., Buttyan, L.: Embedded systems security: threats, vulnerabilities,
and attack taxonomy. In: 2015 13th Annual Conference on Privacy, Security and
Trust (PST), pp. 145–152. IEEE (2015)
37. Parveen, N., Ali, A., Ali, A.: Iot based automatic vehicle accident alert system.
In: 2020 IEEE 5th International Conference on Computing Communication and
Automation (ICCCA), pp. 330–333. (2020)
38. Pycroft, L., et al.: Brainjacking: implant security issues in invasive neuromodula-
tion. World Neurosurgery 92, 454–462 (2016)
39. Ransford, B., et al.: Cybersecurity and medical devices: a practical guide for cardiac
electrophysiologists. Pacing Clin. Electrophysiol. 40(8), 913–917 (2017)
40. Rasmussen, K.B., Castelluccia, C., Heydt-Benjamin, T.S., Capkun, S.: Proximity-
based access control for implantable medical devices. In: Proceedings of the 16th
ACM Conference on Computer and Communications Security, pp. 410–419 (2009)
41. Research, G.V.: Internet of things in healthcare market worth $861.3 billion
by 2030 (March, 2023). https://www.grandviewresearch.com/press-release/global-
iot-in-healthcare-market
42. Rios, B., Butts, J.: Security evaluation of the implantable cardiac device ecosystem
architecture and implementation interdependencies. WhiteScope, sl (2017)
43. Ronquillo, J.G., Zuckerman, D.M.: Software-related recalls of health information
technology and other medical devices: Implications for fda regulation of digital
health. Milbank Q. 95(3), 535–553 (2017)
44. Saraswathi, S., Suresh, G., Katiravan, J.: False alarm detection using dynamic
threshold in medical wireless sensor networks. Wireless Netw. 27, 925–937 (2021)
45. Sidhu, S., Mohd, B.J., Hayajneh, T.: Hardware security in IOT devices with empha-
sis on hardware trojans. J. Sens. Actuator Netw. 8(3), 42 (2019)
46. Spence, N., Niharika Bhardwaj, M., Paul III, D.P.: Ransomware in healthcare facil-
ities: a harbinger of the future? Perspectives in Health Information Management,
pp. 1–22 (2018)
47. Sundaravadivel, P., Lee, I., Mohanty, S., Kougianos, E., Rachakonda, L.: Rm-iot: an
iot based rapid medical response plan for smart cities. In: 2019 IEEE International
symposium on smart electronic systems (iSES)(Formerly iNiS), 241–246. IEEE
(2019)
48. Tyagi, A.K., Sreenath, N.: Cyber physical systems: analyses, challenges and pos-
sible solutions. Internet of Things Cyber-Phys. Syst. 1, 22–33 (2021)
49. Wehbe, T., Mooney, V.J., Javaid, A.Q., Inan, O.T.: A novel physiological features-
assisted architecture for rapidly distinguishing health problems from hardware tro-
jan attacks and errors in medical devices. In: 2017 IEEE International Symposium
on Hardware Oriented Security and Trust (HOST), pp. 106–109. IEEE (2017)
50. Xu, F., Qin, Z., Tan, C.C., Wang, B., Li, Q.: Imdguard: securing implantable
medical devices with the external wearable guardian. In: 2011 Proceedings IEEE
INFOCOM. IEEE, pp. 1862–1870 (2011)
51. Yang, Y., Wu, L., Yin, G., Li, L., Zhao, H.: A survey on security and privacy issues
in internet-of-things. IEEE Internet Of Things J. 4(5), 1250–1258 (2017)
52. Zaeri-Amirani, M., Afghah, F., Mousavi, S.: A feature selection method based on
shapley value to false alarm reduction in icus a genetic-algorithm approach. In:
2018 40th Annual International Conference of the IEEE Engineering in Medicine
and Biology Society (EMBC), pp. 319–323. IEEE (2018)
FAMID: False Alarms Mitigation in IoMT Devices 217
53. Zanetti, R., Aminifar, A., Atienza, D.: Robust epileptic seizure detection on wear-
able systems with reduced false-alarm rate. In: 2020 42nd annual international
conference of the IEEE Engineering in Medicine & Biology Society (EMBC), pp.
4248–4251. IEEE (2020)
54. Zhou, Y., et al.: A contrastive learning approach for icu false arrhythmia alarm
reduction. Sci. Rep. 12(1), 4689 (2022)
Dynamic Task Allocation and Scheduling
for Energy Saving in Edge Nodes
for IoT Applications
1 Introduction
In IoT systems, along with timeliness, response time, and waiting delay, energy
consumption is also a significant design consideration. The usage of an edge
layer with multiple servers is now been proposed to help in satisfying these timing
constraints. An Edge server can be in an active or inactive state consuming static
and/or dynamic energy. Static energy is due to leakage current and dynamic
energy is due to switching current. The inactive state duration and the energy
consumed by the edge server during this period can be further reduced if it can
be transited from an inactive state to a state that consumes an infinitesimally
small amount of static energy. This can be done by shutting down the unutilized
processing elements. To do this, the benefits achieved by transiting to a lower-
energy state have to be more than the energy consumed in this decision-making,
shutting down and waking up the processing elements, leading to a threshold
time. Thus if the edge server remains inactive for a longer duration than the
threshold, static energy consumption can be reduced, thus reducing the overall
energy consumption. The inactive duration can be increased by postponing the
less critical task as much as possible i.e. completing it just before its deadline.
If the edge server remains idle for less than the threshold duration, it is better
to share the workload from other edge servers and transit to an active state. By
doing this, there is a possibility for other servers to remain in a lower-energy
state for a long time.
In a multi-Edge (ME) system, there is a fragmentation of task occupancy
of any Edge Server. One of the reasons is the variations in actual execution
time (AET) which is usually less than or the same as the worst-case execution
time (WCET). Due to this, the schedule has small inactive intervals. Merging the
active intervals leads to a reduction in the number of active and inactive intervals
which effectively produces longer inactive intervals. One of the techniques used
is the Dynamic Procrastination technique, which achieves this merging without
missing any deadlines [1,2]. For homogeneous multi-core systems, job migration
and other techniques have been used to improve the performance [3]. The energy
saving is proportional shutdown period achieved [4,5].
In this work, we propose an energy-saving technique for IoT applications
with hard real-time periodic workload generation on a homogeneous Multi-Edge
(HME) system connected with dedicated links having multi-mode energy levels.
The scheduler produces a valid schedule with optimal energy per productive
work by minimizing inactive intervals of all edge servers with the help of DP,
DVFS, and controlled job migration.
The rest of the paper is organized as follows: Sect. 2 discusses the work done
in the areas related to DP, DVFS, and migration techniques for energy saving.
In Sect. 3 the proposed technique is explained. Section 4 details the experimental
evaluation and results. Section 5 concludes the paper with future directions.
220 S. K. Gawali et al.
2 Related Work
Recent literature on hard real-time schedulers using various slowdown and
shutdown techniques is analyzed in this section. We also explore these tech-
niques when combined with job migration. Researchers address leakage current
energy management i.e. saving the static energy while scheduling by consider-
ing temperature-dependent leakage on a processor with the help of accumulated
execution slack while some modulate the core between active and sleep states
by combining DPS and Real-Time Calculus for computing the core idle intervals
[6–8]. To combine idle durations ES-RHS uses harmonization for unicore and MC
systems [9]. In [10], the optimization goal is to minimize time delay, but power
consumption is not considered. [11] explains Dynamic Computation Offloading
for Mobile-Edge Computing with Energy Harvesting Devices. [12] studied the
tradeoff between time delay and energy consumption of mobile devices. Other
approaches include DVSLK which merges the scattered idle intervals [13], Single
Frequency Approximation (SFA) with procrastination [14], Critical Speed DVS
with Procrastination (CS-DVS-P) [15], systems without constraints on maxi-
mum processor speed [16]. Here SFA and CS-DVS-P show good performance in
all jobs execution with WCET. [17], studied the potential of dynamic procrasti-
nation with AET.
Job migrations have been used for load balancing, but can also exploit for
improving response time and energy saving. In an ME system, in addition, to
which Edge-Server to migrate the jobs other aspects to be considered are when
and which jobs are to be considered for migration. In the ME system, the inac-
tive/shutdown intervals being spread across multiple Edge servers makes opti-
mal energy efficiency difficult. The push and pull migration strategies used by
Linux can help in merging these intervals [18]. A polynomial-time scheduling
algorithm for minimizing energy consumption with job migration was proposed
by [16]. Energy savings in a multi-core system using the push-procrastinate-pull
(Pcube) technique have been studied in [19]. Here we extend their Pcube tech-
nique further with Dynamic Voltage/Frequency scaling (DVFS) technique along
with migration for optimal energy consumption. [20] used DVFS technique to
develop a hierarchical scheduling algorithm called Slack Aware Frequency Level
Allocator (SAFLA) for saving energy while scheduling multiple real-time peri-
odic task graphs on heterogeneous systems.
Researchers working on energy-saving techniques in edge computing plat-
forms try to balance the other parameters like system throughput, latency,
network delay, edge server utilization, and task success rate along with energy
parameters. We explored the literature that uses various strategies for energy
savings in edge computing. One of the reasons for task failure in edge comput-
ing is less energy with the edge nodes to which the task was allocated. Thus
task allocation is significant in edge computing. [21] proposed a hybrid MAC
protocol-based adaptable sleep mode and demonstrated the effectiveness that
improves the network throughput and enhances energy conservation. [22] pro-
posed a framework called EASE for job scheduling and migration within the edge
hosts with distributed renewable energy resources. EASE uses distributed con-
Dynamic Task Allocation and Scheduling for Energy Saving 221
sensus step, to reach the migration agreement. [23] proposed online upload sched-
uler, named FReshness-aware Energy efficient ScHeduler (FRESH), to minimize
the update energy consumption subject to information freshness constraints. [24]
provide insights on the optimal look-ahead time for energy prediction. They for-
mulated an energy-aware scheduler for battery-less IOT devices using the Mixed
Integer Linear Program.
In addition to slowdown and shutdown techniques we look at optimal deci-
sions on migration to reduce energy consumption. Our focus is not only on
reducing the number of idle intervals but also on reducing the idle duration.
This is achieved by migrating the upcoming jobs as their arrival times are known
due to their periodic nature. In push migration, We call the edge server from
where the job is pushed as a source node and to which it is pushed as a target
node. Similarly, in pull migration, vice-versa nomenclature is used. Although
push migration increases the active duration in the target edge node, it helps in
increasing the idle duration in the source edge node. This can potentially extend
to a duration suitable for the shutdown. Similarly, pull migration increases the
utilization of the source edge node and increases inactive duration on the target
edge node. Migration can thus help in increasing the inactive duration for both
pull and push mechanisms. Whenever such migrations do not help in achieving
the shutdown threshold, instead of remaining idle, the edge server executes the
tasks at low frequencies thus saving dynamic energy. Our proposed technique
results in the creation of a schedule combining the DP and DVFS with migra-
tion. We show that our proposed method reduces overall energy consumption. By
optimally making use of affinity features of the ME system with multiple energy-
state support, our proposed algorithm schedules the tasks on edge servers with
optimal energy consumption.
3 Problem Statement
The optimization problem aims to minimize the overall energy consumption of
the multi-edge system having multi-mode energy levels that supports migration.
For the source edge server from where the jobs are migrated, maximize the
shutdown duration and for the target edge server having insufficient duration
for shutdown to which the jobs are migrated, minimize the idle duration. The
idle duration is produced due to variations in AET. The optimization problem
is not only to reduce the idle duration but also the number of idle intervals.
The overall energy consumption is the sum of energy consumed at active, idle,
and shutdown state energy of all the edge servers. This includes the overhead
energy caused by edge server state transitions from active to shutdown and
wakeup which is to be considered while deciding upon the shutdown threshold.
Due to scaled-down frequency, the active duration increases but the energy in
this duration remains low due to low voltage. Thus overall energy consumption
reduces.
222 S. K. Gawali et al.
4 Proposed Technique
The optimization problem aims to minimize the overall energy consumption of
the homogeneous multi-edge system supporting multi-mode energy levels. For
the source edge server from where the jobs are migrated, maximize the shutdown
duration and for the target edge server having insufficient duration for shutdown
to which the jobs are migrated, minimize the idle duration. The idle duration
is produced due to variations in AET. The optimization problem is not only to
reduce the idle duration but also the number of idle intervals. The overall energy
consumption is the sum of energy consumed at active, idle, and shutdown state
energy of all the edge servers. This includes the overhead energy caused by
edge server state transitions from active to shutdown and wakeup which is to
be considered while deciding upon the shutdown threshold. Due to scaled-down
frequency, the active duration increases but the energy in this duration remains
low due to low voltage. Thus overall energy consumption reduces.
Our proposed scheduler follows slowdown and shutdown techniques along
with migration for less energy consumption. Thus named Slowdown or Shut-
down with Migration (SoSM). Whenever it is possible for the edge server to
be transited from an idle state to a shutdown state in the future, it executes
the jobs at full voltage and frequency. To qualify for shutdown, it postpones
the upcoming job execution using the Dynamic Procrastination (DP) technique
and migrates the jobs to other Edge Servers. When it is not possible to push
the jobs to other Edge Servers, the unproductive inactive time is utilized by
pulling the jobs from other Edge Servers to aid in shutting down the other Edge
Servers. If the server still continues to have an idle state, our scheduler SoSM
executes the non-critical jobs at the lowest possible voltage and frequency to save
energy. This technique is called Dynamic Voltage/Frequency technique (DVFS).
Migration with DP and DVFS helps in increasing the overall Multi-Edge system
utilization and shutdown duration resulting in reducing energy consumption.
At the beginning, the scheduler considers the worst-case execution time
(WCET) of jobs to find the active and idle durations. If the idle duration is
not large enough for shutdown, it finds jobs that can be pushed to other Edge
Servers - Push. The non-migratable jobs are then procrastinated to increase the
idle duration -Procrastinate. If after the actions it does not have a duration large
enough to qualify for shutdown, instead of remaining idle, it pulls jobs from other
edge servers -Pull. If there are not enough jobs to pull, it finds the appropriate
voltage and frequency for job executions and scales down the voltage/frequency
accordingly.
When the Edge node becomes idle, SoSM uses Push - Procrastinate - Pull
policy [19] to decide whether to keep the edge node idle or shut it down. This
effectively combines idle/shutdown intervals to longer shutdown duration using
aggressive procrastination and migration. When the edge node wakes up, the
scheduler computes the appropriate voltage and frequency for task execution.
Algorithm 1 shows our scheduler SoSM.
Dynamic Task Allocation and Scheduling for Energy Saving 223
5 Experimental Evaluation
We designed a framework for finding the resultant schedule of a task set
along with measuring various energy parameters like inactive, static, dynamic,
and total energy consumption for the state-of-the-art Dynamic Procrastination
(DPS) and DVFS technique along with the proposed scheduler Shutdown or
Slowdown with Migration (SoSM). Each edge node is allocated a set of randomly
generated tasks using a first-fit bin packing approximation algorithm based on
CPU utilization. The utilization of each edge node is varied between 20% to
90%. The simulations for utilization percentages ranging from 225% to 355% in
a set of 3 and 4 edge nodes are performed. The following ranges for mean,
µ=(WCET+AET)/2, and standard deviation = (WCET–AET)/number of
tasks is given where WCET is the worst-case execution time and AET is the
actual execution time. The AET of the task is varied between 5% and 100% of
its WCET in steps of 5%. For the analysis purpose, the schedules are generated
for one hyper period because the situation of all the periodic tasks is ready at
the same time repeats at the least common multiple of the periods which is
called a hyper period. The total energy consumption (Etot ) during execution is
measured by considering energy components like Static energy (Estat ), Dynamic
energy (Edyn ), Scheduler decision making energy (Edm ), processor shutdown and
wakeup energy (Epsd ).
Jejurikar et al. [15] have given a set of energy parameters which we use in our
simulations as shown in Table 1 [19]. We take the maximum frequency at 1 V to
be 3.1 GHz with 0.43 nF of capacitance [15]. For procrastination with migration
decisions, we take 2 µJ. For the idle duration to qualify for shutdown, the energy
saved must be more than the energy spent shutting down and waking up the
edge server i.e. the static and dynamic energy consumed during idle duration
must be more than the overhead energy of state transition. If T is the shutdown
threshold, (22 nJ + 11 nJ) T > 483 µJ. Thus we have considered T as the time
equivalent to 15000 CPU cycles.
Step 1: Compute the idle duration at the end of the active period.
Step 2: If (idle duration > shutdown threshold)
return state = shutdown and SV = Max voltage
Else Find the list of pushable jobs.
Find the procrastinated idle duration (PID)
by considering the pushable jobs and
procrastination of non-pushable jobs using DP.
If (PID > shutdown threshold)
return state = shutdown and SV = Max voltage
Else Find the pushable jobs from the end of the active period.
This will create extra idle duration (EID).
If (EID + PID > shutdown threshold)
return state = shutdown and SV = Max voltage
Else Find the list of pullable jobs from other edge servers.
Find the appropriate scaled voltage/frequency using DVFS.
If the list of pullable jobs is non-empty
return state = active and SV = scaled voltage
Else return state = idle and SV = scaled voltage
On Event: Empty ready queue i.e when the edge node is idle
Input: ready queue of other edge servers at time t.
Output: state idle/shutdown and scaled voltage (SV)
Step 1: Find the PID by considering the pushable jobs and procrastination of
non-pushable jobs using DP.
Step 2: If (PID > shutdown threshold)
return state = shutdown and SV = zero
Else Find the list of pullable jobs from other edge servers.
Find the appropriate scaled voltage/frequency using DVFS.
If the list of pullable jobs is non-empty
return state = active and SV = scaled volatge
Else return state = idle and SV = minimum voltage
Dynamic Task Allocation and Scheduling for Energy Saving 225
6 Experimental Results
We compare the results of our proposed technique SoSM with state-of-the-art
energy-saving techniques Dynamic Procrastination (DPS) and DVFS schedulers,
combined DPS and DVFS (SoS) i.e. without migration in Figs. 1 and 2.
We find that the utilization of migration in our proposed technique results in
increased shutdown duration. A possible reduction in the chances of core shut-
down occurs due to the scaled active duration. Figure 1 shows that on average
SoSM reduces shutdown duration by 2.76% over DPS and increases by 0.74%
over SoS schedules.
On average, SoSM reduces static energy by 46% over DVFS. Since static
energy is inversely proportional to the shutdown duration, this implies that on
average, our proposed technique increases the static energy by 2.88% over DPS
and reduces it by 0.81% over the SoS algorithm.
On average the dynamic energy consumption of SoSM is more than the DVFS
by 4.5% over and less than by 5% and 1.34% over DPS and SoS algorithms
respectively.
We find that on average, SoSM reduces the total energy consumption by
18.6%, 2.3%, and 1.2% over DVFS, DPS, and SoS algorithms respectively Fig. 2.
226 S. K. Gawali et al.
7 Conclusion
For IoT applications using Edge layer computations having Edge servers con-
nected with dedicated links, energy efficiency is an important criterion for a
successful deployment. In the future, the conservation of energy and the long
life of deployed systems is likely to be deciding factor for large-scale adoption.
A vision of the future world has all household and industrial devices being part
of some IoT network. Satisfying the requirements of collecting, analyzing, and
transmitting data requires optimized and energy-efficient Edge servers.
In this work, we have proposed a dynamic task allocation and scheduling tech-
nique for periodic tasks on multilayer IoT systems, having Edge Servers that can
conserve energy by staying in a very low-energy state like shutdown. Along with
the primary constraint of timeliness on real-time tasks, our proposed scheduler
achieves a reduction in overall energy consumption. It achieves enhancement in
overall shutdown duration with the help of migration, dynamic procrastination of
tasks, and dynamic voltage/frequency scaling of task execution. Migration helps
to utilize any unused idle duration of Edge servers by executing the jobs from
other Edge servers and increasing the possibility of a shutdown in Edge servers.
This results in reduced static and dynamic energy consumption of the system. In
the future, with the deployment of Embedded devices such as Edge Servers, the
use of an effective task-scheduling mechanism will result in conserving energy
and an increase in the lifetime of IoT deployments.
References
1. Liu, B., Foroozannejad, M.H., Ghiasi, S., Baas, B.M.: Optimizing power of many-
core systems by exploiting dynamic voltage, frequency and core scaling. In: IEEE
58th International Midwest Symposium on Circuits and Systems (MWSCAS), pp.
1–4 (2015)
Dynamic Task Allocation and Scheduling for Energy Saving 227
2. Chen, Y.L., Chang, M.F., Liang, W.Y., Lee, C.H.: Performance and energy efficient
dynamic voltage and frequency scaling scheme for multicore embedded system. In:
IEEE International Conference on Consumer Electronics (ICCE), pp. 58–59 (2016)
3. Keng-Mao, C., Chun-Wei, T., Yi-Shiuan, C., Chu-Sing, Y.: A High Performance
Load Balance Strategy for Real-Time Multicore Systems. Hindawi Publishing Cor-
poration, The Scientific World Journal (2014)
4. Devdas, V., Aydin, H.: On the interplay of voltage/frequency scaling and device
power management for frame-based real-time embedded applications. In: IEEE
Transactions on Computers, pp. 31–44 (2012)
5. Legout, V., Jan, M., Pautet, L.: A scheduling algorithm to reduce the static energy
consumption of multiprocessor real-time systems. In: 21st ACM International con-
ference on Real-Time Networks and Systems (RTNS), pp. 99–108 (2013)
6. Yang, C.Y., Chen, J.J., Thiele, L., Kuo, T.W.: Energy-efficient real-time task
scheduling with temperature-dependent leakage. In: Design, Automation Test in
Europe Conference Exhibition (DATE), pp. 9–14 (2010)
7. Awan, M., Petters, S.: Enhanced race-to-halt: a leakage-aware energy management
approach for dynamic priority systems. In: 23rd Euromicro Conference on Real-
Time Systems (ECRTS), pp. 92–101 (2011)
8. Huang, K., Santinelli, L., Chen, J., Thiele, L., Buttazzo, G.: Periodic power man-
agement schemes for real-time event streams. In: Proceedings of the 48th IEEE
Conference on Decision and Control (CDC) with 28th Chinese Control Confer-
ence, pp. 6224–6231 (2009)
9. Rowe, A., Lakshmanan, K., Zhu, H., Rajkumar, R.: Rate-harmonized scheduling
and its applicability to energy management. In: IEEE Transactions on Industrial
Informatics, pp. 265–275 (2010)
10. Liu, J., Mao, Y., Zhang, J., Letaief, K.B.: Delay-optimal computation task schedul-
ing for mobile-edge computing systems. In: IEEE International Symposium on
Information Theory (ISIT), pp. 1451–1455 (2016)
11. Mao, Y., Zhang, J., Letaief, K.B.: Dynamic computation offloading for mobile-edge
computing with energy harvesting devices. In: IEEE Journal on Selected Areas in
Communications, pp. 3590–3605 (2016)
12. Mao, Y., Zhang, J., Letaief, K.B.: Joint task offloading scheduling and transmit
power allocation for mobile-edge computing systems. In: IEEE Wireless Commu-
nications and Networking Conference (WCNC), pp. 1–6 (2017)
13. Han, L., Canon, L.-C., Liu, J., Robert, Y., Vivien, F.: Improved energy-aware
strategies for periodic real-time tasks under reliability constraints. In: IEEE Real-
Time Systems Symposium (RTSS), pp. 17–29 (2019)
14. Pagani, S., Chen, J.J.: IEEE 19th International Conference on Embedded and Real-
Time Computing Systems and Applications (RTCSA), Energy efficiency analysis
for the Single Frequency Approximation (SFA) Scheme, pp. 82–91 (2013)
15. Jejurikar, R., Pereira, C., Gupta, R.: Leakage aware dynamic voltage scaling for
real time embedded systems. In: 41st IEEE Design Automation Conference, pp.
275–280 (2004)
16. Chen, J.J., Heng-Ruey, H., Kai-Hsiang, C., Chia-Lin, Y., Ai-Chun, P., Tei-Wei, K.:
Multiprocessor energy-efficient scheduling with task migration considerations. In:
Proceedings of 16th Euromicro Conference on Real-Time Systems (ECRTS), pp.
101–108 (2004)
17. Gawali, S., Raveendran, B.: DPS: a dynamic procrastination scheduler for multi-
core/multi-processor hard real time systems. In: IEEE International Conference
on Control, Decision and Information Technologies (CoDIT), pp. 286–291 (2016)
228 S. K. Gawali et al.
18. Yu, K., Yang, Y., Xiao, H., Chen, J.: An improved DVFS algorithm for energy-
efficient real-time task scheduling. In: IEEE 22nd International Conference on High
Performance Computing and Communications, IEEE 18th International Confer-
ence on Smart City. IEEE 6th International Conference on Data Science and Sys-
tems (HPCC/SmartCity/DSS), pp. 263–272 (2020)
19. Gawali, S.K., Goveas, N.: P3: A task migration policy for optimal resource uti-
lization and energy consumption. In: 2022 IEEE 1st International Conference on
Data, Decision and Systems (ICDDS), pp. 1–6 (2022)
20. Roy, S.K., Devaraj, R., Sarkar, A.: SAFLA: scheduling multiple real-time periodic
task graphs on heterogeneous systems. In: IEEE Transactions on Computers, pp.
1067–1080 (2023)
21. Al-Janabi, T.A., Al-Raweshidy, H.S.: An energy efficient hybrid MAC protocol with
dynamic sleep-based scheduling for high-density IoT networks. In: IEEE Internet
of Things Journal, pp. 2273–2287 (2019)
22. Perin, G., Meneghello, F., Carli, R., Schenato, L., Rossi, M.: EASE: energy-aware
job scheduling for vehicular edge networks with renewable energy resources. In:
IEEE Transactions on Green Communications and Networking, pp. 339–353 (2023)
23. Zhang, L., Yan, L., Pang, Y., Fang, Y.: FRESH: freshness-aware energy-efficient
Scheduler for cellular IoT systems. In: IEEE International Conference on Commu-
nications (ICC), pp. 1–6 (2019)
24. Delgado, C., Famaey, J.: Optimal energy-aware task scheduling for batteryless IoT
devices. In: IEEE Transactions on Emerging Topics in Computing, pp. 1374–1387
(2022)
Deep Learning Based Framework
for Forecasting Solar Panel Output Power
1 Introduction
The dependency transition from non-renewable energy sources such as oil, nat-
ural gas, coal, and nuclear energy to renewable energy sources such as solar,
mechanical vibration, kinetic, thermal, and wind has been hastened in various
sectors for electricity generation in response to the growing need of restriction
on non-renewable energy uses in power production. Solar energy is one of the
most promising and prominent renewable energy source due to its high energy
density, ubiquitous nature, and cost-effectiveness [1]. Due to recent advance-
ments in Photovoltaic (PV) technology, the use of solar panels and solar cells is
not only restricted to power grids but also has captured immense popularity in
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 229–239, 2023.
https://doi.org/10.1007/978-3-031-45878-1_16
230 P. Mohanty et al.
– A hybrid DL model based on CNN, LSTM, and the attention mechanism has
been proposed.
– Attempts have been made to implement various DL models such as LSTM,
CNN, Auto–LSTM, LSTM–attention, Gated Recurrent Units (GRU), CNN–
LSTM with the same dataset for forecasting the output power of solar panel
to have a comparative analysis of the proposed model.
– A lightweight model of the proposed DL model has been developed using the
pruning technique which can be implemented on low-end microprocessors.
Deep Learning Based Framework for Forecasting Solar Panel Output Power 231
Further, the related prior work has been discussed in Sect. 2. The method-
ology and proposed model are described in Sect. 3 and 4, respectively. Results
and discussions have been presented in Sect. refsec5. The manuscript has been
concluded in Sect. 6.
3 Methodology
Figure 1 represents the functional methodology of the proposed solar panel out-
put power forecasting framework. The dataset for this work has been taken from
232 P. Mohanty et al.
Kaggle’s website [16]. The raw data need preprocessing, which includes data
cleaning, which typically deals with finding the null values and handling the
missing data and data transformations. The preprocessed data is split into train-
ing and testing data in the ratio of 80:20. The DL model itself can extract the
significant features from the dataset. The model iteratively learns during training
process and tunes the hyperparameters to reduce the error while increasing the
forecasting accuracy. Subsequently, the testing data has been used to evaluate
the performance of the model. The proposed DL model can not run in a low-
end microprocessor/microcontroller. Thus, a lightweight model can be developed
in order to implement it in the field to forecast the solar panel output power.
The lightweight model has lesser trainable parameters compared to the original
model thus, it is supposed to consume lesser computation power compared to
the original model.
4 Proposed Work
h7 h2 h1
c7 c6 c2 c1 c0
Repeang Repeang Repeang
Forecasted Aenon h7 h6 h2 h1 h0
Network Network Network
Output Power Layer
x2 x1
x7
LSTM Network
Figure 2 depicts the proposed hybrid CNN and LSTM with an attention
mechanism. The employed CNN extracts meaningful insights about the spa-
Deep Learning Based Framework for Forecasting Solar Panel Output Power 233
error values, low trainable parameters and thus low computational complexity.
The performance of the model is evaluated using four different error metrics,
such as Mean Absolute Error (MAE), Mean Square Error (MSE), Root Mean
Square Error (RMSE), coefficient of determination and (R2 ). The mathemati-
cal formulae for calculating MAE, MSE, RMSE, and R2 are mentioned below.
Figure 6 illustrates the graph between the actual output power of the solar panel
against the predicted output power of the solar panel. It has been plotted only
for 250 samples in the testing data to visualize the difference between predicted
and original values.
N
1
M SE = ( y i − pi ) 2 (1)
N i=1
1 N
RM SE = ( y i − pi ) 2 (2)
N i=1
N
1
M AE = |yi − pi | (3)
N i=1
N
( y i − pi ) 2
i=1
R2 = 1 − (4)
N
( yi − pmean ) 2
i=0
A lightweight model has been developed using pruning technique that reduces
the trainable parameters. It can be noted that a large number of weights is near
zero in a trained fully connected neural network. These insignificant weights
whose values are near zero are removed using pruning technique. The number of
trainable parameters is decreased as a result without decreasing accuracy. This
lightweight model can be deployed into a low-end microprocessor such as rasp-
berry pi to incorporate the proposed framework in field. The proposed model has
been compared with various DL models which are majorly implemented to pre-
dict output power of solar panel. Table 1 summarises the comparative analysis of
the proposed model with similar models. The error metrics is graphically shown
in Fig. 7. Thus, the proposed light weight model can be directly implemented on
off-the-shelf microprocessors and deployed on field to forecast the harvestable
solar energy in short term. Along with the microprocessor, a low-power Data
Acquisition (DAQ) system can be designed and deployed to measure the neces-
sary environmental parameters, which task is to provide real-time data to the
proposed model allowing it to be scaled up for use on a real-world test bench.
The proposed model can receive the data from the DAQ and process them to
forecast the energy that can be harvested in future.
Deep Learning Based Framework for Forecasting Solar Panel Output Power 235
Fig. 4. Correlation between output power of solar panel and various parameters (a)
Output power vs. angle of incidence of the solar panel (b) Output power vs. shortwave
radiation (c) Total cloud cover vs. output power (d) Output power vs. humidity (e)
Output power vs. temperature
236 P. Mohanty et al.
Fig. 5. Correlation between output power of solar panel and various parameters (a)
Loss curve (b) Layers analysis of the proposed model
6 Conclusion
In this manuscript, a hybrid DL model for forecasting the output power of solar
panel has been proposed. The proposed model uses two layers of CNN hybridized
with one layer of LSTM followed by an attention layer. CNN is used for feature
extraction, and LSTM with an attention layer predicts future output power. The
model has been analyzed using four error metrics such as MAE, MSE, RMSE,
and R2 . The proposed model is compared with various popular DL models used
to forecast solar panels in most of the literature. It outperforms other models by
achieving maximum R2 of 0.845. A lightweight model has also been developed
for the proposed model using the pruning technique, which has less number of
trainable parameters and possesses almost the same accuracy. The lightweight
model can be used to implement the framework on low-end microprocessors to
forecast the output power of solar panels in field.
238 P. Mohanty et al.
References
1. Mohanty, P., Pati, U., Mahapatra, K.: Self-powered intelligent street light man-
agement system for smart city. In: IEEE 18th India Council International Con-
ference (INDICON), Guwahati, India, pp. 1–6 (2021). https://doi.org/10.1109/
INDICON52576.2021.9691575
2. Yagil, G., Yang, D., Srinivasan, D.: Automatic hourly solar forecasting using
machine learning models. Renew. Sustain. Energy Rev. 105, 487–498 (2019).
https://doi.org/10.1016/j.rser.2019.02.006
3. Yao, G., Lei, T., Zhong, J.: A review of convolutional-neural-network-based action
recognition. Pattern Recogn. Lett. 118, 14–22 (2019). https://doi.org/10.1016/j.
patrec.2018.05.018
4. Rao, G., Huang, W., Feng, Z., Cong, Q.: LSTM with sentence representations
for document-level sentiment classification. Neurocomputing 308, 49–57 (2018).
https://doi.org/10.1016/j.neucom.2018.04.045
5. Tao, Y., Chen, H., Qiu, C.: Wind power prediction and pattern feature based on
deep learning method. In: IEEE PES Asia-Pacific Power and Energy Engineering
Conference (APPEEC), pp. 1–4. (2014). https://doi.org/10.1109/APPEEC.2014.
7066166
6. Gensler, A., Henze, J., Sick, B., Raabe, N.: Deep Learning for solar power fore-
casting - an approach using AutoEncoder and LSTM Neural Networks. In: IEEE
International Conference on Systems, Man, and Cybernetics (SMC), Budapest,
Hungary, pp. 002858–002865 (2016). https://doi.org/10.1109/SMC.2016.7844673
7. AlKandari, M., Ahmad, I.: Solar power generation forecasting using ensemble app-
roach based on deep learning and statistical methods. Appli. Comput. Inform.
(2019). https://doi.org/10.1016/j.aci.2019.11.002
8. Wang, K., Qi, X., Liu, H.: A comparison of day-ahead photovoltaic power fore-
casting models based on deep learning neural network. Appl. Energy 251, 1–14
(2019). https://doi.org/10.1016/j.apenergy.2019.113315
9. Li, G., Xie, S., Wang, B., Xin, J., Du, S.: Photovoltaic power forecasting with a
hybrid deep learning approach. IEEE Access 8, 175871–175880 (2020). https://
doi.org/10.1109/ACCESS.2020.3025860
10. Li, P., Zhou, K., Lu, X., Yang, S.: A hybrid deep learning model for short-term
PV power forecasting. Appl. Energy 259, 1–11 (2020). https://doi.org/10.1016/j.
apenergy.2019.114216
11. Jebli, I., Belouadha, F., Kabbaj, M., Tilioua, A.: Deep learning based models for
solar energy prediction. Adv. Sci. Technol. Eng. Syst. J. 6(1), 349–355 (2021).
https://doi.org/10.25046/aj060140
12. Liu, C., Gu, J., Yang, M.: A simplified LSTM neural networks for one day-ahead
solar power forecasting. IEEE Access 9, 17174–17195 (2021). https://doi.org/10.
1109/ACCESS.2021.3053638
13. Montoya, A., Mandal, P.: Day-ahead and week-ahead solar PV power forecast-
ing using deep learning neural networks. In: North American Power Sympo-
sium (NAPS), Salt Lake City, UT, USA, pp. 1–6 (2022). https://doi.org/10.1109/
NAPS56150.2022.10012199
Deep Learning Based Framework for Forecasting Solar Panel Output Power 239
14. Chen, M., Chiang, H., Chang, C.: Solar photovoltaic power generation prediction
based on deep learning methods. In: IET International Conference on Engineering
Technologies and Applications (IET-ICETA), Changhua, Taiwan, pp. 1–2 (2022).
https://doi.org/10.1109/IET-ICETA56553.2022.9971676
15. Machina, S., Koduru, S., Madichetty, S.: Solar energy forecasting using deep learn-
ing techniques. In: 2nd International Conference on Power Electronics & IoT Appli-
cations in Renewable Energy and its Control (PARC), Mathura, India, pp. 1–6
(2022). https://doi.org/10.1109/PARC52418.2022.9726605
16. Solar energy power generation dataset. https://www.kaggle.com/datasets/stucom/
solar-energy-power-generation-dataset, (Accessed 1 May 2023)
AI and Big Data for Next-G Internet
of Medical Things (IoMT)
EHR Security and Privacy Aspects: A
Systematic Review
1 Introduction
– Access Control: Access control is the process of dealing with the situation who
can access EHRs and what actions they can perform. It includes authenti-
cation, authorization, and accountability. Authentication verifies the identity
of the user, authorization determines what resources the user can access, and
accountability ensures that the actions of the user are recorded for auditing
purposes [7].
– Data Encryption: Data encryption is the process of converting plaintext data
into ciphertext to prevent unauthorized access. Encryption ensures that the
data is secure during transmission and storage, making it unreadable to unau-
thorized users [30].
– Auditing: Auditing is the process of recording and monitoring EHR access and
use. Auditing helps to detect and investigate any unauthorized access, mod-
ification or disclosure of EHRs, ensuring compliance with regulatory require-
ments and standards [24].
– Risk Management: Risk management refers to the series of activities aimed at
identifying, evaluating, and minimizing risks related to the security and pri-
vacy of electronic health records (EHRs). This involves the creation of policies
and procedures to manage EHRs, training employees on how to maintain EHR
security and privacy, and implementing various technical controls such as fire-
walls, intrusion detection systems, and prevention mechanisms. Through this
process, organizations can effectively safeguard the confidentiality, integrity,
and availability of their EHRs, while complying with relevant laws and regu-
lations governing the handling of medical information [31].
The protection of Electronic Health Records (EHRs) poses significant legal and
ethical challenges that demand thoughtful contemplation. Patients are entitled
to manage their health information, which encompasses the ability to regulate
who can retrieve it and for what reason. Conversely, healthcare providers have
a responsibility to secure patients’ information from being accessed or revealed
EHR Security and Privacy Aspects: A Systematic Review 245
1.4 Motivation
EHR security and privacy are critical aspects of modern healthcare, requiring
a multi-disciplinary approach that involves healthcare providers, IT specialists,
and regulatory bodies. Effective implementation of security and privacy measures
in EHR systems is crucial for protecting patient privacy and ensuring timely
access to critical medical information for quality healthcare delivery. In conclu-
sion, protecting EHRs requires constant attention and vigilance, and healthcare
providers must remain up-to-date with the latest security and privacy measures
to protect their patients’ sensitive personal information.
This work is structured into several sections for clarity and organization. In
Sect. 2, we will explore the role of digitalization in the healthcare sector. Section 3
will delve into the importance of ensuring the security and privacy of Electronic
Health Records (EHRs) in the context of federated learning. Major challenges in
this area will be analyzed in Sect. 4, followed by an examination of the current
state of the art in Sect. 5. Section 6 will analyze the limitations of some existing
approaches. In Sect. 7, we will propose federated learning-based solutions for
ensuring security and privacy in the healthcare sector. Finally, in Sect. 8, we will
discuss possible directions for future work.
Digital technology has become more and more important in healthcare inno-
vation and has introduced several tools and methods for improving healthcare
services. These measures consist of maintaining secure storage of patient informa-
tion in a centralized location and implementing software that enhances the avail-
ability of health-related data for patients. However, the digitization of health-
care is still in its early stages, and several multidimensional problems need to be
addressed.
Healthcare organizations are adopting digital technologies to improve per-
formance and efficiency, save costs, and increase efficacy. This trend is fueled
by the availability of cost-effective and energy-efficient equipment and software,
as well as the success of high-profile projects in many countries. Digital health
systems can be especially beneficial in low-income countries, helping organiza-
tions achieve cost savings and improve healthcare delivery, which is critical in
the time-reactive nature of healthcare.
A periodic survey is carried out by the World Health Organization (WHO) to
gather information on the scope and structure of healthcare digitization across
different countries. However, a recent report on digital healthcare innovation
246 S. Banerjee et al.
Table 1. Some solutions to enhance privacy and security in the healthcare sector.
actors for identity theft, insurance fraud, and other criminal activities. Moreover,
the unauthorized disclosure or misuse of EHRs can harm patients’ reputations,
cause emotional distress, and lead to physical harm.
In FL, multiple institutions collaborate on a machine learning model without
sharing their sensitive data. This collaborative approach can provide significant
benefits, such as improved accuracy, reduced bias, and faster model training.
However, it also introduces new security and privacy risks, such as data breaches,
data poisoning attacks, and model inversion attacks.
Therefore, it is critical to ensure the security and privacy of EHRs in FL
to protect patients’ sensitive personal information. This requires implementing
robust security and privacy measures, such as access control, data encryption,
auditing, and risk management. Additionally, participants in FL must comply
with various regulatory requirements, such as the Health Insurance Portability
and Accountability Act (HIPAA), which sets national standards for protecting
the privacy and security of patient’s health information.
In summary, protecting EHRs in FL is essential to maintain public trust,
protect patient privacy, and comply with regulatory requirements. Healthcare
providers, data scientists, and regulatory bodies must work together to imple-
ment effective security and privacy measures to protect patient’s sensitive per-
sonal information [8,35].
4 Major Challenges
use and sharing of their data can be challenging, especially given the complexity
of the healthcare system and the varied interests of different stakeholders.
Electronic Health Record (EHR) security and privacy are critical aspects that
need to be addressed to protect patient health information. Here are some of the
state-of-the-art measures that are being implemented to ensure EHR security
and privacy:
Access control: EHR systems use access control mechanisms to ensure that
only authorized personnel can access patient health information. These mech-
anisms include password-protected logins, two-factor authentication, and role-
based access controls [7].
Encryption: Encryption is used to protect data stored in EHR systems, mak-
ing it unreadable to unauthorized users. Encryption can be applied to data both
in transit and at rest.
Audit trails: EHR systems maintain an audit trail of all access and modifica-
tions to patient health information. This allows organizations to track who has
accessed the information and when, and to detect any unauthorized access or
modifications.
Data backup and recovery: Electronic health record (EHR) systems are cru-
cial in maintaining patient health information, and backups of this data are regu-
larly created to safeguard against system failure or cyberattacks. These backups
serve as a failsafe mechanism and are frequently tested to guarantee their effi-
ciency in restoring data. In case of a disaster, EHR backups can be relied upon
to restore vital information, ensuring continuity of care and patient safety.
Data minimization: EHR systems implement data minimization principles,
meaning that they collect only the minimum amount of information necessary
to provide patient care. This helps reduce the risk of data breaches and protects
patient privacy.
Regular vulnerability assessments: EHR systems undergo regular vulnerabil-
ity assessments to identify and address potential security weaknesses. This helps
prevent security breaches and ensures that the EHR system remains secure over
time.
Employee training and awareness: EHR systems implement training and
awareness programs to ensure that employees are aware of the security and pri-
vacy policies and procedures. This helps prevent accidental breaches of patient
health information.
Compliance with regulations: EHR systems comply with relevant regulations
and standards, such as HIPAA and GDPR. This ensures that patient health
information is protected and that organizations are not subject to legal or finan-
cial penalties.
So EHR security and privacy are essential components of healthcare IT sys-
tems, and these state-of-the-art measures are crucial to ensuring that patient
health information is protected from unauthorized access, use, and disclosure.
252 S. Banerjee et al.
Table 2. Some solutions to enhance privacy and security in the healthcare sector.
Table 2. (continued)
The framework ensures encryption and anonymity of IoHT data using differ-
ential privacy (DP) and was evaluated for COVID-19 patient data using deep
learning applications, showing potential for widespread adoption.
Wang et al. [34] proposed a smart healthcare framework, known as FRESH,
that aims to facilitate the sharing of physiological data while ensuring data
privacy. This framework leverages Federated Learning (FL) and ring signature
defence to protect against source inference attacks (SIAs).
The data collection process for FRESH involves gathering data from wear-
able devices and processing it using edge computing devices for local training
of machine learning models. The model parameters are subsequently uploaded
to the central server for joint training. The authors utilized ring signature tech-
nology to hide the source of parameter updates, which significantly reduces the
success rate of SIAs. They also introduced a batch verification algorithm to
improve the efficiency of signature verification.
According to the authors, FRESH is highly suitable for large-scale smart
healthcare systems that cater to multiple users. This framework represents a
major milestone in the quest to enhance data privacy and security in the health-
care industry.
8 Future Works
The healthcare industry regards the privacy and security of patient data as
essential issues, and continuous investigation is imperative to enhance current
protocols and confront evolving obstacles. A potential research direction could
be to investigate the effectiveness of current privacy and security regulations,
such as HIPAA and GDPR, in protecting patient information. This research
could examine the gaps and limitations in the existing regulatory framework
and propose recommendations for improvements.
Another potential avenue of research is to investigate how emerging tech-
nologies like artificial intelligence and blockchain can improve privacy and secu-
rity within the healthcare industry. For example, blockchain technology offers a
decentralized and tamper-proof platform for storing and sharing patient infor-
mation, which could reduce the risk of data breaches and ensure the accuracy of
health records. Similarly, artificial intelligence can be used to detect and prevent
potential security breaches and unauthorized access to patient information.
Furthermore, research could be conducted on the impact of privacy and secu-
rity breaches on patient trust and healthcare outcomes. A breach of patient
information can lead to a loss of trust between patients and healthcare providers,
which can have long-lasting effects on patient health and well-being. Thus, under-
standing the effects of privacy and security breaches and developing strategies
to restore patient trust could be a valuable research direction.
256 S. Banerjee et al.
they are more likely to be aware of the risks associated with the use of personal
health information and are more likely to take steps to protect it. Healthcare
providers can encourage patient engagement by providing clear and concise infor-
mation about privacy and security policies, as well as by offering patient educa-
tion resources, such as online portals, educational videos, and other materials.
Healthcare providers can promote patient privacy and security, as well as enable
informed decision-making regarding healthcare, by actively engaging patients in
safeguarding their personal information [20,22,32].
10 Conclusion
In conclusion, digitalization offers significant opportunities and advantages to
the healthcare sector, but its implementation requires careful consideration of
the social, organizational, and collaborative aspects of the workplace. Even if
digitization can enhance healthcare performance and accomplish strategic goals,
it is crucial to be aware of potential difficulties and constraints. By focusing on a
positive attitude and providing necessary support, healthcare organizations can
successfully adopt digital technologies and streamline their procedures, resulting
in improved patient care and cost savings.
References
1. Ahmed, M.I., Kannan, G.: Secure and lightweight privacy preserving internet of
things integration for remote patient monitoring. J. King Saud Univ.-Comput.
Inform. Sci. 34(9), 6895–6908 (2022)
2. Alam, M.A., Al Riyami, K.: Shear strengthening of reinforced concrete beam using
natural fibre reinforced polymer laminates. Constr. Build. Mater. 162, 683–696
(2018)
3. Alloghani, M., Al-Jumeily, D., Hussain, A., Aljaaf, A.J., Mustafina, J., Petrov,
E.: Healthcare services innovations based on the state of the art technology trend
industry 4.0. In: 2018 11th International Conference on Developments in eSystems
Engineering (DeSE), pp. 64–70. IEEE (2018)
4. Bhowmik, T., Banerjee, I.: Eeppda-edge-enabled efficient privacy-preserving data
aggregation in smart healthcare internet of things network. Inter. J. Network
Manag. e2216 (2023)
5. Chae, B.: Mapping the evolution of digital business research: a bibliometric review.
Sustainability 14(12), 6990 (2022)
6. Chen, C.M., Chen, Z., Kumari, S., Lin, M.C.: Lap-ioht: a lightweight authentication
protocol for the internet of health things. Sensors 22(14), 5401 (2022)
7. Dagher, G.G., Mohler, J., Milojkovic, M., Marella, P.B.: Ancile: privacy-preserving
framework for access control and interoperability of electronic health records using
blockchain technology. Sustain. Urban Areas 39, 283–297 (2018)
8. Dang, T.K., Lan, X., Weng, J., Feng, M.: Federated learning for electronic health
records. ACM Trans. Intell. Syst. Technol. (TIST) 13(5), 1–17 (2022)
9. Das, D., Banerjee, S., Chatterjee, P., Ghosh, U., Biswas, U.: A secure blockchain
enabled v2v communication system using smart contracts. IEEE Trans. Intell.
Trans. Syst. (2022)
10. Das, D., Banerjee, S., Chatterjee, P., Ghosh, U., Mansoor, W., Biswas, U.: Design of
a blockchain enabled secure vehicle-to-vehicle communication system. In: 2021 4th
International Conference on Signal Processing and Information Security (ICSPIS),
pp. 29–32. IEEE (2021)
11. Das, D., Banerjee, S., Chatterjee, P., Ghosh, U., Mansoor, W., Biswas, U.: Design
of an automated blockchain-enabled vehicle data management system. In: 2022 5th
International Conference on Signal Processing and Information Security (ICSPIS),
pp. 22–25. IEEE (2022)
EHR Security and Privacy Aspects: A Systematic Review 259
12. Das, D., Banerjee, S., Dasgupta, K., Chatterjee, P., Ghosh, U., Biswas, U.:
Blockchain enabled sdn framework for security management in 5g applications.
In: 24th International Conference on Distributed Computing and Networking, pp.
414–419 (2023)
13. Das, S., Namasudra, S.: Lightweight and efficient scpprivacy-preserving/scp mutual
authentication scheme to secure scpinternet of things/scp-based smart healthcare.
Trans. Emerging Telecommun. Technol. (2023)
14. Dutta, K., Guin, R.B., Chakrabarti, S., Banerjee, S., Biswas, U.: A smart job
scheduling system for cloud computing service providers and users: modeling and
simulation. In: 2012 1st international conference on recent advances in information
technology (rait), pp. 346–351. IEEE (2012)
15. Garg, N., Wazid, M., Das, A.K., Singh, D.P., Rodrigues, J.J., Park, Y.: Bakmp-
iomt: design of blockchain enabled authenticated key management protocol for
internet of medical things deployment. IEEE Access 8, 95956–95977 (2020)
16. Herrmann, M., Boehme, P., Mondritzki, T., Ehlers, J.P., Kavadias, S., Truebel,
H.: Digital transformation and disruption of the health care sector: Internet-based
observational study. J. Med. Internet Res. 20(3), e104 (2018)
17. Kigera, J., Kipkorir, V.: Electronic health records-the ethical and legal issues.
Annals African Surgery 20(1), 1–2 (2023)
18. Kumar, V., Mahmoud, M.S., Alkhayyat, A., Srinivas, J., Ahmad, M., Kumari, A.:
Rapchi: robust authentication protocol for iomt-based cloud-healthcare infrastruc-
ture. J. Supercomput. 78(14), 16167–16196 (2022)
19. Lahiri, P.K., Das, D., Mansoor, W., Banerjee, S., Chatterjee, P.: A trustworthy
blockchain based framework for impregnable iov in edge computing. In: 2020 IEEE
17th International Conference on Mobile Ad Hoc and Sensor Systems (MASS), pp.
26–31. IEEE (2020)
20. Lahiri, P.K., Mandal, R., Banerjee, S., Biswas, U.: An approach towards devel-
opments of smart covid-19 patient’s management and triaging using blockchain
framework (2020)
21. Li, H., et al.: Review on security of federated learning and its application in health-
care. Futur. Gener. Comput. Syst. 144, 271–290 (2023)
22. Mandal, R., Banerjee, S., Islam, M.B., Chatterjee, P., Biswas, U.: Qos and energy
efficiency using green cloud computing. In: Intelligent Internet of Things for Health-
care and Industry, pp. 287–305. Springer (2022). https://doi.org/10.1007/978-3-
030-81473-1_14
23. Manogaran, G., Thota, C., Lopez, D., Sundarasekar, R.: Big data security intel-
ligence for healthcare industry 4.0. Cybersecurity for Industry 4.0: Analysis for
Design and Manufacturing, pp. 103–126 (2017)
24. Parker, M.: Managing threats to health data and information: toward security. In:
Health Information Exchange, pp. 149–196. Elsevier (2023)
25. Rahman, M.A., Hossain, M.S., Islam, M.S., Alrajeh, N.A., Muhammad, G.: Secure
and provenance enhanced internet of health things framework: a blockchain man-
aged federated learning approach. IEEE Access 8, 205071–205087 (2020)
26. Roski, J., Bo-Linn, G.W., Andrews, T.A.: Creating value in health care through big
data: opportunities and policy implications. Health Aff. 33(7), 1115–1122 (2014)
27. Roy, R., Haldar, P., Das, D., Banerjee, S., Biswas, U.: A blockchain enabled
trusted public distribution management system using smart contract. In: Inter-
national Conference on Electronic Governance with Emerging Technologies, pp.
25–35. Springer (2022). https://doi.org/10.1007/978-3-031-22950-3_3
28. Ryu, J., et al.: Secure ecc-based three-factor mutual authentication protocol for
telecare medical information system. IEEE Access 10, 11511–11526 (2022)
260 S. Banerjee et al.
29. Singh, R., Dwivedi, A.D., Srivastava, G., Chatterjee, P., Lin, J.C.W.: A privacy
preserving internet of things smart healthcare financial system. IEEE Internet of
Things J. (2023)
30. Sonkamble, R.G., Bongale, A.M., Phansalkar, S., Sharma, A., Rajput, S.: Secure
data transmission of electronic health records using blockchain technology. Elec-
tronics 12(4), 1015 (2023)
31. Tertulino, R., Antunes, N., Morais, H.: Privacy in electronic health records: a
systematic mapping study. J. Public Health, 1–20 (2023)
32. Tiwari, S., et al.: Applications of machine learning approaches to combat covid-19:
a survey. In: Lessons from COVID-19, pp. 263–287 (2022)
33. Wang, S., Kirillova, K., Lehto, X.: Travelers’ food experience sharing on social
network sites. J. Travel Tourism Market. 34(5), 680–693 (2017)
34. Wang, W., Li, X., Qiu, X., Zhang, X., Zhao, J., Brusic, V.: A privacy preserv-
ing framework for federated learning in smart healthcare systems. Inform. Proc.
Manag. 60(1), 103167 (2023)
35. Xu, J., Glicksberg, B.S., Su, C., Walker, P., Bian, J., Wang, F.: Federated learning
for healthcare informatics. J. Healthcare Inform. Res. 5, 1–19 (2021)
36. Yang, C., Everitt, J.H., Murden, D.: Evaluating high resolution spot 5 satellite
imagery for crop identification. Comput. Electron. Agric. 75(2), 347–354 (2011)
37. Yin, X., Zhu, Y., Hu, J.: A comprehensive survey of privacy-preserving federated
learning: a taxonomy, review, and future directions. ACM Comput. Surv. (CSUR)
54(6), 1–36 (2021)
38. Yu, S., Park, Y.: A robust authentication protocol for wireless medical sensor net-
works using blockchain and physically unclonable functions. IEEE Internet Things
J. 9(20), 20214–20228 (2022). https://doi.org/10.1109/JIOT.2022.3171791
39. Zhao, Y., et al.: Growth traits and sperm proteomics analyses of myostatin gene-
edited Chinese yellow cattle. Life 12(5), 627 (2022)
SNN Based Neuromorphic Computing Towards
Healthcare Applications
1 Introduction
Medical healthcare is an area that can significantly benefit from the integration of neuro-
morphic computing, AI, and ML. Neuromorphic computing is a revolutionary computer
paradigm that mimics the operation of the human brain. It uses spiking neural networks
(SNNs), Deep Neural Networks (DNNs), and in some cases, Recurrent Neural Net-
works (RNNs) to process and analyze data more efficiently and faster than traditional
computing systems. AI and ML are also becoming increasingly important in healthcare,
allowing for analyzing large amounts of data and developing predictive models. In the
proposed work, we explore the applications of these technologies in medical healthcare
and their potential to transform the field. Artificial Neural Networks (ANNs) are widely
used in a variety of applications such as image recognition, audio identification, and
processing of natural languages. Traditional ANNs, on the other hand, are unsuitable
for low-power and real-time applications because to their high computational complex-
ity and energy consumption. Because of their excellent energy economy and processing
capacity, SNNs have emerged as a potential replacement to standard ANNs. In this paper,
we comprehensively study the theory and applications of SNNs [1].
Brain-Computer Interfaces (BCIs): SNNs can be used to develop BCIs, which enable
patients to communicate with computers or other devices using of their brain signals.
SNNs can be used to decode the electroencephalogram (EEG) signals generated by the
patient’s brain and translate them into specific commands.
Epilepsy Prediction: SNNs can be used to predict epileptic seizures in patients with
epilepsy. SNNs can learn to recognize the patterns in EEG signals that indicate an
imminent seizure and issue an alert to the patient or caregiver.
Parkinson’s Disease Diagnosis: SNNs can be used to diagnose Parkinson’s disease
by analyzing the patient’s gait. SNNs can learn to recognize the characteristic changes
in gait associated with Parkinson’s disease and issue a diagnosis.
Drug Discovery: SNNs can be used to accelerate the drug discovery process by pre-
dicting the efficacy and toxicity of candidate drugs. SNNs can be trained on large datasets
of molecular structures and their corresponding biological activities and used to predict
the activity of new compounds.
Medical Image Analysis: SNNs can be used to analyze medical images, such as MRI
and CT scans. SNNs is used to detect and classify abnormalities, such as tumors, and
aid in the diagnosis of diseases.
Medical Robotics: SNNs can be used to control medical robots, such as surgical
robots. SNNs can learn to recognize the desired surgical actions and movements and
issue commands to the robot.
Overall, SNNs have the potential to revolutionize medical healthcare by enabling
faster and more accurate diagnosis, prediction, and treatment of diseases [5]. The salient
contribution of the proposed work is as follows:
1. Spiking Neural Network’s (SNN) biological spiking behavior is used to generate
desired results.
264 P. Maji et al.
2. The proposed work validates a Neuromorphic approach for various disease detection.
3. The proposed Neuromorphic framework considerably speeds up the process of
evaluating required outputs accurately.
The remainder of the work is organized as follows: Sect. 2 contains a literature
review, Sect. 3 contains a proposed method and data for the application of neuromorphic
systems, and Sect. 4 contains a Dataflow Diagram. Part 4 discusses the findings, and
Sect. 5 concludes with a conclusion.
2 Literature Review
Table 1. Comparative table for different neuromorphic approaches in the medical field with
resultant.
immense. Table 1. clearly reflect the comparisons among the work done in the field of
neuromorphic systems and application area.
The proposed work defines and trains a spiking neural network to perform binary classi-
fication on the Pima Indian’s diabetes and Wisconsin Cancer datasets. The US National
Institute of Diabetes and Digestive and Kidney Diseases first gathered the data, then
made it available to the general public for research. It is a dataset that is frequently used
in the study of machine learning and predictive modeling. The dataset has 768 instances,
each with 8 attributes. Dr. William H. Wolberg originally gathered the breast cancer data
at the University of Wisconsin Hospital in Madison, Wisconsin, and made it accessible
to the general public for research. There are 569 instances in the collection, and each
instance has 32 attributes.
The spiking neural network is defined using the Nengo framework. It consists of three
layers: a dense layer with 16 units, a layer of leaky integrate-and-fire (LIF) neurons,
and a dense output layer with a single sigmoid activation unit. The input and output
nodes are defined as well.
The network is then compiled using the NengoDL simulator and trained using the
fit method with the mean squared error loss and a stochastic gradient descent optimizer.
The training is performed for the number of epochs, and we attain the best in 10 epochs.
Finally, the trained model is used to predict the training data and compute the clas-
sification accuracy. The predictions are threshold at 0.5 to obtain binary labels, and the
accuracy is computed as the mean of the correct predictions.
In this proposed work, we explain with the diabetic dataset only; however, we have tested
the proposed model with both the said dataset and also explain in the result section. We
first load the Pima Indians Diabetes dataset from the UCI repository and split it into
input features (X) and output labels (y). We then scale the input features to the range {0,
1} to ensure that all the features are on a similar scale.
We define a spiking neural network using the Nengo and NengoDL frameworks. The
network has two dense layers and a spiking LIF layer. We then compile the model using
NengoDL’s Simulator object and train the model on the dataset.
Once the SNN framework is installed and the data is preprocessed, we begin to build
the SNN. The steps involved in building an SNN model for diabetes prediction are as
follows.
Here’s the step-wise detailed working principle of the algorithm:
SNN Based Neuromorphic Computing Towards Healthcare Applications 267
Step1. Load the Pima Indians Diabetes dataset from a URL and
store the column Names in a list.
Step 2. Load the dataset into a Pandas DataFrame.
Step 3. Split the dataset into input features (X) and output labels
(y).
Step 4. Scale the input features to the range [0, 1].
Step 5. Define a spiking neural network using the Nengo library
5.1. Create a Nengo network.
5.2. Define the input and output nodes.
5.3. Create the first dense layer with 16 units using
the NengoDL Layer.
5.4. Create the second dense layer with 16 units using
the Nengo LIF Neuron model.
5.5. Create the final output layer with a sigmoid
activation Function using the NengoDL layer.
5.6. Create probes to record the output of the final
layer.
Step 6. Configure the NengoDL settings to make the network non-
trainable.
Step 7. Compile the model using the MSE loss function and
With a learning rate of 0.001 along with Stochastic Gradient
descent optimizer
Step 8. Train the model using the training data for 10 epochs.
Step 9. Predict new data using the trained model.
The spiking neural network has two dense layers, each with 16 units, and a final
output layer with a sigmoid activation function. The LIF neuron model is used in the
second layer. The model is trained using the MSE loss function and SGD optimizer with
a learning rate of 0.001 for 10 epochs. The final classification accuracy of the model is
printed to the console, which is more than 90% for the diabetes dataset and 98.5% in
cancer dataset.
The investigations occurred on the Google Colab platform and also in Jupyter Notebook
with Python 3.7 and Keras. Machine learning techniques have been applied using the
“SKLearn” library. Moreover,we monitor the binary cross-entropy with Adam Optimizer
and MSE metric function. For SNN, we employ the Nengo library. The capacity of
Nengo to simulate massive spiking neural networks is one of its key advantages. For that
purpose, a network of interconnected neurons is defined, and each neuron is represented
using LIF neurons (Fig. 2).
268 P. Maji et al.
P1 Trained Adapted
ReLU Neutron Weights Weights
P2
Train Conv1 S_Conv1
Convm S_Convm
Somax
Funcon Pn FC1 S_FC1
Vector Convoluonal & Pooling Fully Connected
Matrix Based Feature Extracon Classificaon FCn S_FCn
Mapping
Biomedical
Compung Engine based on Neuromorphic Spike
Input Signals
O1
0 1 1 0 0
1 1
Spike Encoder
0 00
00 0
Result
0 01 O2 Spike
1 1 11 1
01 01 0 10 Counter
0 0 0 0
1 0 0 1 1
Infer O3
IF Neuron Grid IF Neuron Grid
When assessing the effectiveness of regressors, we also utilize metrics like mean
squared error (MSE) or mean absolute error (MAE) in place of accuracy. The perfor-
mance of the model improves as these metrics’ values decrease. When predicting a
continuous parameter (such as blood sugar level) in regression assignments like the one
in this code, precision is typically not an essential factor. The ROC curve, used to assess
the effectiveness of binary classification models, cannot be calculated because the code
provided is for a regression problem. The ROC curve is used to calculate the trade-off
between true and false favourable rates at various categorization levels. Braindrop is
also used, which is an open-source mixed-signal neuromorphic architecture designed
270 P. Maji et al.
5 Conclusion
Classification of neuron modelling approaches, comparison of the investigated chips,
and identification of present trends and limits impeding the development of neuromor-
phic technologies for medical applications. The application of Neuromorphic systems in
biological healthcare and signal processing holds great potential for medical profession-
als, practitioners and their patients. Neural Networks specially SNNs can be utilized to
improve the quality of life for chronically sick patients by allowing ambient observations
for anomalies, hence reducing the strain on medical resources. In the proposed work,
we can implement SNN for healthcare data with up to 98.5% accuracy compared to
other ML algorithms. Proper application can result in decreased workloads for medical
practitioners, allowing them to focus on time-critical jobs that demand a quality higher
than what neural networks can attain now.
When combined with Neuromorphic hardware, an SNN-based application performs
better than conventional ML methods. Our observation is that neuromorphic computing
have the potential to significantly enhance medical applications and the optimization
profession. Each of these applications is likely to benefit from the massively parallel,
event-driven, and/or stochastic processes of neuromorphic computers when combined
with neural networks.
References
1. Bellec, G., Salaj, D., Subramoney, A., Legenstein, R., Maass, W.: A solution to the learning
dilemma for recurrent networks of spiking neurons. Nat. Commun. 11(1), 1–13 (2020)
2. Nunes, J.D., Carvalho, M., Carneiro, D., Cardoso, J.: Spiking neural networks: a survey. IEEE
Access. 10 (2022). https://doi.org/10.1109/ACCESS.2022.3179968
3. Liu, J., Wu, T., Ma, X., Zhang, Y., Hu, J.: A survey on deep learning-based neuromorphic
computing. Front. Neurosci. 15, 655935 (2021)
4. Koo, M., Srinivasan, G., Shim, Y., Roy, K.: sBSNN: stochastic-bits enabled binary spiking
neural network with on-chip learning for energy efficient neuromorphic computing at the
Edge. IEEE Trans. Circ. Syst. I: Regular Papers, 1–10 (2020). https://doi.org/10.1109/TCSI.
2020.2979826
5. Li, Z., Tang, W., Zhang, B., Yang, R., Miao, X.: Emerging memristive neurons for neuromor-
phic computing and sensing. Sci. Technol. Adv. Mater. 24(1), 2188878 (2023). https://doi.
org/10.1080/14686996.2023.2188878.PMID:37090846;PMCID:PMC10120469
6. Schuman, C., et al.: A Survey of Neuromorphic Computing and Neural Networks in
Hardware (2017)
7. Esser, S.K., Merolla, P.A., Arthur, J.V., et al.: Convolutional networks for fast, energy-efficient
neuromorphic computing. PNAS 113(41), 11441–11446 (2016)
8. Tavanaei, A., Ghodrati, M., Kheradpisheh, S.R., Masquelier, T., Maida, A.: Deep learning in
spiking neural networks. Neural Netw. 111, 47–63 (2019). https://doi.org/10.1016/j.neunet.
2018.12.002, ISSN 0893–6080
SNN Based Neuromorphic Computing Towards Healthcare Applications 271
9. Qiao, N., et al.: Reconfigurable online learning spiking neuromorphic processor comprising
256 neurons and 128K synapses. Front. Neurosci. 9, 141 (2015)
10. Livi, P., Indiveri, G.: A current-mode conductance-based silicon neuron for address-event
neuromorphic systems. In: IEEE International Symposium on Circuits and Systems, Taipei,
Taiwan 2009, pp. 2898–2901 (2009). https://doi.org/10.1109/ISCAS.2009.5118408
11. Yamakawa, T., et al.: Wearable epileptic seizure prediction system with machine-learning-
based anomaly detection of heart rate variability. Sensors 20, 3987 (2020). https://doi.org/10.
3390/s20143987
12. Pereira, S., Pinto, A., Alves, V., Silva, C.A.: Brain tumor segmentation using convolutional
neural networks in MRI images. IEEE Trans. Med. Imaging 35(5), 1240–1251 (2016). https://
doi.org/10.1109/TMI.2016.2538465
13. Michaelis, C., Lehr, A.B., Oed, W., Tetzlaff, C.: Brian2Loihi: an emulator for the neuromor-
phic chip Loihi using the spiking neural network simulator brian. Front. Neuroinform. 9(16),
1015624 (2022). https://doi.org/10.3389/fninf.2022.1015624
14. Hatem, S.M., et al.: Rehabilitation of motor function after stroke: a multiple
systematic review focused on techniques to stimulate upper extremity recovery.
Front. Hum. Neurosci. 13(10), 442 (2016). https://doi.org/10.3389/fnhum.2016.00442.
PMID:27679565;PMCID:PMC5020059
15. Klietz, M., Bronzlik, P., Nösel, P., et al.: Altered neurometabolic profile in early parkinson’s
disease: a study with short echo-time whole brain MR spectroscopic imaging. Front. Neurol.
17(10), 777 (2019). https://doi.org/10.3389/fneur.2019.00777
16. Andreou, A.G.: Real-time sensory information processing using the TrueNorth Neurosynaptic
System. In: IEEE International Symposium on Circuits and Systems (ISCAS), Montreal, QC,
Canada 2016, pp. 2911 (2016). https://doi.org/10.1109/ISCAS.2016.7539214
17. Navamani, T.M.: Deep Learning and Parallel Computing Environment for Bioengineering
Systems (2019)
18. Wei, O., Shitao, X., Chengyu, Z., Wenbao, H., Qionglu, Z.: An overview of brain-like com-
puting: Architecture, applications, and future trends. Front. Neuro. 16 (2022). https://doi.org/
10.3389/fnbot.2022.1041108, ISSN=1662–5218
Crossfire Attack Detection in 6G
Networks with the Internet of Things
(IoT)
1 Introduction
As the internet continues to evolve and become more essential to daily lives
than ever, there is a growing population looking to destroy it. Denial of service
(DoS) attacks have been a threat to the internet for years and continue to cause
issues even in today’s networks. The continued improvement and introduction
of internet-of-things (IoT) devices have pushed mobile carriers to update their
existing infrastructure to support the increased number of devices. With the
use of Internet of Medical Things (IoMT) servers, securing the network has
become even more critical. If an adversary were to disrupt the operations of
IoMT devices, entire hospitals and healthcare networks may be affected, leading
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 272–289, 2023.
https://doi.org/10.1007/978-3-031-45878-1_19
Crossfire Detection in IoT 273
also suffers from the fact that oftentimes these routes are non-ideal, meaning
legitimate traffic is degraded at the expense of security.
The summary of contributions are:
– The paper proposes a statistical detection model for crossfire attacks using
Analysis of Variance (ANOVA) and neural networks.
– The proposed mechanism only uses packet headers and not packet content
to determine if a packet is adversarial that achieves an accuracy of 95.3% in
detecting these packets
– We evaluated the proposed defense mechanism on a real-world network topol-
ogy from the ATT North America backbone topology using the Mininet sim-
ulation environment.
The rest of the article is organized as follows: Sect. 2 discusses some back-
ground information about IoT, IoMT, DoS attacks, and crossfire attacks.
Section 3 details the threat model of the attack. Section 4 discusses defense strate-
gies against the crossfire attack. Section 5 Shows the execution of the crossfire
attack.
2 Background
2.1 6G
The Internet of Things has its own distinct architecture that works with the
Internet. Typically IoT is categorized into 4 layers. Figure 1 details the types of
devices that are present on each layer. The perception layer is the lowest level of
the IoT architecture. The perception layer contains the sensor-enabled physical
objects which act as endpoints to the IoT ecosystem. The next layer, the network
layer, consists of various communication protocols, edge computing, and network
connectivity. This layer transfers information securely from IoT end-points to a
276 N. Perry and S. Bhunia
processing device. The middleware layer receives data from the network layer
and stores it in a database. Cloud computing servers and database management
systems are typically middleware devices that allow applications and sensors to
connect. Finally, the top layer of the four-layer IoT architecture is the application
layer. This layer IoT exists as a result of many technologies. These technologies
work together to create a holistic system that is able to communicate across the
internet. Radio frequency identification (RFID) was a large precursor to IoT as
it allowed machines to record metadata, recognize objects, and control devices
through radio waves [8].
1. On-Premises - The edge services should be located at the edge of the network,
meaning it should be able to run isolated from the core of the network
2. Proximity - By being close to the source of the data/information, MEC is
useful for analytics and data collection
3. Lower Latency - By being closer to the edge devices, latency is considerably
reduced. This can be used to reduce latency or improve user experience.
4. Location Awareness - When connected to WiFi or cellular, services can use
low-level signaling to determine the location of connected devices.
5. Network Context Information - Real-time network statistics can be used by
applications to provide context-specific services that can be monetized and
change the experience of mobile communication.
Mobile edge computing can be used in many sectors to offload core ser-
vices. Augmented reality (AR) systems typically require high computational
power. Many users use (AR) on their mobile devices, so computations have to
be offloaded to servers. Edge computing would allow these high-demand, low-
latency tasks to remain at the edge of the network [1]. Edge computing also
will play a key role with respect to web performance and caching HTML con-
tent. By deploying content delivery servers at the edge, HTTP requests would
travel through these servers that would handle many of these requests, reducing
traffic across the core network [1]. MEC services allow 5G to continue to work
towards the core goal of “zero-latency” as reducing congestion in the core allows
more traffic to be routed. This in turn improves the experience for users of 5G
technology.
Crossfire Detection in IoT 277
Cell phone
Company
Server
Game
Console
downstream traffic is affected [7]. The crossfire attack is more difficult to detect,
as crossfire is an indirect attack, contrary to most other attacks. Since the attack
spreads low-intensity traffic to various destinations, this allows the attack traffic
to blend in with legitimate traffic and is virtually undetectable in standard DoS
detection and mitigation protocols, at least until after substantial damage has
been done [7].
To execute a crossfire attack, first, a potential adversary would select a list
of public servers within the targeted areas and a set of decoy servers away from
the target area. Since these servers are publicly accessible, they can be easily
found. Next, the adversary would have to perform recognisance and generates
a link map. This link map would be a map of layer 3 links that connect their
decoy servers to public servers. The map enables the adversary to select a set of
“target servers” that when flooded, would be able to cut off the target area from
the open internet. The adversary then coordinates the decoy servers to flood the
target link, effectively blocking most flows to the target area. Each individual
server sends low-intensity traffic so as to not arouse suspicion of anomaly-based
detection system. The individual low-intensity flows are indistinguishable from
legitimate traffic. Finally, the adversary begins to flood each target link one at
a time to not trigger automatic packet route mutation [7].
MTD. One proposed crossfire defense solution is the moving target defense
(MTD). Traditional networks are static in nature, allowing attacks to spend as
much time as needed to gather information and find vulnerabilities [4]. MTD
has historically been a warfare strategy but recently has been adopted into the
IT world. A moving target defense can first delay network mapping and recon-
naissance. ICMP and UDP scans can be disrupted by imposing fake hosts and
services on random ports that do not exist. The fake listeners can increase the
time and workload an attacker would need in order to launch an attack [6]. Addi-
tionally, a mutable network can be created that changes IP addresses and ports
of network hosts dynamically. This means that machines will not be able to be
located at the same address at any given time [4]. Oftentimes, MTD takes the
form of random route and address mutations. Randomization has been a com-
mon strategy in moving target defenses. This strategy is based on the idea that
if addresses of targets within the network constantly change or access policies
between the attacker and target change, then the attack success rate will dras-
tically reduce. An attacker’s ability to effectively do reconnaissance is sharply
diminished as well due to the ever-changing network landscape [6]. Obfuscation
of links via SDN has also been proposed to confuse and thwart attackers [2]. By
obfuscating links between the attacker and target, an adversary would not be
able to identify common links, a key step in performing a crossfire attack.
the virtual IPs assigned by the network consistently changed frequently and
synchronously. However, the higher the rate of mutation, the more overhead is
required to run the network [4]. This type of approach is often used by load-
balancers to send traffic to different destinations depending on network load.
Another proposed way to mitigate large-scale DDoS attacks is by using a
routing around congestion (RAC) defense. The RAC defense works by rout-
ing traffic between a service deployer and a critical autonomous system around
degraded links. RAC defense asserts that attack traffic is irrelevant and does not
need to be filtered when using this defense [14]. The RAC defense offers path
isolation by dynamically creating detour routes for critical flow [15].
3 Threat Model
To successfully execute a crossfire attack, adversaries must be able to mask
malicious traffic behind legitimate traffic to avoid detection by the system. The
aforementioned mitigation techniques focus on adversaries using the same attack
nodes repeatedly. If an adversary were to have a botnet sufficiently large, they
would be able to slowly introduce attack nodes into the attack, and remove
attack nodes that have become ineffective. By staying within the confines of
thresholds, an attack could be executed for longer without being detected.
During the reconnaissance phase of the attack, an adversary would use mul-
tiple attack nodes to execute trace route commands. These commands would be
done at low intensity, and low rate to avoid route mutation from the SDN. By
spreading out these trace route commands, common links can be discovered and
mapped with minimal error due to route mutation. This would allow the adver-
sary to create a route map, and understand where secondary nodes are used
when the primary link is being flooded. For maximum efficiency, the adversary
would choose a time during routine peak demand, as SDNs would anticipate this
stress on the system, and additional strain may be attributed to regular demand
fluctuations.
Once launching the attack, the nodes would monitor the route of the low-
intensity traffic to destination servers, to ensure that the traffic is being routed
through the link node. If a node determines that it has been rerouted, it shall
continue directing traffic to the target node, and wait before disconnecting and
changing targets. This can prevent the “common denominator” defense. The
moving target defense can be leveraged in itself to disrupt legitimate traffic. By
forcing the SDN to continually change routes, legitimate traffic can be slowed
down beyond usability.
The adversary would not launch all attack nodes at once, as this may cause
a spike in demand, which the system would detect. Instead, the adversary would
gradually increase attack nodes in order to mask the increase in demand as
organic demand, thereby potentially circumventing anomaly-based detection [7].
As the attack propagates on the network, constant monitoring of network
routing would be required. As the system responds to the attack, we would
monitor the change in performance during route mutation, and when the attack
is taking place undetected. This would allow for the practicality of leveraging
route mutation-based mitigation to be measured.
4 Defense Mechanism
Since crossfire attacks are so lethal, it is important to detect when they are occur-
ring as soon as possible. Therefore, using a software defined network (SDN) is
ideal so that a holistic view of the network can be obtained. The use of SDN
is proposed as an ideal approach to obtain a holistic view of the network. In
an SDN, the control plane is decoupled from the data plane, allowing central-
ized control and management of the network. The SDN architecture consists of
Crossfire Detection in IoT 281
SDN
Controller
3. T
Sen hreat
2P t ba A
a ck ck ssess
et H to C m
ea d ont ent
ers roll
Sen er
t to
IDS
Cloud
IDS
4. Threat
Assessments
Sent to all 1. Packet Info
Switches Sent to SDN Controller
Benign
Packet Permitted
Adversarial
Packet Blocked
tion model, and a cloud-based IDS to detect and mitigate crossfire attacks. By
inspecting packet headers, identifying adversarial packets, and dropping them in
real time, the mechanism helps protect the network from the detrimental effects
of crossfire attacks, minimizing potential damage and maintaining network per-
formance. Figure 3 details the mechanism.
5 Experiment Setup
The execution of a crossfire attack, in theory, appears straightforward. By flood-
ing a specific link, the attacker aims to overwhelm it with traffic. This process
involves identifying potential routes that utilize the targeted link and generating
low-intensity requests across those routes to flood the link effectively. However,
in practice, executing a crossfire attack can be challenging due to the limited
availability of information regarding the specific routes taken by packets. The
lack of public access to this crucial routing data presents a significant hurdle
for attackers attempting to orchestrate such attacks. In this section, we delve
into the intricacies of executing a crossfire attack, exploring the methodologies
used to overcome these obstacles and the implications of this type of attack on
network performance and security.
PTLD
PTLD
NY54
NY54 CMBR
CMBR
CHCG
CHCG
SCRM
SCRM SLKC
SLKC PHLA
PHLA
SNFN
SNFN DNVR
DNVR CLEV
CLEV
KSCY
KSCY WASH
WASH
STLS
STLS
LA03
LA03 NSVL
NSVL
RLGH
RLGH
ATLN
ATLN
SNDG
SNDG PHNX
PHNX
DLLS
DLLS
NWOR
NWOR
Core Deployed
HSTN ORLD
ORLD
HSTN
Completion 2007 – 2008 SNAN
SNAN
Completed
Zoo [9] as the basis for the network configuration. The network topology, as
depicted in Fig. 4, provides a visual representation of the structure and inter-
connections of the various network components. It showcases the arrangement
of links within the simulated network. To introduce additional functionality and
explore specific scenarios, a Mobile Edge Computing (MEC) server was strategi-
cally placed between the ORLD and CLEV nodes. The MEC server served as a
centralized computing platform that brought computing resources closer to the
network edge, enabling efficient processing and analysis of data generated within
the network. In this particular setup, the MEC server had a direct connection
between the ORLD and CLEV nodes, facilitating seamless communication and
data exchange between them. The choice of the ATT North America Backbone
network from The Internet Topology Zoo [9] was driven by its complexity and
size, which allowed for a more realistic simulation of network traffic. By utiliz-
ing a network with a sufficient number of components and diverse connections,
researchers and analysts could better understand and evaluate the performance,
scalability, and security aspects of network systems under various conditions.
in this scenario, the network would only have access to information contained in
the packet headers to analyze and identify the attack.
After running the network for 5 min, the ping round trip times were collected
and plotted in Fig. 5. The moving average was plotted as a line. As the attack
continues, the ping increases on average. During the first stage, the average
ping remained around 50–100 ms. Once the zombie servers began attacking, the
average increased to about 500 ms. After the second phase, the ping increases
to about 800 ms. Finally, during the third phase the ping increases and hovers
around 1500 ms.
6 Crossfire Detection
Detecting a crossfire attack directly can be difficult. Since HTTPS encrypts the
packets, the contents of the packets cannot be inspected. Only the headers of
each packet are able to be inspected. Therefore, each model created only analyzes
the headers of the packets and makes decisions based on those headers.
After running the experiment described in Sect. 5, packet headers were col-
lected for every packet sent on the network. About 30,000 packets were collected.
After the packets were collected, the data was aggregated. First, a standard Anal-
ysis of Variance (ANOVA) was conducted. After conducting the initial ANOVA.
Crossfire Detection in IoT 285
An analysis of variance was first performed with all the data. The initial ANOVA
gave the results depicted in Table 1 and Table 2. As shown, the largest predictor
of adversarial packets were the window size, the ack rtt, and the time relative.
After running the original ANOVA, we removed any non-significant factors to
achieve the following ANOVA shown in Table 3 and Table 4. The test as a whole
is able to determine whether or not a packet is adversarial based on the window
size, ack rtt, frame time, and tcp time. This model may not be practical given
a node that consistently has a significant delay. If a node has significant delay,
all packets may be marked as adversarial. Additionally, during an attack, the
delay of packets through the congested links may present a problem where the
model detects all packets as adversarial and blocks essentially all connections,
worsening the effects of the attack.
286 N. Perry and S. Bhunia
Table 5. Confusion Matrix for Training Table 6. Confusion Matrix for Validation
Data for 1 × 3 neural network Data for 1 × 3 neural network
Predicted Predicted
Adversarial Benign Adversarial Benign
Actual Adversarial 19206 33 Actual Adversarial 6373 12
Benign 242 4928 Benign 72 1679
Table 7. Confusion Matrix for Validation Table 8. Confusion Matrix for Validation
Data with 2 × 25 neural network Data with 2 × 25 neural network
Predicted Predicted
Adversarial Benign Adversarial Benign
Actual Adversarial 19308 28 Actual Adversarial 6375 10
Benign 140 4933 Benign 31 1720
7 Conclusion
and the deployment of intrusion detection systems that inspect packet headers
can also contribute to the detection and prevention of adversarial packets. As the
number of interconnected devices continues to grow, with an estimated 60 billion
devices expected to be connected by 2025, the importance of securing critical
servers and networks cannot be overstated. Ongoing research and collaboration
among cybersecurity experts, network administrators, and device manufacturers
are essential to developing effective defense mechanisms and ensuring the unin-
terrupted operation of vital services, particularly in healthcare settings. Over-
all, protecting critical servers from DoS attacks, including the evolving cross-
fire attack, requires a multi-faceted approach that combines advanced technolo-
gies, robust security protocols, and proactive defense strategies. By addressing
these challenges and investing in cybersecurity measures, we can safeguard the
integrity and reliability of the Internet and its essential services for the benefit
of all.
References
1. Abbas, N., Zhang, Y., Taherkordi, A., Skeie, T.: Mobile edge computing: a survey.
IEEE Internet Things J. 5(1), 450–465 (2017)
2. Aydeger, A., Saputro, N., Akkaya, K., Rahman, M.: Mitigating crossfire attacks
using SDN-based moving target defense. In: 2016 IEEE 41st Conference on Local
Computer Networks (LCN), pp. 627–630 (2016). https://doi.org/10.1109/LCN.
2016.108
3. Balaji, S., Nathani, K., Santhakumar, R.: IoT technology, applications and chal-
lenges: a contemporary survey. Wirel. Pers. Commun. 108(1), 363–388 (2019).
https://doi.org/10.1007/s11277-019-06407-w
4. Gudla, C., Sung, A.H.: Moving target defense application and analysis in software-
defined networking. In: 2020 11th IEEE Annual Information Technology, Elec-
tronics and Mobile Communication Conference (IEMCON), pp. 641–646 (2020).
https://doi.org/10.1109/IEMCON51383.2020.9284847
5. Gupta, A., Jha, R.K.: A survey of 5G network: architecture and emerging technolo-
gies. IEEE Access 3, 1206–1232 (2015). https://doi.org/10.1109/ACCESS.2015.
2461602
6. Kampanakis, P., Perros, H., Beyene, T.: SDN-based solutions for moving target
defense network protection. In: Proceeding of IEEE International Symposium on a
World of Wireless, Mobile and Multimedia Networks 2014, pp. 1–6 (2014). https://
doi.org/10.1109/WoWMoM.2014.6918979
7. Kang, M.S., Lee, S.B., Gligor, V.D.: The crossfire attack. In: 2013 IEEE Sym-
posium on Security and Privacy, pp. 127–141 (2013). https://doi.org/10.1109/SP.
2013.19
8. Kavre, M., Gadekar, A., Gadhade, Y.: Internet of Things (IoT): a survey. In: 2019
IEEE Pune Section International Conference (PuneCon), pp. 1–6 (2019). https://
doi.org/10.1109/PuneCon46936.2019.9105831
9. Knight, S., Nguyen, H., Falkner, N., Bowden, R., Roughan, M.: The internet topol-
ogy zoo. IEEE J. Select. Areas Commun. 29(9), 1765–1775 (2011). https://doi.org/
10.1109/JSAC.2011.111002
10. Narayanadoss, A.R., Truong-Huu, T., Mohan, P.M., Gurusamy, M.: Crossfire
attack detection using deep learning in software defined its networks. In: 2019
Crossfire Detection in IoT 289
Abstract. The tomato has a high market value and is one of the vegetables grown
in the most significant quantity globally. Tomato plants are susceptible to diseases,
which can negatively impact the fruit’s yield and quality. Detecting these illnesses
at an early stage and their accurate identification is necessary for successfully man-
aging diseases and reducing losses. In recent years, deep learning methods such
as convolutional neural networks (CNNs) have demonstrated significant promise
in identifying plant diseases from images. This research suggested a CNN-based
strategy for detecting tomato leaf diseases using transfer learning. Transfer learn-
ing enables us to enhance the performance of our disease detection model using
a smaller dataset by leveraging pre-trained CNN models that have been trained
on large datasets. The proposed transfer learning model through Resnet50 and
Inception V3 is effective by applying it to a dataset of tomato leaf images. As a
result, a high level of accuracy is achieved and could be indulged for practical
applications in agriculture.
1 Introduction
Tomato plants are vulnerable to various diseases that can severely harm the plant’s leaves,
fruit, and overall health. These diseases may result from multiple causes, including bac-
terial, fungal, and viral infections, nutrient deficiencies, and extreme weather conditions.
Some common tomato leaf diseases include early blight, late blight, Septoria leaf spot,
bacterial spot, and tomato yellow leaf curl virus, which can cause symptoms like yellow-
ing and wilting of leaves, brown spots on leaves, stem cankers, and stunted growth [1].
Farmers can take preventative measures such as proper plant spacing, adequate watering,
and regular sanitation practices to minimize the risk of tomato leaf disease. Additionally,
early detection and timely treatment with suitable fungicides or bactericides can help
control the spread of these diseases and ensure a healthy tomato crop.
© IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 293–305, 2023.
https://doi.org/10.1007/978-3-031-45878-1_20
294 B. S. Vidhyasagar et al.
allows the model to utilize the information available better. Transfer learning can
contribute to more accurate predictions, thereby reducing the number of false positives
and negatives.
In this study, we identify tomato illnesses using two deep-learning models.
1.1 Inception v3
Google researchers in 2015 brought into the limelight Inception v3, a convolutional
neural network designed to classify images and features in a deep neural network with
over 23 layers. The architecture of the model is given in Fig. 1.
Inception v3 uses a unique module called the Inception module, which consists of
multiple convolutional filters with different sizes applied to the same input, allowing for
efficient feature extraction at different scales [6]. It also incorporates batch normalization
and regularization techniques to enhance the accuracy of the architecture.
Inception v3 has accomplished benchmark performance on several image recognition
tasks. Additionally, it has a relatively small model size compared to other deep neural
network architectures, making it efficient to train and deploy.
1.2 Resnet 50
Researchers from Microsoft devised the ResNet50 convolutional neural network archi-
tecture first used in the company’s research in 2015. It is a subset of the ResNet (Residual
Network) family of deep neural networks and was developed to enhance the training of
very deep neural networks with more than 50 levels [6]. ResNet50 uses skip connections,
allowing information to flow from earlier layers to the last layers without passing through
any non-linear transformations. This makes it much simpler to train very deep networks.
It has attained state-of-the-art performance on various image recognition tasks, such as
recognizing objects and parsing scenes, among other image recognition tasks. In addi-
tion to this, it has been demonstrated to be effective in transfer learning, which is the
296 B. S. Vidhyasagar et al.
process whereby a model that has been pre-trained can be fine-tuned on a new task using
relatively small quantities of data. The architecture of Resnet50 is described in Fig. 2.
Using these two pre-trained models, train them on the tomato leaf dataset once it has
been preprocessed, and then assess how well the trained models perform on the image
assessment set. The effective measures, such as accuracy, precision, recall, and F1-score,
will be computed to determine how successful the model is. Then, compare these two
models to the remaining four models to choose which models have the most outstanding
performance out of the six.
2 Related Works
The author primarily emphasizes creating a CNN model that can differentiate between
five distinct diseases that can affect tomato leaves. They utilized a dataset containing
1,583 images of healthy tomato leaves and images of leaves that were affected by these
diseases according to the research findings, a CNN model that included data augmenta-
tion obtained an accuracy of 98.9%, outperforming a control group that used a model that
did not have data augmentation, which received an accuracy of 91.2%. The author offers
insightful recommendations on the application of CNNs to diagnose diseases affecting
tomato leaves and emphasizes the significance of data supplementation techniques for
enhancing the model’s overall performance [1].
Kurup et al. investigate how capsule networks can be used to classify plant varieties
and detect plant diseases. They introduced a novel method that improves the accuracy
of plant disease classification and plant species recognition by combining capsule net-
works and transfer learning. The authors describe the architecture of their suggested
model, which is based on the well-known Convolutional Neural Network (CNN) archi-
tecture, ResNet-50. The authors have utilized Transfer learning by seeding the model
with weights that have already been pre-trained using the ImageNet dataset. The algo-
rithm is then fine-tuned on two different plant datasets, namely PlantVillage and Flavia,
Prediction of Tomato Leaf Disease Plying Transfer Learning Models 297
to classify plant diseases and recognize plant species. The capsule network-based model
outperformed other state-of-the-art techniques by achieving an accuracy of 98.95% for
plant disease classification on the PlantVillage dataset. The suggested model achieved an
accuracy of 96.27% when it came to recognizing plant species using the Flavia dataset
[2]. A pre-trained CNN model, the VGG16 model, has been demonstrated to perform
admirably when given the image classification task. They describe how they got rid of
the model’s last fully connected layer and used the layer’s output before it as input to
the SVM classifier has fewer layers. An overall accuracy of 95.8% was achieved on the
authors’ dataset, which the authors describe as promising results[3]. Followed by various
researchers, the Inception-v3 model is used in image categorization tasks. They describe
the training and testing process and the metrics used to evaluate the performance of their
method and fine-tuned model with the tomato leaf dataset [4].
To identify tomato diseases, the authors applied various state-of-the-art convolu-
tional neural networks (CNNs) classification network architectures, such as ResNet18,
MobileNet, DenseNet201, and InceptionV3, to a total of 18,162 plain tomato leaf images.
They describe the process of training and testing, as well as the evaluation metrics that
were used to evaluate the performance of their method [6]. A deep learning-based tech-
nique with Inception V3 architecture is employed for detecting potato leaf diseases. The
accuracy of the suggested approach for detecting five different types of potato leaf dis-
eases was 95.12%. The authors propose that this method can be used for early detection
and monitoring of potato leaf diseases, which can help reduce crop losses and increase
agricultural productivity [8].
Ahmad et al. suggest a method for detecting diseases that affect tomato leaves
using pre-trained convolutional neural networks. Four distinct models of CNNs VGG16,
ResNet50, InceptionV3, and Xception are used in addition to data augmentation strate-
gies to expand the scope of their information. The authors analyzed the performance of
their model in terms of accuracy, precision, recall, and F1 score, and they compared it
to the performance of other techniques that are considered to be state-of-the-art. They
discovered that the Xception model achieved the highest level of precision, which was
99.6% [9].
Agarwal et al. suggest a method for detecting tomato leaf diseases called ToLeD that
makes use of a convolutional neural network to hypothesize that their technique can be
adapted to identify a wider variety of plant diseases. Because the dataset that was used
in the research is on the smaller side, the generalizability and scalability of the technique
may be compromised as a result [10].
Traditional deep learning models have been trained over various plant disease eval-
uations. When Transfer learning in conjunction with deep neural networks is involved,
it has shown a par improvement in diagnosing leaf diseases in grape and mango plants.
They used two pre-trained models, namely Inception V3 and ResNet50, and then fine-
tuned them using a dataset consisting of grape and mango leaf images. Their approach
had an accuracy of 97.4% when detecting three different diseases that can affect grape
leaves and 94.6% when detecting two other diseases that can affect mango leaves [11].
A technique for detecting tomato plant diseases using Transfer learning with syn-
thetic images generated by a conditional generative adversarial network is proposed by
Abbas et al. The authors suggest that their method can be used for the early detection
298 B. S. Vidhyasagar et al.
and monitoring of tomato plant diseases, which can help increase crop yields and reduce
losses. They also suggest expanding their approach to identify other plant diseases uti-
lizing synthetic images generated by C-GAN. It is possible for the synthetic images
included in the dataset to introduce noise or biases, both of which can harm the model’s
performance. The dataset may be flawed if the synthetic images are not authentic or have
artifacts. Training is necessary for both the C-GAN and the classifier models, which are
parts of the technique. It may cause the method to become more complicated and call
for additional computational resources and training time [12].
The paper proposes a novel method called aGROdet, designed to detect plant dis-
eases and estimate leaf damage severity within an Agriculture Cyber-Physical System
implemented at the edge platform of IoT systems. A convolutional neural network-based
model trained on large publicly available datasets achieves over 97% accuracy in ini-
tial experiments. The study also addresses damage estimation challenges, such as leaf
shadows and surrounding areas [17].
The article emphasizes the significance of early detection and disease severity esti-
mation for effective disease management and prevention. The proposed solution sug-
gests a fully automated approach utilizing deep neural networks, specifically the Mask
R-CNN network, for disease detection and localization in leaf images. The method
employs image augmentation and transfer learning to enhance precision and save time.
The approach is applicable to various imaging devices, such as smartphone cameras or
low-altitude unmanned aerial vehicle (UAV) cameras. It has been demonstrated using
apple leaves as a case study [18].
3 Dataset
Each image in the Plant Village dataset is associated with a label indicating the type of
plant and the type of disease or pest. The dataset covers over 20 crops, including tomato,
potato, apple, grape, and soybean, and includes more than 80 different plant diseases
and pests. The dataset also contains additional metadata, such as geographic location,
disease severity, and time of the year the image was captured. We separated the dataset
into ten directories with ten different tomato leaf diseases with 22,998 images. The ten
directories of tomato leaf images are taken for the project and divided into training and
validation datasets. The training dataset has 18403 images, and the validation dataset
has 4595 images belonging to 10 classes. The details is depicted in Table 1.
4 Methodology
4.1 Resnet50
The dataset was obtained from Kaggle and is named plant village. Before training the
model, Images are preprocessed to ensure they were in a standard format. Firstly, resized
the images to 224x224 pixels to reduce the computational complexity of the model. Then
converted the images to RGB format to ensure they were in the same color space as the
pre-trained ResNet 50 model. Lastly, normalization of the pixel values is done to ensure
they were between 0 and 1, which helped the model converge faster during training.
To further improve the model’s performance and prevent overfitting, we augmented
the dataset using various data augmentation techniques such as random rotation, flip,
and zoom. This helped increase the dataset’s size and improve the model’s ability to
generalize to new data.
Transfer learning is employed to fine-tune the pre-trained ResNet 50 model on the
tomato disease dataset. Frozing the initial layers of the model and retraining the last few
layers to adapt to the new dataset. This approach allowed to leverage the pre-trained
weights of the ResNet 50 model, which were learned from millions of images from the
ImageNet dataset, and apply them to our tomato disease detection task. During train-
ing, experimentation with different hyperparameters, such as batch size and optimizer,
has been done to find the optimal combination with the best performance. Various 32
batch sizes and Stochastic gradient descent (SGD) optimizers are used for the proposed
model’s best hyperparameters. Finally, the model’s performance is evaluated on a hold-
out validation set by calculating accuracy, precision, recall, and F1-score metrics. This
helped to assess the model’s ability to detect tomato diseases accurately and efficiently.
By evaluating the model’s performance on a separate validation set, an evaluation on
how well the model would perform on new, unseen data could be justified. The proposed
Transfer learning architecture for Resnet is shown in Fig. 3.
300 B. S. Vidhyasagar et al.
4.2 Inceptionv3
To build a tomato disease detection model using Inceptionv3, the first step is to collect a
large dataset of tomato images [13–16]. This dataset should include images of healthy
tomatoes and tomatoes affected by various diseases, such as bacterial spots, late blight,
and early blight. The dataset should contain various tomato images captured in different
lighting conditions and angles. The images can be obtained from multiple sources, such
as online image repositories or by taking pictures of actual tomatoes. Once the dataset is
done, the images should be preprocessed to ensure consistency and reduce noise in the
data. The images should be resized to a standard size, such as 224x224, to ensure that all
images have the exact dimensions. The pixel values should be normalized to provide the
input data has a similar range of values. Image enhancement strategies, including random
rotations, flips, and other transformations, may be utilized to expand the dataset’s size
and enhance the model’s performance. Separating the dataset into a training group and
a validation set is necessary.
It has been demonstrated that the Inceptionv3 architecture is a robust convolutional
neural network (CNN) that is capable of achieving state-of-the-art outcomes when it
comes to image classification tasks. Convolutional layers, pooling layers, and ultimately
connected layers are some of the types of layers that can be found in the deep neural
network known as Inceptionv3. Because the architecture was created to process images
of various sizes and resolutions, it is particularly well-suited for detecting tomato dis-
eases. To categorize the input picture into a variety of tomato diseases, the final layer of
Inceptionv3 should be removed and replaced with a global average pooling layer, which
should then be followed by a fully connected layer that uses softmax activation. Figure 4
Represents the Inception v3 model over transfer learning.
Prediction of Tomato Leaf Disease Plying Transfer Learning Models 301
5 Results
A Comparison of accuracy and loss of over ten epochs of three models, Resnet50, Alex
net, and Lenet has been performed. The graph shows that Inceptionv3 has the highest
validation accuracy (0.9535) while Alex net has the lowest (0.5959). However, Resnt50
also had the lowest validation loss (0.1400), while Alex net had the highest (1.1171).
Resnet50 performed the best in accuracy and loss, followed by Lenet and Alex net.
However, it is essential to note that further testing and analysis would be needed to
confirm this conclusion. Figure 5 Shows the graphical representation of accuracy and
loss of the proposed model across the state of art models.
302 B. S. Vidhyasagar et al.
The bar graph comparison shows that Inceptionv3 had the highest validation accu-
racy (0.9876) while Alex net had the lowest (0.5959). However, Inceptionv3 also had
the lowest validation loss (0.0414), while Alex net had the highest (1.1171). Overall,
Inceptionv3 performed the best in accuracy and loss, followed by Lenet and Alex net.
We suggest that Inceptionv3 may be the most effective model for this task. Figure 6
Depicts the performance of the proposed architecture over benchmark models.
Prediction of Tomato Leaf Disease Plying Transfer Learning Models 303
6 Conclusion
Tomato plants are vulnerable to various leaf diseases. Finding these diseases early would
save the plant for better yield. The proposed transfer learning models Resnet50 and
Inception V3 show the model is effective in disease prediction. In turn, it opens the
door to numerous potential applications in agriculture. Transfer learning is on the run,
which reduces the training time of a model from scratch as well improves the models’
performance. The model can be further enhanced with GAN architectures and extended
for predicting various plant diseases.
304 B. S. Vidhyasagar et al.
References
1. N. K. E., K. M., P. P., A. R., V. S.: Tomato Leaf Disease Detection using Convolutional Neural
Network with Data Augmentation. In: 2020 5th International Conference on Communication
and Electronics Systems (ICCES), Coimbatore, India, pp. 1125–1132 (2020). https://doi.org/
10.1109/ICCES48766.2020.9138030
2. Kurup, R.V., Anupama, M.A., Vinayakumar, R., Sowmya, V., Soman, K.P.: Capsule Network
for Plant Disease and Plant Species Classification. In: Smys, S., Tavares, J.M.R.S., Balas,
V.E., Iliyasu, A.M. (eds.) ICCVBIC 2019. AISC, vol. 1108, pp. 413–421. Springer, Cham
(2020). https://doi.org/10.1007/978-3-030-37218-7_47
3. Arunnehru, J., Vidhyasagar, B.S., Anwar Basha, H.: Plant Leaf Diseases Recognition Using
Convolutional Neural Network and Transfer Learning. In: Bindhu, V., Chen, J., Tavares,
J.M.R.S. (eds.) International Conference on Communication, Computing and Electronics
Systems. LNEE, vol. 637, pp. 221–229. Springer, Singapore (2020). https://doi.org/10.1007/
978-981-15-2612-1_21
4. Salvi, R.S., Labhsetwar, S.R., Kolte, P.A., Venkatesh, V.S., Baretto, A.M.: Predictive analysis
of diabetic retinopathy with transfer learning. In: 2021 4th Biennial International Conference
on Nascent Technologies in Engineering (ICNTE), pp. 1–6. IEEE (2021)
5. Jebadas, D.G., Sivaram, M., M, A., Vidhyasagar, B.S., Kannan, B.B.: Histogram Distance
Metric Learning to Diagnose Breast Cancer using Semantic Analysis and Natural Language
Interpretation Methods. In: Johri, P., Diván, M.J., Khanam, R., Marciszack, M., Will, A. (eds.)
Trends and Advancements of Image Processing and Its Applications. EICC, pp. 249–259.
Springer, Cham (2022). https://doi.org/10.1007/978-3-030-75945-2_13
6. Xiaoling X., Xu, C., Nan, B.: Inception-v3 for flower classification. In: 2017 2nd International
Conference on Image, Vision and Computing (ICIVC), pp. 783–787 (2017)https://doi.org/
10.1109/ICIVC.2017.7984661
7. Kumar, R., Singh, D., Chug, A., Singh, A.P.: Evaluation of Deep learning based Resnet50 for
Plant Disease Classification with Stability Analysis. In: 2022 6th International Conference
on Intelligent Computing and Control Systems (ICICCS), Madurai, India, pp. 1280–1287
(2022). https://doi.org/10.1109/ICICCS53718.2022.9788207
8. Nawaz, M., Nazir, T., Javed, A., et al.: A robust deep learning approach for tomato plant leaf
disease localization and classification. Sci. Rep. 12, 18568 (2022). https://doi.org/10.1038/
s41598-022-21498-5
9. Chowdhury, E.H., et al.: Tomato leaf diseases detection using deep learning technique.
Technol. Agric. 453 (2021)
10. Belal A.M.A, Abu-Naser. S.S.: Image-based tomato leaves diseases detection using deep
learning (2018)
11. Mosin, H., Tanawala, B., Patel, K.J.: Deep learning precision farming: Tomato leaf disease
detection by transfer learning. In: Proceedings of 2nd international conference on advanced
computing and software engineering (ICACSE) (2019)
12. Kumari, CU., Jeevan P.S., Mounika, G.: Leaf disease detection: feature extraction with K-
means clustering and classification with ANN. In: 2019 3rd international conference on
computing methodologies and communication (ICCMC). IEEE (2019)
13. Qiang, Z., He, L., Dai, F.: Identification of Plant Leaf Diseases Based on Inception V3 Transfer
Learning and Fine-Tuning. In: Wang, G., El Saddik, A., Lai, X., Martinez Perez, G., Choo,
K.-K. (eds.) iSCI 2019. CCIS, vol. 1122, pp. 118–127. Springer, Singapore (2019). https://
doi.org/10.1007/978-981-15-1301-5_10
14. Serawork, W., Polceanu, M., Buche, C.: Soybean plant disease identification using convolu-
tional neural network. In: FLAIRS conference (2018)
Prediction of Tomato Leaf Disease Plying Transfer Learning Models 305
15. Hasan, M.M.M., et al.: An efficient disease detection technique of rice leaf using AlexNet. J.
Comput. Commun. 8(12), 49 (2020)
16. Gunjan, C., et al.: Potato leaf disease detection using inception V3. Int. Res. J. Eng. Technol
(IRJET) 7(11), 1363–1366 (2020)
17. Mitra, A., Mohanty, S.P., Kougianos, E.: aGROdet: a Novel framework for plant disease
detection and leaf damage estimation. In: Proceedings of the IFIP International Internet of
Things Conference (IFIP-IoT), pp. 3–22 (2022)
18. Mitra, A., Mohanty, S.P., Kougianos, E.: A smart agriculture framework to automatically
track the spread of plant diseases using mask region-based convolutional neural network. In:
Proceedings of the IFIP International Internet of Things Conference (IFIP-IoT), pp. 68–85
(2022)
Video Captioning Based on Sign Language
Using YOLOV8 Model
1 Introduction
Facial expressions, hand gestures, and body movements are among the visual cues uti-
lized in sign language to convey meaning. Sign language is very useful for people who
have difficulty hearing or speaking. The process of converting these gestures into words
or the rudiments of formally spoken languages is known as recognition of sign language.
Sign language, unlike spoken language, is based on concepts. There are 26 hand sym-
bols used to symbolize each letter of the alphabet in this language, although it cannot be
written. These gestures are expressed using the fingers, and words are written out. Words
or names are communicated by combining finger spellings and motions. The deaf and
dumb use two widely used sign languages that have developed over time: American Sign
Language (ASL) and British Sign Language (BSL). A technique called full communi-
cation, which incorporates spoken language, lip reading, BSL or ASL, and both, is used
in several deaf schools. Hand gestures are a powerful tool for human communication.
There are many signs that represent complicated meanings; Therefore, it can be difficult
for persons who don’t comprehend that language to identify them. Given its increasing
popularity, gesture-based communication has been the subject of numerous studies. In
the past, sensor-based technology was primarily used for sign language interpretation.
Gloves with sensors that link to a recipient on one side are used in this method. However,
this strategy has flaws of its own. Convolutional neural networks, deep learning, and AI
have all improved signing interpretation communication in this method, to name a few
instances. Regional categories for sign language include Indian, American, Chinese,
Arabic, and so forth. To improve applications and explain them at the most basic levels,
supposedly countries also conduct research on gesture recognition, pattern recognition,
and image processing.
2 Literature Survey
The article Real-Time Voice to Sign Language Automatic Translation for the Deaf and
Dumb People view the entire approach as a pipeline in which YouTube recordings or
videos are delivered as the source and smooth films with sign language are generated
and transferred to the appropriate voice [1]. The pipeline was broken up into several
smaller sections. The first module uses NLP to process the input video (from YouTube
or any other.mp4 file) and captions, extracting the appropriate keyword from the text.
The database receives this keyword as 3D symbol signals. Additionally, sign language
is represented in the final video. This paper’s shortcoming is that it only uses a small-
scale framework and the limited gesture-based communication that is available to them.
The framework is otherwise adaptable and straightforward to employ in a study hall
environment. It features a particular module that makes it simple for the teachers to
communicate with the students during exams.
The image frames are first acquired from the video captured via OpenCV. Then
the region and the hand gestures are segmented and detected. Finally, the CNN model
developed determines the label and converts it to text [2].
The workflow of the model is divided into many phases in the study Real Time
Sign Language Recognition and Translation to Text for Vocally and Hearing-Impaired
People [3]. When a user uploads a video, it is converted into adequate frames in the first
step, which is also where the video procurement takes place. The following stage will
preprocess the acquired frames or photos after receiving these frames. In the following
step, the frames are preprocessed to get rid of the noise and blur from the video that
was shot in poor lighting or under unfavorable circumstances. By using the surrounding
entry to replace each one, the median filtering removes the noise and blurriness. The
308 B. S. Vidhyasagar et al.
feature is then derived using the HOG picture processing technique in the following
stage. It extracts the feature from the image and outputs vectors, allowing the classifica-
tion of items using this shape. After features are derived, they are given into the SVM
classification, which uses the feature to put them in groups. Utilizing prepared video
and group testing video with a specified representation, the characterization supplies
the outcome. At the conclusion of the classification procedure, the proportional content
portrayal will be provided using the class names that were assigned during the training
phase. The sole drawback of this paper is that the algorithm and the picture both play
a role in how precisely differentiation is accomplished using this method. Otherwise, it
achieves the goal of higher precision and lower processing overhead and has a higher
recovery accuracy compared to conventional processing systems.
Sign Language Recognition has been separated into two levels in this research pro-
cessing level and classification [4]. Real-time images are initially sent to the processing
stage, where they are preprocessed by turning from color to grayscale to improve classifi-
cation. The image is preprocessed by adding noise as needed and eliminating undesired
noise with a median filter. The OTSU algorithm is used to extract the features from
the image, and the image is then classified into the appropriate sign by comparing it
to database images created using the augmentation technique, where each sign will be
trained from various angles to produce a classification that is more accurate and pre-
cise. SVM (Support Vector Machine), the most well-liked technique for sign language
recognition, is the classification approach applied in this case. Also, MATLAB is used
to extract the relevant text and voice when the features obtained match the features in
the database.
The author insists on a system which involves three equipment: Flex sensors for the
hand (5 fingers), Arduino UNO and LCD for the display [5]. All the flex sensors help
in collecting real-time data collected from the position of the fingers and the obtained
data is sent to Arduino UNO for processing. After the data is processed, the message is
displayed on the LCD. Commonly used sentences and phrases are represented by Mode
1, which is triggered by the combination 01000. When switching on Mode-1, the desired
message can be selected from the available messages in the database/table. A particular
message is associated with each bend of the five flex sensors on the five fingers. For
instance, straightening the little finger while bending the other fingers sends the word
“Good Morning.“ Another example is saying “I’m sorry” by straightening the thumb
and little finger while bending the other three fingers. Alphabets are represented by
Mode-2, which is enabled by the sequence 01100. The desired alphabets can be selected
from the English alphabets and few special characters once Mode-2 has been activated.
With Mode 2, the user can represent words in addition to the alphabet, removing the
limitation found in Mode 1. There were only enough flex sensors to define a maximum
of 29 widely used phrases. This mode offered the ability to construct new phrases using
different alphabets. Also, it should be noted that only 29 of the possible 32 possibilities
can be used for communication.
Video Captioning Based on Sign Language Using YOLOV8 Model 309
The study Modelling of Sign Language Smart Glove Based on Bit Equivalent Imple-
mentation Using Flex Sensor focuses on the designing a low-cost wearable for the
speech-impaired which identifies the sign-language made by the user using Machine
learning and produces audio clip for the sign-language interpreted [6]. This system
involves three modules: the Gesture sensing module, the data acquisition module and
the sign recognition and presentation module. To sense both the orientation and the flex-
ion of fingers, an economical sensing glove that does not disturb the normal functioning
of the hands was developed in Gesture sensing module. A bend sensor based on Velostat
was created to sense flexion while a three-axis MEMS accelerometer was employed to
detect hand orientation. Each bend sensor was connected in series with a 5k resistor to
generate a voltage divider. A 5V DC supply was given to this arrangement and voltage
Vbend is measured as a function of Resistance Rbend. Each glove contains eight analogue
output lines, of which five are used to transmit bend sensor data and three are used to
provide accelerometer data to the data acquisition module. The analogue signals are
transformed into digital for further processing by the data acquisition module. The A/D
conversion procedure has been carried out using the National Instruments portable data
acquisition systems. Using trial and error, it was chosen to set the sampling rate at 50
samples per second to reduce the processing time. The Recognition and Presentation
module processes the sensor outputs and identifies the sign being made using the gloves.
There are two parts in this module: quantization of sign recognition series and sign
recognition and classification. Inside the Quantization part, once the sampled data from
the sensors is available, the Symbolic Aggregate ApproXimation (SAX) algorithm is
used to code it into symbols. In the Sign Classification part, the ability of three different
classifiers—Naive Bayes, Classification Tree, and Support Vector Machines (SVM)—
to recognise and differentiate between ISL gestures was tested. SVM gave the highest
accuracy of 93.04% and the only limitation of this system is it is only limited to 56 ISL
signs.
3 Proposed Methodology
The American Sign Language Letters dataset is an object detection dataset that
includes a bounding box for each ASL letter. This project has a collection of 720 pho-
tographs take with different hand postures held at different places. As this is a tiny
dataset, manual labelling with bounding box coordinates was performed using the label-
ing software in Roboflow, and a probability of transformations function was used to gen-
erate many instances of the same picture, each with different bounding boxes. Figure 1
Represents the sample from the dataset with identified sign.
Once this dataset is ready, the next step would be to train a deep learning model on
the data. The model would need to be optimized for real-time performance to ensure
fast processing of video input. To ensure accuracy, the model is trained with a large and
diverse set of augmented data and regularly evaluated on test data. The evaluation should
measure the model’s performance on recognizing signs in different lighting conditions,
orientations, and hand shapes.
3.2 Develop a Wrapper that can get Real-Time Video from Online Meetups
Developing a wrapper for obtaining real-time video from online meetups involves creat-
ing a software module that can interface with the video streaming capabilities of popular
online meeting platforms, such as Zoom, Microsoft Teams, Google Meet, or other similar
platforms. The wrapper would be responsible for capturing and processing video streams
from online meetups in real-time, allowing developers to build additional functionalities
on top of the video streams.
The first step in developing the wrapper would be to understand the APIs (Application
Programming Interfaces) provided by the online meeting platforms. These APIs typically
provide access to video streams, audio streams, and other data related to online meetings.
The wrapper would need to utilize these APIs to capture the video stream from the online
meetup platform.
Video Captioning Based on Sign Language Using YOLOV8 Model 311
Next, the wrapper would need to process the video stream in real-time. This could
involve tasks such as decoding the video stream, extracting frames, and performing
image processing operations like resizing, cropping, and color correction. Real-time
processing would be crucial to ensure minimal latency and smooth performance.
Once the video stream is processed, the wrapper could provide various functionalities
depending on the intended use case. For example, it could perform real-time video
analysis, such as face recognition, emotion detection, or object detection, to extract
meaningful information from the video stream. It could also overlay additional visual
elements, such as annotations, graphics, or subtitles, on top of the video stream to enhance
the user experience.
3.3 Recognize the Hand Motions from the Acquired Image Frame
Yolo-v8 is one of the currently leading models in this industry. YOLO can identify
hand gestures in image frames because it uses deep neural networks to locate objects
in real-time. Image acquisition is followed by pre-processing to assure image quality,
object recognition with YOLO, hand gesture classification using further machine learn-
ing approaches, post-processing to fine-tune results, and output of the recognised hand
gestures. The quality of the training data, the model architecture, and the hyperparameters
all have an impact on how accurate the hand motion detection system is. The reliability
and sturdiness of the hand motion recognition system must be ensured through careful
examination and testing. Also, to extract the exact sign from the frame or the input video,
it removes the background and precisely captured the hand from the camera input during
real time sign language detection.
Even though Yolo-v8 was not produced by the original YOLO writers, YOLO v8
is believed to be quicker and lighter, with accuracy comparable to YOLO v5, which is
widely regarded as the quickest and most accurate real-time object recognition model.
85% of the enhanced photos were used for training, with the remaining 15% set aside
for testing and validation. The model was trained for 64 epochs with the YOLOv8
pre-trained weights using transfer learning.
Images of sign language can be converted into text using the YOLO object identification
technique and the spellchecker library. The spellchecker library is used to process the
labels produced by YOLO, and it can detect and rectify any potential spelling mistakes in
the text. As a result, sign language gestures can be converted into written text and the mes-
sage expressed by the sign language image can be represented textually. The algorithms
and dictionaries used in the spellchecker library assist in recommending adjustments
depending on context, ensuring that the final literary material is grammatically and lin-
guistically accurate. It is easier for a larger audience to access and comprehend because
to this combined use of YOLO and the spellchecker library.
312 B. S. Vidhyasagar et al.
4 Results
4.1 Mean Average Precision
Mean Average Precision (mAP) at 50 is a commonly used evaluation metric for object
detection models, including those based on YOLO (You Only Look Once) algorithm.
It measures the accuracy of the model in detecting objects across different categories
(e.g., person, car, bicycle) at a certain intersection over union (IoU) threshold of 0.5. The
[email protected] score of the created model is 95.7%. Figure 2 represents the predicted sample
test results.
F1-Confidence curve, the highlighted one showing the F1-Confidence calculated for 64
epochs. The F1 score is a measure of a model’s accuracy that considers both precision
and recall. It is defined as the harmonic mean of precision and recall:
F1 score = 2 * (precision * recall) / (precision + recall).
The fact that the model performs admirably with such a tiny dataset cannot be overlooked!
Even in fresh locations with different hands, it detects well. There are a couple constraints
that can be easily overcome by just providing more data to train on. With a few tweaks
and a lot more data, we anticipate having a functional model that can be extended far
beyond the ASL alphabet.
314 B. S. Vidhyasagar et al.
5 Conclusion
Video captioning based on sign language is a crucial solution for improving accessibility
for individuals who are deaf or hard of hearing, providing them with equal access to
information, education, and multimedia content. With accurate and synchronized cap-
tions that convey sign language expressions, video content becomes more inclusive and
promotes effective communication, bridging gaps between sign language users and non-
sign language users in various settings. Continued advancement and promotion of sign
language video captioning can create a more inclusive world, fostering social inclusion
and enhancing communication for individuals with hearing loss.
References
1. Mehta, A., Solanki, K., Trupti Rathod, T.: Automatic translate real-time voice to sign language
conversion for deaf and dumb people. Int. J. Eng. Res. Technol. (IJERT) ICRADL – 2021
9(5), 174–177 (2021)
2. Rani, R.S., Rumana, R., Prema, R.: A review paper on sign language recognition for the deaf
and dumb. Int. J. Eng. Res. Technol. (IJERT) 10(10), (2021)
3. Khallikkunaisa, Kulsoom A.A., Chandan Y.P, Fathima Farheen, F., Halima, N.: Real time
sign language recognition and translation to text for vocally and hearing impaired people. Int.
J. Eng. Res. Technol. (IJERT) IETE 8(11) (2020)
4. Kodandaram, S.R., Kumar, N., Gl, S.: Sign language recognition. Turkish J. Comput. Math.
Educ. (TURCOMAT). 12, 994–1009 (2021)
5. Muralidharan, N. T. , R. R. S., R. M. R., S. N. M., H. M. E.: Modelling of sign language
smart glove based on bit equivalent implementation using flex sensor. In: 2022 International
Conference on Wireless Communications Signal Processing and Networking (WiSPNET),
Chennai, India, pp. 99–104 (2022). https://doi.org/10.1109/WiSPNET54241.2022.9767137
6. Singh, A.K., John, B.P., Subramanian, S.R.V., Kumar, A.S., Nair, B.B.: A low-cost wearable
Indian sign language interpretation system. In: 2016 International Conference on Robotics
and Automation for Humanitarian Applications (RAHA), Amritapuri, India, , pp. 1–6,
(2016) https://doi.org/10.1109/RAHA.2016.7931873
Video Captioning Based on Sign Language Using YOLOV8 Model 315
7. Kartik, P.V.S.M.S., Sumanth, K.B.V.N.S., Ram, V.N.V.S., Prakash, P.: Sign language to text
conversion using deep learning. In: Ranganathan, G., Chen, J., Rocha, Á. (eds) Inventive
Communication and Computational Technologies. Lecture Notes in Networks and Systems,
vol 145. Springer, Singapore (2021). https://doi.org/10.1007/978-981-15-7345-3_18
8. Islam, S., Dash, A., Seum, A., et al.: Exploring video captioning techniques: a comprehensive
survey on deep learning methods. SN COMPUT. SCI. 2, 120 (2021)
9. Tamilselvan, K.S., Balakumar, P., Rajalakshmi, B., Roshini, C., Suthagar, S.: Translation of
sign language for deaf and dumb people. Int. J. Recent Technol. Eng. 8, 2277–3878 (2020).
https://doi.org/10.35940/ijrte.E6555.018520
10. Juju, R.: Video captioning and sign language interpreting (2022)
11. earn2Sign: Sign Language Recognition and Translation using Human Keypoint Estimation
and Transformer Model - September 09, 2020 - Institute for Natural Language Processing
University of Stuttgart Pfaffenwaldring 5 bD-70569 Stuttgart
12. Vidhyasagar, B.S., Raja, J., Marudhamuthu, Krishnamurthy.: A novel oppositional chaotic
flower pollination optimization algorithm for automatic tuning of hadoop configuration
parameters. Big Data. 8, 218–234 (2020). https://doi.org/10.1089/big.2019.0111
13. Cihan Camgöz, N., Koller, O., Hadfield, S., Bowden, R.: Sign language transformers: joint
end-to-end sign language recognition and translation. In: 2020 IEEE/CVF Conference on
Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, pp. 10020–10030
(2020), https://doi.org/10.1109/CVPR42600.2020.01004
14. Bantupalli, K., Xie, Y.: American sign language recognition using deep learning and computer
vision. In: 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA
(2018)
15. Suri, K., Gupta, R.: Convolutional neural network array for sign language recognition using
wearable signal processing and integrated networks. In: 2019 6th International Conference
on Signal Processing and Integrated Networks (SPIN), Noida, India (2019)
16. Byeongkeun, K., Tripathi, S., Nguyen, T.Q.: Real- time sign language fingerspelling recog-
nition using convolutional neural networks from depth map. arXiv preprint https://arxiv.org/
abs/1509.03001 (2015)
17. Anand, M.S., Kumar, N.M., Kumaresan, A.: An efficient framework for indian sign language
recognition using wavelet transform. Circuits Syst. 7, 1874- 1883 (2016)
18. Kumud, T., Baranwal, N., Nandi, G.C.: Continuous dynamic indian sign language gesture
recognition with invariant backgrounds. In: 2015 International Conference on Advances in
Computing, Communications and Informatics (ICACCI), 2211–2216 (2015)
19. Manikandan, A.B.R.K.: AAyush patidar, and pallav walia, hand gesture detection and
conversion to speech and text. Int J. Pure Appl Math. 120(6), 1347–1362 (2018)
20. Dutta, K.K., Swamy, S.A.: Double Handed Indian Sign Language to Speech and Text. In:
2015 Third International Conference on Image Information Processing (ICIIP), pp. 374–377
(2017)
Improvement in Multi-resident Activity
Recognition System in a Smart Home
Using Activity Clustering
1 Introduction
Human Activity Recognition (HAR) is a system that is built to monitor human
activities, especially the elderly person living alone in the house. The HAR inte-
grates ambient assisted living (AAL) into smart environments such as smart
homes to recognize the activities automatically [1]. The AAL uses simple perva-
sive, ubiquitous sensors with artificial intelligence (AI) to connect smart devices
with Internet-of-Things (IoT). The sensors used in the AAL are categorized into
vision sensors, ambient sensors, wearable sensors/devices, and smartphones for
monitoring the simple, complex, and postural transitions [2]. The smart home
environment with AAL facility has been primarily used for activity monitoring,
healthcare, fall detection, sports tracking, etc. [3]. The vision-based has specific
issues with the person’s privacy while monitoring and requires installation in
all the actuated areas. Wearable and smartphone devices must be worn for the
entire day, which is difficult for the elders. Sometimes, the elders may forget to
carry or recharge the battery due to age or other diseases [4]. Recently, the HAR
with ambient sensors has attracted most researchers as it is non-invasive and
can indirectly alert the family members or caretakers in case of any anomaly in
the residents’ activity [5].
Various researchers have proposed a HAR system in the last decade that
primarily focuses on single resident activity recognition [6]. The single-resident
activity is simple and easy to detect. Many state-of-the-art machine and deep
learning algorithms have shown promising results for the single-resident activity
in the AAL environment [1,3,6]. However, the smart home environment may
have more chances for multi-residents such as friends, neighbors, pets, etc., and
the single-resident is not always the case. Multi-resident activity recognition in an
ambient sensor environment is slightly complex and needs certain improvements
in activity recognition and detection. Researchers have developed multi-resident
activity recognition benchmark datasets for further research and implementa-
tion. However, these datasets have certain limitations in data collection, anno-
tation scarcity, data association, computational cost, etc. This paper solves such
a challenge of the benchmark dataset by clustering the activity labels based on
the nature of the activity. Experimentation has been carried out by utilizing the
ARAS [7] - an ambient sensor-based multi-resident activity recognition dataset.
The remaining sections of the paper are organized as follows. Section 2 deals
with the motivation of the challenges discovered in the benchmark datasets.
Section 3 discusses the proposed clustering of activity labels, Sect. 4 demonstrates
the experimentations and results using the clustered dataset, and Sect. 5 con-
cludes the paper.
2 Motivation
In the last decade, many machine and deep learning techniques have been
employed for multi-resident human activity recognition (MRHAR) in an ambi-
ent sensor smart environment [8,9]. However, MRHAR has certain limitations
that need to be addressed for better improvement in the system performance.
The limitations are discussed in subsections as follows.
ARAS Dataset. [7] - Two pairs of residents in two smart Houses, A and B,
perform 27 activities. House A contains two males, while House B is a married
couple. The dataset reflects on the natural behaviors of the residents to 30 days
of recording for each house. Each house contains 30 days of recording, and each
day consists of 22 * 86,400 instance matrix results in 30 * 22 * 86400 instances.
The first 20 columns are the binary sensor values that say the state of the sensor,
either fired or not. Columns 21 and 22 represent the activity label of residents 1
and 2. Annotation of the activities was achieved by the residents using a simple
graphical user interface (GUI) placed in the most convenient places of the houses.
CASAS Dataset. [10]- The Centre for Advanced Studies in Adaptive Systems
has collected several scripted and unscripted multi-resident activity recognition
datasets. “twor.2009”, “twor.summer.
2009”, “twor.2010”, “Tulum,” “tulum2”, and “Cairo” are unscripted, and “Mul-
tiresidentADLs” is a scripted dataset of CASAS. Activities are annotated by
recording the start and end time of the activity via a handwritten diary. The
data are delimited in a specific format for recording, and it requires preprocess-
ing before the feature engineering process. Data collection uses 14–21 ambient
sensors to carry out 11 activities.
VanKasteren Benchmark. [11] - One resident in three smart houses was consid-
ered for the data collection. It has 14–23 different sensors to collect data and
demonstrate approximately 10–16 activities. The dataset is available for 25 days
but not collected continuously. Activities are annotated using the handwritten
diary and a Bluetooth device.
SDHAR-Home Dataset. [13] - The dataset has been developed through non-
intrusive technology in the multi-resident smart home environment. Set of non-
intrusive sensors integrated with activity wristbands to capture the events in
the house, positioning the user through triangulation using beacons respectively.
Improvement in Multi-resident Activity Recognition System 319
As per the research survey by [8], collecting two ambient sensing smart home
datasets with multi-residents offers a better opportunity to study and compare
the activity recognition algorithm more realistically. The vanKasteren dataset
may not opt for such a fair comparison as it has only one resident perform-
ing various activities in all three smart homes. Similarly, the MARBLE dataset
deals with the multi-habitant nature of the single-resident activity and data col-
lected in the Inlab environment. CASAS data are mostly predetermined and
were repeatedly performed in a controlled laboratory environment. This collec-
tion accounts for inter-subject variability and is thus not suitable for real-world
data analysis. Further, the UJA dataset has been collected using multi-model
heterogeneous hardware items, and this may be infeasible in multiple cases for
the recognition of ADLs. However, the ARAS dataset accounts for intra-subject
variability and does not account for inter-subject one. This leads to a better
opportunity to study and compare the MRHAR system. UJA, MARBLE, and
SDHAR are the recently evolved datasets. They have been created using multi-
modal heterogeneous devices for data collection, so it has not been considered
here for experimentation.
320 E. Ramanujam et al.
It has also been proven by collecting the citations of datasets [7,10–14] from
the year of dataset generation (mostly 2011) to till date (April 30, 2023) as per
the Scopus and Google scholar records. Alternatively, CASAS has the highest
number of citations, 642. VanKasteren has 250 citations, and ARAS has 146
citations. ARAS has fewer citations because the collected data is completely
binary and has no discriminating features to easily classify the activities labeled
in each instance. The performance of the ARAS dataset still needs to improve
due to certain factors of data collected during experimentation. It motivates us to
analyze the ARAS dataset to improve its performance apart from implementing
hybrid deep learning models to make the system more complex to recognize the
activities.
The complexity of data association is the second major challenge in the HAR
system, which refers to the number of persons and activities associated with
it. Concurrent activity is one of the challenges in activity recognition driven by
data association. The activity occurs when the resident participates in multiple
activities simultaneously-for instance, sitting and watching TV, conversing while
watching TV, etc. A sample challenge with data association is shown in Fig. 1,
which represents the activity of watching TV and using the Internet in four
different instances. In the first instance, there is no firing of the sensor. In the
second instance, the IR sensor (sensor 3) of the TV receiver is fired, which
confirms the activity of watching TV. However, there is no sensor firing for the
activity using the Internet. In this case, the resident may be standing and using
the Internet. In the third instance, the force sensor of the couch (sensor 4) is
fired, representing that the resident(s) may be sitting on the couch and watching
TV or using the Internet. However, there is no proper sensor firing stating the
activity of watching TV and using the Internet. Finally, in the last instance, the
force sensor in the couch and chair are fired, representing one resident sitting on
the couch and another sitting on the chair. However, the activities are labeled
to be again watching TV and using the Internet. This label associated with the
binary sensor makes the system complex to recognize the activities.
A similar example is shown in Fig. 2 where the activities labeled are preparing
dinner and watching TV. However, it is unclear which resident is preparing the
dinner and which is watching TV. Moreover, for watching TV, sensor 4 is fired
correctly in all three cases, whereas for preparing dinner (sensor 16), it is not
adequately fired in Instances 1 and 4. This makes inevitable proof for improper
labeling and association, which makes any machine/deep learning models more
complex and consumes time for model generation.
In a specific case, multi-resident cooperation activities are also associated
with the complexity of data association. For example, the residents may be in
the process of cooking. However, the respective sensor fired relates to cooking
and washing dishes. So one resident may cook, and the other may wash dishes.
This causes inevitable confusion. Moreover, the system cannot identify who is
Improvement in Multi-resident Activity Recognition System 321
washing the dish and cooking. These limitations must be addressed to improve
the performance of any learning model to handle the ARAS dataset.
Fig. 3. Percentage of timestamp without any sensor activations for ARAS House A
and House B [20]
a shower also. However, the resident may need to annotate the activity correctly,
which uses the toileting label even for the shower with the respective sensor fired.
The third major factor that needs to be concerned is the feasibility of implement-
ing the MRHAR in real-time. A considerable effort must be made to make the
system acceptable to many end-users as it is close to human life. However, the
state-of-the-art systems proposed for MRHAR on the ARAS dataset have certain
challenges in attaining a preferable recognition rate (Accuracy) or the computa-
tional time taken to recognize the event. The research works that considered the
ARAS dataset for experimentation, and the results are shown in Table 2. In com-
paring the performances of the state-of-the-art methods, the accuracy reaches
the maximum of 89% for House A and 97% for House B, respectively. However,
there is no mention of the execution time for the models represented by various
state-of-the-art methods. Alternatively, the research work [19] has attained the
accuracy of 86.315% and 87.975% for Houses A and B, respectively, but the
execution time is 10-fold high than our previous work [20]. The system should
be recourse intensive, so it fits portable devices and can respond instantly. Thus,
the computational cost and recognition issue should be addressed.
– Preparing food is the common primary activity, whereas its secondary activ-
ities are washing dishes, cutting vegetables, etc., In addition, the food can be
further categorized into breakfast, lunch, and dinner. Thus, the primary and
secondary activities are clubbed into one food preparation activity.
– The researchers categorize Eating/having food at all times, namely break-
fast, lunch, and dinner, or eating snacks in the evening (secondary activities)
based on the timestamp where they recorded the instance. All these activities
have been carried out at the dining table kept in the house. Considering this
in-depth, the resident may sit in a chair or at the dining table for all eating
purposes. There may be fewer chances of firing different sensors for the activ-
ity of eating. Thus, the activities are merged into Eating food as a common
term.
– Sleeping and napping are carried out on the couch that fires the same binary
sensor installed in the house. Thus, the activity has been merged into sleeping,
which represents snapping also.
– Watching TV, conversing, and having guests resembles firing the binary sen-
sor of the sofa/couch in the hall where the guest usually arrives. In addition,
the cleaning of the house activity closely resembles the resident sitting on the
sofa/couch. Thus, watching TV, conversing with guests, and cleaning clubbed
into the activity of watching/cleaning.
– Studying and Reading books are carried out by sitting in a chair, and thus
respective binary sensors are fired for this activity. Studying is commonly
represented here for studying and reading books.
– Having a shower, Toileting, Laundry, shaving, and brushing teeth are the
activities that fire the sensors, bathroom cabinet, bathroom door, water
closet, etc. As all these activities are secondary activities carried out in the
bathroom/washroom, it has been managed in the bathroom/washroom in a
single term.
– Using the Internet, talking on the phone, and listening to music represent the
entertainment process, and most fire the couch/chair in the hall. Thus, it has
been fixed into the activity termed entertainment.
326
E. Ramanujam et al.
Fig. 4. The resemblance of sensor firing for various activity labels on the ARAS House B dataset
Improvement in Multi-resident Activity Recognition System 327
– Finally, changing clothes, going out (outdoors), and other activities have dis-
criminate binary pulse sensors that never mix with any other activities. Thus
the same name has been used for the process.
All 27 activities of the ARAS dataset have been clustered into 10 modified
clustered label annotations, and the detailed labeling is shown in Table 3. The
clustering process is completely manual based on the primary and secondary
activities of the dataset.
The experimental results on clustered activity labels with various deep learn-
ing architectures, as mentioned above, are shown in Tables 5 and 6 for Houses
Improvement in Multi-resident Activity Recognition System 331
Fig. 6. Accuracy rate with the number of epochs for ARAS Houses A and B respectively
Similarly, the learning rate has been analyzed for three values (lr = 0.0001,
0.001, and 0.1), as shown in Fig. 7. Among the three values, the deep learning
332 E. Ramanujam et al.
models have shown promising results on lr = 0.001, and thus, it has been fixed
for experimentation as reported in Tables 6 and 4.
4.1 Limitations
5 Conclusion
This paper deals with the limitations of the multi-resident human activity recog-
nition (MRHAR) data collected on ambient sensing smart home environments.
The limitations have evolved during the data collection to data association, anno-
tations, and cost of computation. These limitations are profoundly analyzed with
the benchmark MRHAR datasets such as ARAS, CASAS, vanKasteren, etc. The
clustering of activity labels has been made in this work to analyze the perfor-
mance variation using simple deep-learning models regarding recognition rate
and computation time. The results have also proven that the clustering of activ-
ity labels has a multi-fold increase in computational time efficiency and recogni-
tion. In the near future, the clustering can be done with certain algorithms such
as Expectation-Minimization, Canopy, etc., with the deep convolution models
to outperform state-of-the-art works.
References
1. Bianchi, V., Bassoli, M., Lombardo, G., Fornacciari, P., Mordonini, M., De Munari,
I.: IoT wearable sensor and deep learning: an integrated approach for personalized
human activity recognition in a smart home environment. IEEE Internet Things
J. 6(5), 8553–8562 (2019). https://doi.org/10.1109/jiot.2019.2920283
2. Sharma, V., Gupta, M., Pandey, A.K., Mishra, D., Kumar, A.: A review of
deep learning-based human activity recognition on benchmark video datasets.
Appl. Artif. Intell. 36(1), 2093705 (2022). https://doi.org/10.1080/08839514.2022.
2093705
3. Almeida, A., Mulero, R., Rametta, P., Urošević, V., Andrić, M., Patrono, L.: A
critical analysis of an IoT-aware AAL system for elderly monitoring. Futur. Gener.
Comput. Syst. 97, 598–619 (2019). https://doi.org/10.1016/j.future.2019.03.019
Improvement in Multi-resident Activity Recognition System 333
4. Gupta, N., Gupta, S.K., Pathak, R.K., Jain, V., Rashidi, P., Suri, J.S.: Human
activity recognition in artificial intelligence framework: a narrative review. Artif.
Intell. Rev. 55(6), 4755–4808 (2022)
5. Anikwe, C.V., et al.: Mobile and wearable sensors for data-driven health monitor-
ing system: state-of-the-art and future prospect. Expert Syst. Appl. 202, 117362
(2022). https://doi.org/10.1016/j.eswa.2022.117362
6. Babangida, L., Perumal, T., Mustapha, N., Yaakob, R.: Internet of things (IoT)
based activity recognition strategies in smart homes: a review. IEEE Sens. J. 22(9),
8327–8336 (2022). https://doi.org/10.1109/jsen.2022.3161797
7. Alemdar, H., Durmaz Incel, O., Ertan, H., Ersoy, C.: ARAS human activity
datasets in multiple homes with multiple residents. In: Proceedings of the ICTs for
Improving Patients Rehabilitation Research Techniques (2013). https://doi.org/
10.4108/icst.pervasivehealth.2013.252120
8. Chen, K., Zhang, D., Yao, L., Guo, B., Yu, Z., Liu, Y.: Deep learning for
sensor-based human activity recognition. ACM Comput. Surv. 54(4), 1–40 (2021).
https://doi.org/10.1145/3447744
9. Shiri, F.M., Perumal, T., Mustapha, N., Mohamed, R., Ahmadon, M.A.B., Yam-
aguchi, S.: A survey on multi-resident activity recognition in smart environments.
arXiv preprint: arXiv:2304.12304 (2023)
10. Cook, D.J., Crandall, A.S., Thomas, B.L., Krishnan, N.C.: CASAS: a smart home
in a box. Computer 46(7), 62–69 (2013). https://doi.org/10.1109/mc.2012.328
11. van Kasteren, T.L.M., Englebienne, G., Kröse, B.J.A.: Human activity recogni-
tion from wireless sensor network data: benchmark and software. In: Chen, L.,
Nugent, C., Biswas, J., Hoey, J. (eds.) Activity Recognition in Pervasive Intelli-
gent Environments. Atlantis Ambient and Pervasive Intelligence, vol. 4. Atlantis
Press, Amsterdam (2011). https://doi.org/10.2991/978-94-91216-05-3 8
12. De-La-Hoz-Franco, E., Bernal Monroy, E., Ariza-Colpas, P., Mendoza-Palechor, F.,
Espinilla, M.: UJA human activity recognition multi-occupancy dataset. In: Pro-
ceedings of the 54th Hawaii International Conference on System Sciences (2021).
https://doi.org/10.24251/hicss.2021.236
13. Ramos, R.G., Domingo, J.D., Zalama, E., Gómez-Garcı́a-Bermejo, J., López, J.:
SDHAR-HOME: a sensor dataset for human activity recognition at home. Sensors
22(21), 8109 (2022). https://doi.org/10.3390/s22218109
14. Arrotta, L., Bettini, C., Civitarese, G.: The MARBLE dataset: multi-inhabitant
activities of daily living combining wearable and environmental sensors data. In:
Hara, T., Yamaguchi, H. (eds.) MobiQuitous 2021. Lecture Notes of the Institute
for Computer Sciences, Social Informatics and Telecommunications Engineering,
vol. 419, pp. 451–468. Springer, Cham (2022). https://doi.org/10.1007/978-3-030-
94822-1 25
15. Chen, D., Yongchareon, S., Lai, E.M.-K., Yu, J., Sheng, Q.Z., Li, Y.: Transformer
with bidirectional GRU for nonintrusive, sensor-based activity recognition in a
multiresident environment. IEEE Internet Things J. 9(23), 23716–23727 (2022).
https://doi.org/10.1109/jiot.2022.3190307
16. Lentzas, A., Dalagdi, E., Vrakas, D.: Multilabel classification methods for human
activity recognition: a comparison of algorithms. Sensors 22(6), 2353 (2022).
https://doi.org/10.3390/s22062353
17. Natani, A., Sharma, A., Peruma, T., Sukhavasi, S.: Deep learning for multi-resident
activity recognition in ambient sensing smart homes. In: 2019 IEEE 8th Global
Conference on Consumer Electronics (GCCE) (2019). https://doi.org/10.1109/
gcce46687.2019.9015212
334 E. Ramanujam et al.
18. Jethanandani, M., Sharma, A., Perumal, T., Chang, J.-R.: Multi-label classification
based ensemble learning for human activity recognition in smart home. Internet
Things 12, 100324 (2020). https://doi.org/10.1016/j.iot.2020.100324
19. Natani, A., Sharma, A., Perumal, T.: Sequential neural networks for multi-resident
activity recognition in ambient sensing smart homes. Appl. Intell. 51(8), 6014–6028
(2021). https://doi.org/10.1007/s10489-020-02134-z
20. Ramanujam, E., Perumal, T.: MLMO-HSM: multi-label multi-output hybrid
sequential model for multi-resident smart home activity recognition. J. Ambient.
Intell. Humaniz. Comput. 14(3), 2313–2325 (2023)
21. Perumal, T., Ramanujam, E., Suman, S., Sharma, A., Singhal, H.: Internet of
things centric-based multiactivity recognition in smart home environment. IEEE
Internet Things J. 10(2), 1724–1732 (2023). https://doi.org/10.1109/jiot.2022.
3209970
22. The multilayer perceptron. Neural Computing - An Introduction (1990). https://
doi.org/10.1201/9781420050431.ch4
23. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8),
1735–1780 (1997). https://doi.org/10.1162/neco.1997.9.8.1735
24. Cho, K., et al.: Learning phrase representations using RNN encoder-decoder for
statistical machine translation. In: Proceedings of the 2014 Conference on Empir-
ical Methods in Natural Language Processing (EMNLP) (2014). https://doi.org/
10.3115/v1/d14-1179
25. O’Shea, K., Nash, R.: An introduction to convolutional neural networks. arXiv
preprint: arXiv:1511.08458 (2015)
Metaverse for IoT (MIoT)
Forensics Analysis of Virtual Reality
Social Community Applications
on Oculus Quest 2
1 Introduction
Virtual Reality (VR) is currently used in a wide range of industries such as edu-
cation, healthcare, and social communities. In education, VictoryXR has created
immersive learning experiences by using VR to provide training to educators,
teaching many subjects [26]. They expect to operate up to 100 digital twin cam-
puses by 2023 [20]. In healthcare, VR goes hand-in-hand with Internet of Med-
ical Things (IoMT) and Internet of Health Things (IoHT) based technologies.
In November 2021, the U.S. Food and Drug Administration (FDA) approved
a VR system called EaseVRx as back pain treatment for individuals above 18
years old [24]. For social communities, the metaverse has allowed people to meet
up in virtual spaces such as restaurants, cinemas, and even concerts. Bouck-
aert [11] found that cheaper prices, visual effects, and accessibility are some of
the motivators for people preferring virtual concerts to live concerts. Social VR
platforms such as Meta’s Horizon Worlds and Microsoft’s AltspaceVR enable
people to traverse virtual worlds, play games, and interact with friends as well
as strangers online.
While VR brings many benefits, it can also lead to some drawbacks. For
example, VR can provide an immersive and interactive learning experience that
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 337–352, 2023.
https://doi.org/10.1007/978-3-031-45878-1_23
338 S. Ho and U. Karabiyik
2 Related Work
The following section discusses a review of the literature that was conducted
regarding virtual reality social applications, forensic analysis of virtual reality,
and mobile forensic investigations.
of Detroit case where text messages were used as evidence in a court of law.
Ultimately, this led to perjury and misconduct charges while Mayor Kwame
Kilpatrick and his Chief of Staff Christine Beatty were serving as public officers.
The current study forensically analyzes the Oculus Quest 2 for artifacts such as
message logs which can be used as possible inculpatory evidence.
3 Methodology
The goal of this study was to discover the amount of recoverable forensically
relevant user data from four social community applications on the Oculus Quest
2 through mobile forensic analysis. A virtual reality environment was used to
achieve this. A private Wi-Fi network was used for the devices to have internet
connectivity. DHCP assignment will be used to assign IP addresses to all devices.
All assigned IP addresses follow an RFC 1918 [23] compliant addressing scheme
and are in the 10.0.0.0/8 subnetwork. The following sections describe the scenario
setup, measures, design, virtual environment, and data analysis methods for the
present study.
Forensics Analysis of Virtual Reality Social Community Applications 341
The physical location of the investigation was within a single research lab. A
possible scenario where this investigation would be useful is a case involving
child grooming. Choo [12] defines child grooming as “a premeditated behavior
intended to secure the trust and cooperation of children prior to engaging in
sexual conduct.” Online grooming has become increasingly popular online due
to new technologies such as cyberspace [13,19]. Pedophiles also share information
such as tricks to groom children for abuse with other pedophiles [9]. An offender
would need a profile to communicate with children online. Given the offender
used any of the social community applications investigated, forensically relevant
information such as user height, message logs, and interactions can be used as
both inculpatory and exculpatory evidence.
3.2 Measures
The single-group posttest-only design was employed for this study as the inde-
pendent variable is manipulated with a single group. The independent variable
is the social community application used. The dependent variable is the amount
of recoverable forensically relevant user data from the social community applica-
tions. This study attempts to answer the question of “What forensically relevant
user data is recoverable from social community applications on the Oculus Quest
2?” A potential threat to internal validity is due to the data population, acqui-
sition, and analysis being done in a controlled virtual environment. This erases
any factors and challenges that may influence the study in a real-world situation.
To address this, real-world scenarios were replicated as closely as possible within
the virtual environment. This involves designing virtual scenarios that mimic the
characteristics, interactions, and usage patterns found in real-world VR applica-
tions. By creating realistic conditions, we increase the ecological validity of the
current study.
3.3 Design
Data Population. The social community applications that were used are free
software available in the Oculus Quest Store. A test user account (AnnaOup525)
342 S. Ho and U. Karabiyik
was created and used throughout this study. This account was used to log in
to the Meta Quest application on the Samsung Galaxy Tab S7 as well as the
Oculus Quest 2. After logging in, the four social community applications were
downloaded and installed. A full list of applications and software used can be
seen in Table 1.
The same activities were done on each of the four social community appli-
cations to generate user data on the device. For the purposes of interaction, a
Forensics Analysis of Virtual Reality Social Community Applications 343
second account (**********hpa) was used. All user interactions were made with
the researchers’ second account. A short summary of the activities done on the
device is as follows:
Data Analysis. This study used the following methods to analyze the Oculus
Quest 2.
Forensics Imaging. The first method was to acquire an image of the Oculus
Quest 2. The Oculus Quest 2 was connected to the laptop via a USB-C cable.
AXIOM Acquire was used to create a forensically sound image of the device
and its storage media. As the Oculus Quest 2 was not rooted, there was no
authorized access to the full file system. This resulted in a quick image being
made instead of a logical or full image. Hash values were used to verify the
integrity of the image and ensure that it is an exact copy of the original data.
Once verified, Magnet AXIOM Process was used to load the image into Magnet
AXIOM Examine which was then used to extract data from the image. Foren-
sically relevant files and artifacts, such as application data, system logs, and
user data were examined to check for any recoverable forensically relevant user
data. Examples of forensically relevant user data include personally identifiable
information (PII), login credentials, and interaction history.
Log Analysis. The second method was to connect the Oculus Quest 2 to the
laptop and manually perform log analysis on available files and folders.
4 Results
The following section describes the results obtained from the analysis section.
Results are reported in two parts - general findings and application-specific find-
ings.
As seen in Fig. 2, the latest WiFi scan results were stored in a file called
OculusQuest2QuickImage.zip sdcard sdcard Android logs wifi-
scanner.txt. We can see that the ***3.0 network was scanned. From Fig. 3, we
can see that a successful connection was logged. The connection was logged in with
the researcher’s username (***76) as well as the domain (******.edu) which can
be seen in Fig. 4.
344 S. Ho and U. Karabiyik
The following findings were found from the image created through Magnet
AXIOM. There were several forensically relevant user and application data that
could be found from Horizon Worlds. Figure 5 shows the last used time and dura-
tion used for the Horizon Worlds application. Unlike for Multiverse, Rec Room,
and VRChat, the total time used and last time used were 00:00 and 1969-12-31
19:00:00 respectively. As these are default UNIX timestamps, it can be assumed
that this data was either not available or it is encrypted.
Another artifact that was found was a photo that was taken showing the
interaction of the two user accounts. This can be seen in Fig. 6.
The following findings were found through log analysis on the
Quest 2 Android data com.facebook.horizon Horizon Logs
socialvr 2023-04-24 12.51.01.log log file.
The first artifact was the username and EPOCH time of the user name tag
being created. This can be seen in Fig. 7.
One of the actions taken was sending follow requests. As seen in Fig. 8,
the second account’s username, ID, avatar head shot picture, event type, and
EPOCH time of event were logged.
Additionally, another action is navigating to different rooms. For Horizon
World, both accounts navigated to a room called “Arena Clash Winter.” As
seen in Fig. 9, the name of the room and EPOCH time entered were logged.
4.3 Multiverse
The following findings were found from the image created through Magnet
AXIOM. There were several forensically relevant user and application data that
could be found in Multiverse. Figure 10 shows the last used time and duration
used for the Multiverse application. The total time used and last time used were
32:01 and 2023-04-24 12:48:58 respectively. These timestamps were very accurate
as they matched the logged data population time.
In addition, there were Binary Large Object (BLOB) cache files that had
signature mismatches. Figure 11 and Fig. 12 were recovered using file carving.
These were the profile pictures of the first and second test user accounts in
Multiverse.
No useful forensically relevant user data was found through log analysis.
346 S. Ho and U. Karabiyik
No useful forensically relevant user data was found through log analysis.
4.5 VRChat
The following findings were found from the image created through Magnet
AXIOM. There were several forensically relevant user and application data that
could be found in VRChat. Figure 14 shows the last used time and duration used
for the VRChat application. The total time used and last time used were 17:39
and 2023-04-24 13:31:33 respectively. These timestamps were very accurate as
they matched the logged data population time.
Another artifact was a picture taken using the camera feature in VRChat.
This picture shows the interaction between the two user accounts. This can be
seen in Fig. 15.
No useful forensically relevant user data was found through log analysis.
tested to see if an image of the Oculus Quest 2 could be made. Another limitation
of the current study is the small sample size of social community applications
that were analyzed (n = 4). Future research could expand upon these limitations
to gain a more comprehensive understanding of the forensic artifacts present in
VR applications and the potential risks to user privacy.
via virtual reality social community applications, which can help support the
investigation and prosecution of crimes involving the use of these technologies.
Overall, this study highlights the importance of considering privacy concerns
when using VR applications. Developers need to prioritize user data protection
in their designs. As VR technology continues to advance and become more inte-
grated into our daily lives, it is crucial to ensure that our personal information
remains secure.
References
1. Android debug bridge. https://developer.android.com/studio/command-line/adb
2. Horizon worlds. https://www.meta.com/horizon-worlds/
3. Magnet axiom forensics. https://www.magnetforensics.com/products/magnet-
axiom. Accessed 03 Feb 2023
4. Meta quest. https://play.google.com/store/apps/details?id=com.oculus.twilight&
hl=en US&gl=US&pli=1
5. Meta quest 2 forensic extraction (testing). https://revo4n6.com/blog-posts/f/
meta-quest-2-forensic-extraction-testing. Accessed 18 May 2023
6. Multiverse. https://www.oculus.com/experiences/quest/4140152679403866/
7. Rec room. https://recroom.com/
8. Vrchat. https://www.oculus.com/experiences/quest/1856672347794301/
9. Adam, A.: Cyberstalking and internet pornography: gender and the gaze. Ethics
Inf. Technol. 4(2), 133 (2002)
10. Adamski, D.: The influence of new technologies on the social withdrawal (hikiko-
mori syndrome) among developed communities, including poland. Soc. Commun.
4(1), 58–63 (2018)
11. Bouckaert, L.: Virtual reality as new concert space. Ph.D. thesis, Universiteit Gent,
Belgium (2021)
12. Choo, K.K.R.: Online child grooming: a literature review on the misuse of social
networking sites for grooming children for sexual offences (2009)
13. Choo, K.K.R., Smith, R.G., McCusker, R., Choo, K.K.R.: Future directions in
technology-enabled crime: 2007-09. Citeseer (2007)
14. Hammond, T.: Text messages and the detroit mayor: Kwame kilpatrick. Online J.
Commun. Media Technol. 3(3), 146–177 (2013)
15. Hassenfeldt, C., Jacques, J., Baggili, I.: Exploring the learning efficacy of digital
forensics concepts and bagging & tagging of digital devices in immersive virtual
reality. Forensic Sci. Int. Digit. Invest. 33, 301011 (2020)
16. Hutchinson, S., Stanković, M., Ho, S., Houshmand, S., Karabiyik, U.: Investigating
the privacy and security of the simplisafe security system on android and iOS. J.
Cybersecur. Priv. 3(2), 145–165 (2023)
17. Johnson, H., Volk, K., Serafin, R., Grajeda, C., Baggili, I.: Alt-tech social forensics:
forensic analysis of alternative social networking applications. Forensic Sci. Int.
Digit. Invest. 42, 301406 (2022)
18. Jones, G.M., Winster, S.G.: Forensics analysis on smart phones using mobile foren-
sics tools. Int. J. Comput. Intell. Res. 13(8), 1859–1869 (2017)
19. Kierkegaard, S.: Cybering, online grooming and ageplay. Comput. Law Secur. Rev.
24(1), 41–55 (2008)
20. Kshetri, N., Rojas-Torres, D., Grambo, M.: The metaverse and higher education
institutions. IT Prof. 24(6), 69–73 (2022)
352 S. Ho and U. Karabiyik
21. Lyle, J.: Computer forensic tool testing at NIST (2007). http://www.cftt.nist.gov/
documents/Amalfi-04.ppt
22. Mahan, M.E.: Exploring ransomware on the oculus quest 2. Ph.D. thesis, Louisiana
Tech University (2022)
23. Rekhter, Y., Moskowitz, B., Karrenberg, D., Groot, G.D., Lear, E.: RFC 1918:
address allocation for private internets (1996). https://www.rfc-editor.org/rfc/
rfc1918.html. Accessed 01 Mar 2022
24. Sato, T., Ishimaru, H., Takata, T., Sasaki, H., Shikano, M.: Application of internet
of medical/health things to decentralized clinical trials: development status and
regulatory considerations. Front. Med. 9, 903188 (2022)
25. Statista: Number of internet and social media users worldwide as of January
2023 (in billions). https://www.statista.com/statistics/617136/digital-population-
worldwide/#:∼:text=As%20of%20January%202023%2C%20there,population
%2C%20were%20social%20media%20users. Accessed 17 Apr 2023
26. Wang, M., Yu, H., Bell, Z., Chu, X.: Constructing an edu-metaverse ecosystem: a
new and innovative framework. IEEE Trans. Learn. Technol. 15(6), 685–696 (2022)
27. Yarramreddy, A., Gromkowski, P., Baggili, I.: Forensic analysis of immersive virtual
reality social applications: a primary account. In: 2018 IEEE Security and Privacy
Workshops (SPW), pp. 186–196. IEEE (2018)
Objective Emotion Quantification
in the Metaverse Using Brain Computer
Interfaces
1 Introduction
The advent of better and inexpensive virtual reality (VR) consumer hardware
has stimulated (neuro) psychological attention and emotional research with mul-
timedia based immersive environments. Particularly, the Metaverse-based VR
experiences offers opportunities for individuals to express emotions in diverse
ways which provide a range of expressions that contribute to a richer emotional
data for analysis using classification algorithms. The multimedia environments
are dynamic and customizable and emotion classification may contribute to emo-
tional analytics for virtual spaces. In this study, we analyze human emotion by
capturing EEG data using BCI’s (IoT real-time sensors) while the users are
exploring various multimedia scenarios suitable for the metaverse. By assessing
emotional cues and patterns during virtual interactions, emotion classification
algorithms can support features such as emotion-based matchmaking, emotional
chat-bots, or emotional sentiment analysis of online communities. Understand-
ing the feelings of other people is part of successful social interactions in humans’
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 353–361, 2023.
https://doi.org/10.1007/978-3-031-45878-1_24
354 A. O. Muresan et al.
seen as a crucial element for research in VR. Evidence suggests that immersion,
presence, and emotion are related only to arousing, but not to non-arousing
neutral content. However, the exact relationship between immersion, presence,
and emotion remains unclear [2,4], hence the importance of research on emotion
and attention effects in VR. Alternatively, in the field of human-computer inter-
action, the correlation of attention and emotion can tell us about the quality
of the Metaverse experience. It has been shown that emotional responses, such
as arousal ratings, are enhanced in Metaverse experiences, e.g., by using either
a “low immersion” or a “high immersion” with the surrounding presentation.
The following is the outline of the remaining segments of this paper. Section 2
provides a literature review on EEG-based techniques for human emotion clas-
sification both in virtual and non-virtual environments. Section 3 elaborates on
the methodology and experimental setup. Section 4 presents the experimental
results while Sect. 5 presents the overall findings and conclusion.
2 Related Work
Several researchers worked on emotion recognition utilizing the signals from
EEG data by combining the Virtual Reality environment. Longo B et al. [1]
356 A. O. Muresan et al.
integrated BCI technology with VR to enable direct brain control for stroke
patients. The results demonstrated promising outcomes in motor performance
improvements, functional gains, and changes in brain activity and connectivity.
Yan et al. [5] examined the existing research landscape, identifies key trends, and
themes, and proposes future research directions in the field of emotion recogni-
tion using BCIs. The findings and research agenda contributed to the advance-
ment of emotion recognition and provide directions for further exploration and
development in the field. Özerdem et al. [10] focused on emotion recognition
using EEG features extracted from movie clips and employing channel selec-
tion techniques. The results demonstrate the potential of EEG-based emotion
recognition in response to multimedia stimuli. Javier Marı́n-Morales et al. [12]
developed a system that can accurately detect and classify emotions based on
physiological signals obtained from individuals immersed in a VR environment.
The study described the experimental setup, which involved participants wearing
wearable sensors while engaging in a VR scenario designed to elicit specific emo-
tions. The captured EEG data were preprocessed and analyzed with Machine
Learning algorithms, such as Support Vector Machines (SVM) or Deep Neu-
ral Networks (DNN). The findings suggested that wearable sensors can provide
valuable insights into emotional states during VR experiences. Similarly, David
Schubring et al. [13] investigated the effects of multimedia data on alpha/beta
brain oscillations in relation to emotions and cognitive tasks. Their findings have
implications for understanding the neural processes underlying emotions and
cognition in VR environments. Hongyu Guo et al. [14] developed an emotion-
based analysis method for designing English language lessons in the Metaverse.
They have put forward a compelling argument on integrating emotional fac-
tors into the instructional design process would help to enhance learner engage-
ment, motivation, and language acquisition outcomes. Their work emphasized
the potential benefits of the metaverse in language instruction and provides ped-
agogical considerations for successful implementation. Eduardo Perez-Valero et
al. [15] used a VR and EEG combo to quantitatively analyze the stress level
during a stress-relaxation session. They used changes in power spectral density
and coherence within these frequency bands were examined to identify stress
and relaxation patterns. Similarly, Rhaı́ra Helena Caetano E Souza et al. [16]
explored the feasibility of utilizing EEG to measure the attentional states of indi-
viduals immersed in virtual scenarios. Techniques such as Power Spectral Analy-
sis (PSA), Time-Frequency (TFA) Analysis, or ERP components extraction were
employed to identify attention-related patterns. The International Affective Pic-
ture System (IAPS) is a standardized set of images widely used in psychological
and neuroscience research to elicit emotional responses. Lang et al. [17] worked
on preparing the manual providing comprehensive information on the IAPS,
including detailed instructions on its use and a compilation of affective ratings
for the image stimuli. The manual outlines the procedures for image selection,
and normative ratings, and provides an extensive database of affective ratings
for the stimuli. Yan et al. [18] successfully implemented an emotion recognition
model that utilizes EEG data, focusing on the rhythm and time characteristics
Objective Emotion Quantification in the Metaverse Using BCI 357
of brain-wave signals. They used features such as power spectral density, signal
entropy, and wavelet transform coefficients to capture the unique rhythmic and
temporal properties of the EEG signals. The data were classified using Rhythmic
Time EEG Emotion Recognition Model generating the highest average recog-
nition accuracy 0.69 which is 0.07 higher than the traditional SVM and KNN
models. Nazmi Sofian Suhaimi et al. [19] examined the current trends and oppor-
tunities in the field, focusing on the use of EEG signals to detect and classify
human emotions.
Both neural network evaluations consist of 8 input nodes, 3 hidden nodes, 2 out-
put nodes. The first type of data analysis utilizes Tensor Flow Keras Sequential
Neural Networks Software to objectify and forecast possible VAD values against
the SAM reported values (outputs) and considers the users’ brainwaves as inputs,
materializing the experiment with correlation indexes for each VAD parameter,
ranging from 0 to 1. The methodology outcomes suggest that there might be a
connection between brainwave data and VAD parameters that can be predicted
with 89% accuracy for Dominance, and around 70% accuracy for Valence and
Arousal, which may lead to promising results if confirmed by enlarging the data
pool. Figure 2 showcases the performance for each VAD parameter.
The second type of data analysis implements MatLab Neural Networks and
still considers the collected brainwave data as predictor variables (inputs), and
the attention and meditation levels provided by the headset as outputs. The
outcome of this appraisal are the regression indexes, showcasing the degree of
correspondence between the predictor variables and response variables (outputs).
The dataset is split in 3 main categories: training, validation and test; each
358 A. O. Muresan et al.
Fig. 2. Juxtaposition of actual and forecasted VAD parameters using TFKS Neural
Networks
of them has its respective regression value, as well as computing the overall
regression index, all ranging from 0 to 1. In examining the results, one may
observe that the training regression level is quite high at 0.85, this might suggest
that the SAM subjective analysis and both the meditation and attention levels
are intertwined to some degree. The overall results infer that the Dominance
emotion is slightly more intensely experienced by the participants in the study.
The neural network-based analysis utilized in the present setup supports this
theory, however for a higher accuracy analysis, a larger pool of subjects may
be required. Nevertheless, our results prove that emotion recognition can be
objectively quantified using BCI’s data. Moreover, a hybrid approach where
EEG data is combined with facial expression or other form of emotion detection
may significantly improve prediction accuracy. The data process is implemented
by two investigative pathways, both utilizing neural networks as an engine for
Objective Emotion Quantification in the Metaverse Using BCI 359
data analysis and considering the collected brainwave data as inputs. Both which
are trained to recognize patterns and classify emotions based on the extracted
features (Fig. 3).
5 Conclusion
References
1. Longo, B., Castillo, J., Bastos, T.: Brain-computer interface combined with virtual
reality environment (VRE) for inferior limbs rehabilitation in post-stroke subjects.
In: SBBIOTEC, BMC 2014, vol. 8 (2014). https://doi.org/10.1186/1753-6561-8-
S4-P18
2. Statthaler, K., Schwarz, A., Steyrl, D.: Cybathlon experiences of the Graz BCI
racing team Mirage91 in the brain-computer interface discipline. J. Neuro Eng.
Rehabil. 14, 1–16 (2017). https://doi.org/10.1186/s12984-017-0344-9
3. Torres, E., Hernández-Álvarez, M., Yoo, S.G.: EEG-based BCI emotion recogni-
tion: a survey. Sensors 20, 5083 (2020). https://doi.org/10.3390/s20185083
4. NeuroSky Homepage. https://neurosky.com/2016/11/bci-what-is-it-and-where-
is-it-going/. Accessed 3 Nov 2016
5. Yan, W., Liu, X., Shan, B., Zhang, X., Pu, Y.: Research on the emotions based on
brain-computer technology: a bibliometric analysis and research agendas. Front.
Psychol. 12, 771591 (2021). https://doi.org/10.3389/fpsyg.2021.771591
6. UWA. https://online.uwa.edu/news/emotional-psychology/. Accessed 22 June
2020
7. Liu, Y., Zhang, Y., Tao, D.: Wearable sensors for emotion recognition: a compre-
hensive review. IEEE Trans. Affect. Comput. 10 (2019)
8. Ekman, P.: An argument for basic emotions. Cogn. Emot. 6, 169–200 (1992).
https://doi.org/10.1038/s41598-021-85163-z
9. Özerdem, M.S., Polat, H.: Emotion recognition based on EEG features in movie
clips with channel selection. Brain Inform. 4(4), 241–252 (2017). https://doi.org/
10.1007/s40708-017-0069-3
10. Russell, J.A.: Core affect and the psychological construction of emotion. Psychol.
Rev. 110, 145 (2003). https://doi.org/10.1037/0033-295X.110.1.145
11. Baumeister, R.F., Stillwell, A.M., Heatherton, T.F.: Guilt: an interpersonal app-
roach. Psychol. Bull. 115, 243 (1994). https://doi.org/10.1037/0033-2909.115.2.
243
12. Llinares, C., et al.: Affective computing in virtual reality: emotion recognition from
brain and heartbeat dynamics using wearable sensors. Sci. Rep. 8, 13657 (2018).
https://doi.org/10.1038/s41598-018-32063-4
13. Schubring, D., et al.: Virtual reality potentiates emotion and task effects of
alpha/beta brain oscillations. Brain Sci. 10, 537 (2020). https://doi.org/10.3390/
brainsci10080537
14. Hongyu, G., Wurong, G.: Metaverse-powered experiential situational english-
teaching design: an emotion-based analysis method. Front. Psychol. 13, 859159
(2022). https://doi.org/10.3389/fpsyg.2022.859159
15. Perez-Valero, E., et al.: Quantitative assessment of stress through EEG during a
virtual reality stress-relax session. Front Comput. Neurosci. 15, 684423 (2021).
https://doi.org/10.3389/fncom.2021.684423
16. Souza, R., et al.: Attention detection in virtual environments using EEG signals: a
scoping review. Front. Physiol. 12, 727840 (2021). https://doi.org/10.3389/fphys.
2021.727840
17. Lang, P.J., et al.: Technical manual and affective ratings of the International Affec-
tive Picture System (IAPS), NIMH Center for the Study of Emotion and Attention
(1997)
Objective Emotion Quantification in the Metaverse Using BCI 361
18. Yan, J., Chen, S., Deng, S.: An EEG-based emotion recognition model with rhythm
and time characteristics. Brain Inf. 6, 1–8 (2019). https://doi.org/10.1186/s40708-
019-0100-y
19. Suhaimi, N.S., Mountstephens, J., Teo, J.: EEG-based emotion recognition: a state-
of-the-art review of current trends and opportunities. CIN 2020 (2020). https://
doi.org/10.1155/2020/8875426
MetaHap: A Low Cost Haptic Glove
for Metaverse
1 Introduction
Virtual reality (VR) technology has come a long way since its inception [1]. It is
now being used in a wide range of applications, from gaming and entertainment
to education, healthcare, and industrial training [2]. One of the main benefits
of VR is its ability to provide an immersive and interactive experience to users,
allowing them to feel as if they are part of the digital environment. However,
this sense of immersion can be limited by the lack of tactile feedback in most
VR systems. This is where haptic gloves come in [3].
Haptic gloves are a type of wearable device that use sensors and actuators to
simulate the sense of touch, allowing users to feel virtual objects and textures
as if they were real. These gloves are equipped with small motors or other types
of actuators that apply pressure or vibrations to the user’s fingers, creating
the illusion of touch [4]. This technology can enhance the immersion of VR
experiences and make them more realistic and interactive [5].
The development of VR haptic gloves has been an active area of research
for several years. There have been numerous prototypes developed, each with
its own unique design and set of features [6]. The early versions of haptic gloves
were bulky, expensive, and not very practical for most applications. However,
recent advancements in technology have led to the development of more compact
and affordable gloves, making them more accessible to a wider range of users [7].
In recent years, there has been growing interest in the potential applications
of haptic gloves in various fields, such as gaming, music, education, and health-
care. In gaming, haptic gloves can be used to provide a more immersive and
realistic experience by allowing users to feel the impact of virtual objects and
interactions. In music, haptic gloves can simulate the experience of playing dif-
ferent types of instruments, making it easier for users to learn and practice. In
healthcare, haptic gloves can be used to simulate medical procedures and train-
ing scenarios, allowing medical students to gain hands-on experience in a safe
and controlled environment. Overall, haptic gloves have the potential to enhance
the immersion and realism of VR experiences, making them more engaging and
interactive for users. The development of haptic gloves is an exciting area of
research that has the potential to revolutionize the way we experience and inter-
act with digital content [8–10].
1.1 How is Our Model Used in MetaHap Differs from the Existing
Ones
Its an easy to design and implement model that provides the user accurate and
reliable experience of the virtual world. This model will let the user experience
the virtual world, in the same way they experience the real world with their own
hands. This model has been implemented with efficient and accurate position
tracking algorithms along with finger tracking and haptic feedback algorithms,
that will allow the user to freely move their hands around in the virtual world.
However this model will not be as advanced as the ones available out their as
this is being implemented on a single processor system, but this model will be
364 S. Sibi Chakkaravarthy et al.
2 Literature Review
Optical methods for position tracking provide diverse and effective approaches
for accurately determining the position and movement of objects. Camera-based
tracking utilizes computer vision techniques and image analysis to track objects
based on visual data captured by cameras. Marker-based tracking relies on
MetaHap: A Low Cost Haptic Glove for Metaverse 365
Figure 2 shows the workflow of PTU algorithm. The position tracking unit con-
sists of three sensors Accelerometer, Gyroscope and GPS. To track the yaw, roll,
and pitch angles, a fusion of accelerometer and gyroscope sensors is employed
[13]. This fusion is achieved through a meticulously designed on-board algorithm
based on quaternions. Additionally, on-board differentiation algorithms are uti-
lized to extract the desired values. These calculated angles are then merged and
filtered using a combination of filters, ensuring the output attains a high level
of reliability and stability. By employing these techniques, the system achieves
accurate and consistent tracking of orientation, enhancing the overall perfor-
mance and user experience.
The combined utilization of accelerometer and GPS sensors enables both
absolute and relative position tracking of the glove within a three-dimensional
space. The accelerometer readings undergo processing through an on-board inte-
MetaHap: A Low Cost Haptic Glove for Metaverse 367
gration algorithm. Subsequently, these processed readings are merged with abso-
lute position data obtained from the GPS module, employing sensor fusion
algorithms. The resulting merged data is then subjected to filtering techniques,
enhancing the accuracy and reliability of the readings. This approach empowers
the system to accurately calculate the real-time position of the glove in three-
dimensional space, relative to an absolute reference point.
Here, 0 <= Current f inger pose <= 90, 0 <= counter value <= 180 (2)
368 S. Sibi Chakkaravarthy et al.
The overall design of the model comprises two distinct components: the primary
system and the secondary system. This design allows for flexibility and extensi-
bility, enabling users or developers to modify or enhance the model with ease.
If desired, additional features can be incorporated by simply programming and
attaching a separate secondary system to the primary system via the I2C pro-
tocol. Alternatively, lightweight programs or features can be directly added to
the primary system.
For example, consider the model described in Fig. 4, which encompasses
finger pose tracking and haptic feedback. If a developer wishes to introduce
pressure sensing for each finger to determine the user’s grip strength on virtual
objects, they can implement this functionality using an additional secondary
system. This secondary system would be connected to the appropriate sensors
and subsequently linked to the primary system via the I2C protocol. The primary
system would then collect data from this new secondary system and transmit
it to the virtual environment to generate the desired output, such as causing
an object to break if held too tightly. Also notice in the provided diagram, the
position tracking unit has been implemented on a separate secondary system to
accommodate heavy algorithms that may be impractical to execute on a single
system. This modular approach enhances the model’s scalability and adaptability
to evolving requirements.
In this section, the development of the VR haptic glove using SBC and other var-
ious sensors, such as gyroscopes, accelerometers, GPS, servos, and encoders, has
370 S. Sibi Chakkaravarthy et al.
been successfully achieved. The glove incorporates haptic feedback, finger pose
tracking, and motion tracking capabilities, providing users with a highly immer-
sive and realistic virtual reality experience [6]. Figure 5 shows the experimental
hardware setup of MetaHap.
The model’s finger pose tracking mechanism, utilizing rotary encoders and
advanced algorithms, accurately calculates the angles of finger joints, enabling
precise tracking and realistic interactions within the virtual environment. The
integration of accelerometer and gyroscope sensors allows for accurate motion
tracking, enhancing the user’s immersion and control. The fusion of accelerom-
eter and GPS sensors enables absolute and relative position tracking in three-
dimensional space, providing users with an enhanced sense of spatial presence.
The implementation of collision detection and prevention using servos ensures
the safety and stability of the glove’s movements. The servo-based mechanism
effectively responds to collision signals, freezing the movement of the affected
finger and preventing any unintended actions. This feature enhances the overall
user experience and reduces the risk of accidental collisions or improper finger
poses.
MetaHap: A Low Cost Haptic Glove for Metaverse 371
6 Conclusion
The proposed MetaHap, a VR haptic glove demonstrates promising capabili-
ties in providing a highly immersive and interactive virtual reality experience.
The integration of various sensors, along with advanced algorithms, allows for
accurate finger pose tracking, motion tracking, and position tracking within the
virtual environment. The implemented features, such as collision detection and
prevention, ensure the safety and stability of the glove’s movements, providing
users with a seamless and reliable interaction experience. The modularity of the
design allows for easy expansion and customization, enabling users and devel-
opers to incorporate additional features or modify the glove according to spe-
cific requirements. MetaHap serves as a foundation for further advancements in
VR haptics using simple microcontrollers. The use of cost-effective components
and the integration of efficient algorithms have the potential to reduce overall
costs and increase the number of features in future iterations. With continuous
research and development, the VR haptic glove holds a promising development
for a wide range of applications, including gaming, virtual musical instrument
simulation, medical training, and beyond.
References
1. Helmold, M.: Extended reality (XR) in QM. In: Helmold, M. (ed.) Virtual and
Innovative Quality Management Across the Value Chain: Industry Insights, Case
Studies and Best Practices, pp. 21–26. Springer, Cham (2023). https://doi.org/10.
1007/978-3-031-30089-9 3
2. Wider, W., Jiang, L., Lin, J., Fauzi, M.A., Li, J., Chan, C.K.: Metaverse chronicles:
a bibliometric analysis of its evolving landscape. Int. J. Hum.-Comput. Interact.
1–14 (2023)
3. Qi, J., Gao, F., Sun, G., Yeo, J.C., Lim, C.T.: HaptGlove-untethered pneumatic
glove for multimode haptic feedback in reality-virtuality continuum. Adv. Sci. 10,
2301044 (2023)
4. Kuhail, M.A., Berengueres, J., Taher, F., Alkuwaiti, M., Khan, S.Z.: Haptic sys-
tems: trends and lessons learned for haptics in spacesuits. Electronics 12(8), 1888
(2023)
5. Sinciya, P.O., Orethu, J.A., Philip, M.A., Prakash, N., Jacob, J.: Multipurpose
immersive virtual reality environment. In: 2023 2nd International Conference
on Applied Artificial Intelligence and Computing (ICAAIC), pp. 855–860. IEEE
(2023)
6. Civelek, T., Arnulph, F.: Virtual reality learning environment with haptic gloves.
In: 2022 3rd International Conference on Education Development and Studies, pp.
32–36 (2022)
7. Pezent, E., Agarwal, P., Hartcher-O’Brien, J., Colonnese, N., O’Malley, M.K.:
Design, control, and psychophysics of tasbi: a force-controlled multimodal haptic
bracelet. IEEE Trans. Rob. 38(5), 2962–2978 (2022)
8. Dangxiao, W., Yuan, G., Shiyi, L., Zhang, Y., Weiliang, X., Jing, X.: Haptic display
for virtual reality: progress and challenges. Virtual Real. Intell. Hardw. 1(2), 136–
162 (2019)
372 S. Sibi Chakkaravarthy et al.
9. Wu, C.M., Hsu, C.W., Lee, T.K., Smith, S.: A virtual reality keyboard with realistic
haptic feedback in a fully immersive virtual environment. Virtual Reality 21, 19–29
(2017)
10. Perret, J., Vander Poorten, E.: Touching virtual reality: a review of haptic gloves.
In: ACTUATOR 2018; 16th International Conference on New Actuators, pp. 1–5.
VDE (2018)
11. Edwards, B.I., Bielawski, K.S., Prada, R., Cheok, A.D.: Haptic virtual reality and
immersive learning for enhanced organic chemistry instruction. Virtual Reality 23,
363–373 (2019)
12. Antonov, V.O., Arustamov, D.A., Zavolokina, U.V., Apurin, A.A.: A method for
controlling groups of unmanned aerial vehicles in a virtual environment using hap-
tic gloves. In: IOP Conference Series: Materials Science and Engineering, vol. 1069,
no. 1, p. 012043. IOP Publishing (2021)
13. Li, F., Chen, J., Ye, G., Dong, S., Gao, Z., Zhou, Y.: Soft robotic glove with sensing
and force feedback for rehabilitation in virtual reality. Biomimetics 8(1), 83 (2023)
Technologies for Smart Agriculture
(TSA)
CroPAiD: Protection of Information in
Agriculture Cyber-Physical Systems Using
Distributed Storage and Ledger
Abstract. The agricultural domain has had a significant role throughout history
in human societies across the globe. With the fast growth of communication and
information systems, the structure of farming procedures has evolved to new
modern standards. Although multiple features helped gain from these advance-
ments, there are many current and rising threats to security in the agricultural
domain. The present paper gives novel methods and architectural designs and
implements distributed ledger through the Tangle platform. Initially, the article
discusses the threats and vulnerabilities faced in the farming sector and presents
an extensive literature survey, and later conducts an experiment for distributing
data through a tangle distributed ledger system. The authors highlight the limita-
tions of central, cloud, and blockchain and suggest mitigation measures through
distributed IOTA systems and distributed storage facilities for data and the pos-
sible influence these solutions can bring in the aspects of data security in the
agricultural sector.
1 Introduction
there are two main concerns, including the collection and storage of farming data. As
there is a variety of data residing in fields, conventional monitoring methods cannot be
applied, and also the data can be exposed to human errors. With the increase in infor-
mation and communications systems, the number of cyber crimes have also increased
worldwide, stealing and harming a variety of assets. Existing cloud systems have mul-
tiple limitations of central storage, continuous connectivity, security risks through dif-
ferent providers, bandwidth constraints, and faint backup procedures. The cyber secu-
rity procedures combine different techniques to give high protection against attacks on
data [2]. Blockchain (BC) is a Distributed Ledger Technology (DLT), a new way to
share data securely through cryptographic hashes in a distributed platform that is being
applied in various domains nowadays. The agricultural field is also taking advantage
of the decentralized features of blockchain, but these systems are going through com-
plex steps to generate blocks and consume energy with on-chain storage [3]. To face
issues in blockchain, the applications are designed with off-chain storage solutions and
exercise other methods to avoid cost, latency, and energy consumption.
Tangle is another Distributed Ledger Technology (DLT) that does not require fees
and validates the transaction nodes at full speed. The transactions are said to be valid
if the previous two transaction’s history does not conflict with the current transaction.
The consensus mechanism, proof of Work (PoW), is not used for validating the trans-
action but for keeping the network secure from spam. The overall throughput of the
tangle is infinite, and the consensus mechanism is used for defining the limits of the
throughput. A distributed storage (IPFS) system is a platform for storing data, web-
sites, applications, and accessing files [4] and does everything a central system does
but without a central storage system. Some of the motivations for implementing current
paper CroPAiD are listed in the Fig. 2. Combining both tangle and IPFS can bring more
security and privacy to sensitive agricultural data. The information stored on the IPFS
network generates a cryptographic hash through a content identifier to retrieve the data
later securely.
CroPAiD: Protection of Information in Agriculture Cyber-Physical Systems 377
The paper follows the given order: We discuss various prior works using conven-
tional and modern methods for transmitting and storing sensitive agricultural data in
Sect. 2. The Sect. 3 elaborates on problems raising through modern and traditional
methods and list the novel solutions provided through current system. In Sect. 4 and
Sect. 5, we define various components, provide a state of the architecture for CroPAiD
and give algorithm steps for navigating the data over IPFS and Tangle, respectively. The
implementation of the system and the results obtained are shown in Sect. 6 followed by
the conclusions and further research aspects in Sect. 7.
2 Related Works
For agricultural data storage, usually, the data is collected in conventional local
databases or cloud systems. In the current agricultural 4.0 era, many researchers and
scholars are conducting profound studies on how modern data storage methods can be
introduced. Launching ledger technology into IoT for data security can be done in two
scenarios; the first is making use of off-chain data storage with the help of distributed
storage-IPFS or traditional local databases. The second is the direct storage of data on
distributed ledger systems.
The paper [5] G-DaM sends the data collected from the Internet of Things to the
near edges for storing the data in distributed platforms and public blockchain technol-
ogy. The application overcomes the traditional data sharing and limitations of central
and cloud systems and increases the quality and integrity of the data. The agroString
[6] proposed an intelligent IoT-based edge system for the management of data through
a private corDapp application. The system sends the information collected from the
IoT edge sensors through the private blockchain to avoid traditional public blockchain
systems’ costs and energy consumption and evade bottlenecks of central and cloud sys-
tems. The application implements an IoAT-edge for collecting temperature and humid-
ity datasets and sends those readings to the corDapp to bring integrity, trust, visibility,
and data quality to each supply chain stakeholder.
378 S. L. T. Vangipuram et al.
The paper [7] uses blockchain for fruit and vegetable traceability to overcome tra-
ditional centralized systems. With the help of a dual storage structure- “database +
blockchain” the system designs the application using an on-chain and off-chain storage
technique to reduce the load pressures and increase the data integrity throughout the
supply chain. The results of the system exhibit improved security of secluded infor-
mation and further enhance the authenticity and reliability of data. Information modi-
fications and tampering with the sensitive data collected in a supply chain may lead to
serious issues regarding the quality and safety of the end product. The article [8] makes
use of blockchain for time stamping, traceability, and tamper-proofing of data with the
help of smart contracts. The solidity language contract manages the agricultural product
transactions with access control and improves the upload and response times.
Crop monitoring is essential to keep a check for pests, weeds, and diseases in the
crops. Monitoring is done using different sensors to see the current state of the prod-
uct to project and predict what will be the next state and issues arising in the crops.
The farmer takes preventive measures accordingly based on the information collected
in the crop monitoring. Field monitoring plays a vital role in increasing crop yield,
and modern IoT technology and communication systems are beneficial in fulfilling this
requirement. An efficient crop monitoring system is proposed in sFarm [9] through a
sensor to collect the data and share the real-time data securely using IOTA Tangle dis-
tributed ledger platform. With the help of IOTA, the central, cloud, public, and private
blockchain limitations are overcome, saving energy and time for uploading and vali-
dation. Many distributed access control technologies through blockchain are already in
practice for dealing with centralized and cloud network limitations, but they, too, inherit
some drawbacks, such as high fee transactions and low throughput. The paper [10] pro-
poses a novel access control framework based on IOTA that enables free transactions
with higher throughput.
Using Ciphertext-Policy Attribute-Based Encryption (CP-ABE) technology, access
rights are encrypted to provide access control and store the data on the distributed
ledger Tangle. IOTA Tangle has some disadvantages and security threats, such as a
parasite chain attack that is a common double-spending attack. To decrease these types
of attacks, the paper [11] gives an efficient method for detecting a parasite chain. The
authors measure a score function at each IOTA transaction to see the importance level.
Any change in this importance is reflected in the 1st and 2nd order of the derivates, thus
giving accurate results in detecting the parasite chain attack. All the above-discussed
prior works try to improve the security in transmitting and storing agricultural data, but
Table 1. Comparing Prior works with Current application CroPAiD.
the current system adds additional features of distributed storage of IPFS to the IOTA
distributed ledger platform to overcome conventional and modern limitations as given
in the Table 1.
With the increase in demand for food in the global market, the need for reduced costs
and increased agricultural production has given way to using new technologies, which
is an attractive choice for farmers and companies [12]. Some of the advantages of IoT
applications in the agricultural sector include crop health monitoring, pest infestation,
water management, frost protection, and decision support. This new method of using
novel communication technologies in agriculture is denoted as precision agriculture or
precision farming (PF) [13]. The use of satellites and GPS in farming helps in digitizing
agricultural measurements to see the accuracy and efficiency of the crop. Based on
the measurements collected from these precision agriculture tools, the farmers and the
experts in the field, study and analyze the variations of crops and livestock data. To
collect different types of information from the fields, the farmer would use IoT nodes
that come with specific features that make them useful in limited domains [14].
With the emergence of new communication systems and the addition of the Inter-
net of Things (IoT) in farming, unknown security risks and data threats arise in the
cyber-physical environment. These data risks are mainly related to cyber security, data
integrity, and data loss disturbing the stakeholder businesses [15]. The constraints of the
IoAT are given in Fig. 3. Precision Farming uses vast modern machinery in the fields,
leading to higher consequences and threats. Farming is possible in open grounds where
weather and environmental conditions are inconsistent, leading to malfunctioning of the
machinery and technical equipment in the fields, resulting in wrong measurements and
hence wrong analysis [16]. Additionally, the temperature and humidity conditions can
affect sensor things for communication, which can lead to data loss [17]. The cyber-
security issue is a worldwide severe threat activity that uses a smart device to access
sensitive personal and government information. Although strict restrictions have been
implanted through law enforcement, the hackers take advantage of internet anonymity
and attack middleware, network, and application layers [2].
– A unique system is designed with Tangle to increase the quality of data and avoid
drawbacks of sensor things.
– To move bulk data to IOTA and avoid double spending issues of Tangle, the current
system uses distributed storage systems near the edges.
– The imitations of conventional storage databases, cloud, and central systems are
circumvented using the IOTA distributed ledger platform.
– Increasing security, data integrity, and evading data tampering by the IOTA system.
– Overcoming blockchain high transaction fees and energy usage through distributed
ledger system of Tangle.
– Using Double hashing procedure for the agricultural data through IPFS and Tangle
to increase security and privacy of data.
– A state-of-the-art architecture is presented for the current system CroPAiD.
– Designing a Cost-efficient infrastructure and presenting results with zero transaction
fees and secured hashes.
The cyber-physical systems (CPS) combines the software and hardware components to
execute a well-defined task. A system that connects and manages the physical attributes
towards its computing capabilities and a design that connects and controls the physi-
cal organizations with virtual structures through networks. The combination of wireless
sensor networks that supervise the physical entities can enhance itself in real-time sce-
narios. The CPS are applied in multiple domains to help in substituting conventional
CroPAiD: Protection of Information in Agriculture Cyber-Physical Systems 381
methods and integrating various platforms and technologies together [18]. Smart Agri-
culture is one of the domains that can benefit from CPS due to its modern and smarter
applicability in monitoring and controlling farming activities and gathering the infor-
mation associated with crops, soil, livestock hygiene, and weather in real-time, along
with maintaining the environment and preserving energy. Figure 4 gives different layers
of cps and their connectivity in physical systems through smart devices to control and
manage the data in an intelligent way.
One of the limitations of Tangle is that the attackers can implant several duplicates of the
data that can lead to double-spending transactions [19]. A user can create and spend the
same digital asset multiple times, which must be checked and prevented. A distributed
storage-IPFS or Interplanetary File System is an internet protocol mainly to store data,
avoid data or asset duplicates across the network, and collect the addresses of the data
in the network. By stopping asset duplication, the IPFS can help in avoiding double
spending issues. By using IPFS as off-chain storage for IOTA tangle, the information is
stored in a distributed platform, reducing local database, central, and cloud constraints.
The data is recognized through content, and every piece of information is divided into
256 kb maximum length blocks. Every block is labeled with a unique identifier for the
content through the cryptographic hash. The distributed hash table of IPFS is based
on the principle of distributed key-value store. Both node identifiers and distance met-
rics strategies in IPFS help in storing and retrieving the data quickly. When reading or
writing the data from or to the edges, the end devices search for the nodes close to the
key attribute values using buckets inside the networks to identify the nodes [4, 20]. The
382 S. L. T. Vangipuram et al.
S/Kademlia algorithm is used for DHT in IPFS to register the nodes whenever a file
gets uploaded and links to nodes through an identifier for file retrieval.
servers. A distributed storage technology-IPFS is embedded into the edge along IOTA
Tangle that serves as a gateway between IoT devices and servers. With distributed stor-
age on edge, the limitations of IOTA, such as double spending and other attacks, are
evaded. Each agricultural sensor data file is transmitted to IPFS to generate a hash,
as explained in Subsect. 4.2, through its unique content identifier. A Merkle-directed
acyclic graph (Merkle-DAG) calculates a root that can retrieve the original file from
the segments. The distributed storage hash of the crop’s sensor data is then moved
toward the Tangle residing in the edge. The IOTA node receives the hash from the
IPFS and further secures it by generating Tangle hashes using MAM and STREAMS
tools, as discussed in Subsect. 4.3 above. The distributed ledger technology is feasible
for point-point, point-multipoint, and multipoint-multipoint communications between
various sensor devices on the field and multiple servers.
The data from the Internet of Things moves toward the edge layer that has both
Distributed Storage-IPFS (DS) and IOTA Tangle systems implemented. Algorithm 1
presents phases in transferring crop data(Cd ) to IPFS and generating 256 kb buffer files
to give a root hash at the end. In the edge distributed storage system (DSE ), both pri-
vate (DSpr ) and public keys (DSpu ) are generated to incorporate access control through
digital signatures and signing crop data files. The IPFS converts the crop data (Cd ) into
a 256 kb buffer file (Cd ) and signs the buffer file (Cdbf 265KB )to get a root hash file of
the crop data (H(Cd Cbf )) where H denotes the hash of the crop data file.
384 S. L. T. Vangipuram et al.
Each input data present in the Tangle creates the following fields: data-length, data,
public key, private key, index, index-next, sign, and auth-sign. The IOTA tangle gen-
erates a seed (Sd ) from a random source and produces a key pair for input data using
the edwards25519 curve algorithm. Each input data calculates the index and the index-
next via private and public keys. The hash of the public key is the index, and the hash
of the public key for the following input data is the index-next. A different key pair
is generated for the next input data from another random source, and for hashing, the
algorithm used in IOTA is BLAKE2b [24]. Computing index and index-next are signif-
icant because they help in continuous data streaming, data ownership, verification, and
authentication. A digest ‘d’ is given by hashing the data, data-length, public key, and
index-next. The sign field is then calculated by signing the digest with the private key.
This will be helpful in verification later for the user. If the user has to verify the data,
compare the hash and sign field values with the public key in the input data. If both are
equal, then the data is verified correctly. The sign field helps in only verification of the
data but does not give authenticity or the author’s identity. The field auth-sign is calcu-
lated by the key pair associated with the sensor device. This authorization signature is
calculated by the private key of the IoT device and stored in a hardware source along
with the public key certificate. To validate and see the authentication of the data, the
user compares the signature with the public key through a trusted third-party certificate
authority. The algorithm 2 and the Fig. 6 show the flow of Crop data in the Edge layer
between IPFS and IOTA Tangle in detail and explains how the data is moved, verified,
and authenticated in Tangle.
‘H(Cdbf 265KB )’ is taken as the input data to the Tangle ledger System. We represent
that input data ‘H(Cdbf 265KB )’ in the following algorithm as ‘Iniota ,’ data-length as
‘Iniota len,’ the public key as ‘Intangle Pukey,’ private key as ‘Intangle Prkey,’ index as ‘I’
and next-index as ‘n-I,’ sign field as ‘sign’ and auth-sign as ‘authsign .’ For Hashing and
digest, we represent with letters ‘H’ and ‘d’.
CroPAiD: Protection of Information in Agriculture Cyber-Physical Systems 385
Fig. 6. CroPAiD detailed Data flow in the Edge layer between IPFS and IOTA Tangle.
6.1 Datasets
The agricultural datasets we used are from the Kaggle [25] source. Each data belonged
to different vegetables and fruits containing images of healthy and diseased crops. These
data collected are sensitive and usable for further research and analysis in bringing
improvements in farming and are also beneficial in the field of agricultural science. We
uploaded the crop data in the current application to test and validate. Table 2 demon-
strates different crop statistics we used for the present paper.
The Fig. 10 shows some sample dataset images we used for storing and sharing in
the CroPAiD application through IPFS and IOTA. When a crop gets infected, it dam-
ages and changes all the primary functions of the food that can harm humans when con-
sumed. This type of crop infected data is beneficial in predicting future crop damage
CroPAiD: Protection of Information in Agriculture Cyber-Physical Systems 387
and helps improve crop yield. Therefore, such data is crucial for farmers and scientists
to take precautions and perform research and study. This data need to be transmitted
in a secure manner without any tampering for correct analysis. The Fig. 10(a), 10(b),
10(c), 10(d), 10(e), 10(f) show pictures of a healthy crop and a diseased crop of apple,
potato, cherry, corn, grape and tomato correspondingly.
To obtain the results for the current application, we have used Intel(R) Core(TM) i9-
10885H CPU @ 2.40 GHz, 32.0 GB RAM as the edge layer. We have deployed the
application logic of IPFS and IOTA tangle in this edge system. We upload the crop data
file to the IPFS node to get the hash of the file, as shown in the Table 3. The IPFS hash
file generated does not have the time stamp but avoids duplicates and double-spending
attacks on the data transferred. The application further takes the IPFS hash as an input
to the IOTA node to give another hash from the tangle platform. The Table 3 shows the
double hashes produced by both IPFS and Tangle. The application has been tested with
different sizes of crop data to produce two hashes with both technologies. Once both the
hashes were received from the application, we used the message unique ID to retrieve
the original file. The time to upload the files was very minimal, and the data transactions
costs were zero compared to blockchain latencies and transaction fees. The paper we
implement combines distributed storage IPFS and IOTA Tangle successfully, resulting
in higher data security with reduced energy consumption nullifying the limitations of
central, cloud, conventional database, and blockchain systems.
388 S. L. T. Vangipuram et al.
File Name Reduced Size IPFS Hash Tangle Hash Txn Time (Sec)
Apple-healthy 23.8 MB QmXWpe6Q5v9qH7Wwgr SKJYF76R3947IRYREIU59 35
5HH5BmB78Q2u4wP 8475FHKEUR834759IFKR30W
WFd7NkvooFZrP PWEKDSVLDKFROIRFHDKJ
Apple-Cedarapplerust 2.9 MB QmYuERUhBu8fuXRa GZSDUAYR87R675RWRYGJDH 3.45
b7RkWwDqDZKHcn8Dp FU9586ERUFJBLDIR43950
kUwpopaNMjAB3 35RTHGKVJS579048EOIHK
Cherry-healthy 14.06 MB QmPRkovGVUgYx2ue JHSGFUY5R635RWGFJSH 21
hy1g5QHwqECXpd1No VET875985WIGDSHVLUSP5T98
CXAsUehznjU5t FHDVJDOYW8R76487RITHK
Cherry-Powdery mildew 11.41 MB QmTy9g2ENwSP66D CMVNGGF653RFHHKJLLOUU 7
V2qkUP7XchCd9AQ ERWEQSCCBBJH87966453FDJ
maznM8saZbvz1xcY GHKJUYRTEESXZVFMHKJO
Corn-healthy 13.39 MB QmZkM4ymQCXKThL LQREWRR5473FCVVNGH67 14
hY6igBMPxcjwaNa 892DHGNCSK53FHSFFKJOIW
uGj6Khvnr1rfuHNh RW9345FDGERSBHYUKIOUQW
Corn-Commonrust 16.72 MB QmVCm8uXgyvnQEfvC FR5476HYHKHNCVZSA3386 12.34
bDpPxZ95XNuTyS 87UYKJNGGFTR544333DEH
ir7thRMMLfoNzFi GJUIIPKNMNBBFVDFSEW4YU
Grape-healthy 6.29 MB QmX1ohMDQqRqtvDG MNXBFYO5I73RGKLD78 5.13
PYVGZjyfVx3zuVEK 79HSJRY764934UTWJHEUFQJO
TXxKRmj6VJxc75 7GDAPOLKLKOIUSDWKN-
MND4
Grape-Esca(Black-Measles) 27.30 MB QmZw4X69QyptuNWj D564837HTYCBHGDJDUR7 37
bA3o6NWAK6x9ve 595HFYE54658THG84658HRI
eb3CcXZdPWQV6qcY 746595RHFI76HJGDTYRIR
Peach-healthy 5.74 MB QmUCANWk22uX6JC MNZCJHARU8473EIDHKSJ 4.61
Bew8SCRXXDbMfru FLJG9485029QPWADJSLKFWOR
XyfCj7YJmSJesmYz IAJFKZJFKSDJLLKPLSKJ
Peach-Bacterial spot 29.89 MB QmdWXdTy8LaTHaL BVKJSDYFIWUR23OUOQFH 43.2
wFAPe49FCBd5eii SKDLSEORIQPOWASJCDKFLKI
jaM43kMd16yj13S7 KDFIY98T4OIP4O549TIDH
Potato-healthy 3.05 MB QmenHxheRqXnXE57D XJSTF346TIUWFH7W6457VI 3.52
mL6Ncgvr3pTJ9Ed SU6WILQURW87RIFI8479WR
g9KFXW58ei5R6z IUFLSJKAS511OQALSJWP
Potato-Lateblight 16.5 MB QmbY6uwyER8WYXbz U6785GHFVDBXDSEWR5687I 23.4
C8ES9xS6iXumS JKGNBMCVXDSWQUTIOUPIK
yK2oy757EgUp2gcxR BMNVVDGTR6E4R7T8987JJ
Tomato-healthy 35.29 MB QmTozqarvDLCzaqX VDFER4557YHGDDSXZMNK 61.3
rt2895H9jBVPsiFx JOU865GGJJLDVXVGSAWUWO
12JedBc9Jy4NFA IWNVHZFQ5E7TIUGVJHIFJH
Tomato-Bacterial spot 27.5 MB Qmbzvc2Pk4qN9vR MBSJAOEUGD7847KI387HOW 38.6
13vvvuFhWiDNhWjh SKDHGVXMSLE-
TtMEB12PUcDZwGP DUR6E6R9TUWSBXKBO
FIF8EE6RWFSBVK
392 S. L. T. Vangipuram et al.
References
1. Farooq, M.S., Riaz, S., Abid, A., Abid, K., Naeem, M.A.: A survey on the role of IoT in
agriculture for the implementation of smart farming. IEEE Access 7, 156237–156271 (2019).
https://doi.org/10.1109/ACCESS.2019.2949703
2. Ivanov, I.: Cyber Security and Cyber Threats: Eagle VS “New Wars”? Academia.edu (2018).
https://www.academia.edu/38462737/CYBER SECURITY AND CYBER THREATS
EAGLE VS NEW WARS
3. Henry, R., Herzberg, A., Kate, A.: Blockchain access privacy: challenges and directions.
IEEE Secur. Priv. 16(4), 38–45 (2018). https://doi.org/10.1109/MSP.2018.3111245
4. Musharraf, M.: What is InterPlanetary File System (IPFS)? (2021). https://www.ledger.com/
academy/what-is-ipfs
5. Vangipuram, S.L.T., Mohanty, S.P., Kougianos, E., Ray, C.: G-DaM: a distributed data stor-
age with blockchain framework for management of groundwater quality data. Sensors 22,
8725 (2022). https://doi.org/10.3390/s22228725
6. Vangipuram, S.L.T., Mohanty, S.P., Kougianos, E., Ray, C.: agroString: visibility and prove-
nance through a private blockchain platform for agricultural dispense towards consumers.
Sensors 22(21), 8227 (2022). https://doi.org/10.3390/s22218227, https://www.mdpi.com/
1424-8220/22/21/8227
7. Yang, X., Li, M., Yu, H., Wang, M., Xu, D., Sun, C.: A trusted blockchain-based traceability
system for fruit and vegetable agricultural products. IEEE Access 9, 36282–36293 (2021).
https://doi.org/10.1109/ACCESS.2021.3062845
8. Yi, W., Huang, X., Yin, H., Dai, S.: Blockchain-based approach to achieve credible traceabil-
ity of agricultural product transactions. J. Phys. Conf. Ser. 1864(1), 012115 (2021). https://
doi.org/10.1088/1742-6596/1864/1/012115
9. Bapatla, A.K., Mohanty, S.P., Kougianos, E.: sFarm: a distributed ledger based remote crop
monitoring system for smart farming. In: Camarinha-Matos, L.M., Heijenk, G., Katkoori, S.,
Strous, L. (eds.) Internet of Things. Technology and Applications. IFIPIoT 2021, vol. 641.
Springer, Cham (2022). https://doi.org/10.1007/978-3-030-96466-5 2
10. Nakanishi, R., Zhang, Y., Sasabe, M., Kasahara, S.: IOTA-based access control framework
for the Internet of Things. In: 2020 2nd Conference on Blockchain Research & Applications
for Innovative Networks and Services (BRAINS), pp. 87–95 (2020). https://doi.org/10.1109/
BRAINS49436.2020.9223293
11. Ghaffaripour, S., Miri, A.: Parasite chain attack detection in the IOTA network. In: 2022
International Wireless Communications and Mobile Computing (IWCMC) (2022). https://
doi.org/10.1109/IWCMC55113.2022.9824318
12. Calicioglu, O., Flammini, A., Bracco, S., Bellù, L., Sims, R.: The future challenges of food
and agriculture: an integrated analysis of trends and solutions. Sustainability 11, 222 (2019).
https://doi.org/10.3390/su11010222
13. Wolf, S.A., Wood, S.D.: Precision farming: environmental legitimation, commodification
of information, and industrial coordination. Rural. Sociol. (1997). https://doi.org/10.1111/j.
1549-0831.1997.tb00650.x
14. Demestichas, K., Peppes, N., Alexakis, T.: Survey on security threats in agricultural IoT and
smart farming. Sensors 20, 6458 (2020). https://doi.org/10.3390/s20226458
15. West, J.: A prediction model framework for cyber-attacks to precision agriculture tech-
nologies. J. Agric. Food Inf. 19(4), 307–330 (2018). https://doi.org/10.1080/10496505.2017.
1417859
16. Devendra, C.: Climate Change Threats and Effects: Challenges for Agriculture and
Food Security. ASM Series on Climate Change (2012). https://pdf.usaid.gov/pdf docs/
PBAAK550.pdf
394 S. L. T. Vangipuram et al.
17. Elijah, O., et al.: Effect of weather condition on LoRa IoT communication technology in
a tropical region: Malaysia. IEEE Access 9, 72835–72843 (2021). https://doi.org/10.1109/
ACCESS.2021.3080317
18. An, W., Wu, D., Ci, S., Luo, H., Adamchuk, V., Xu, Z.: Chapter 25 - Agriculture cyber-
physical systems. In: Cyber-Physical Systems, Intelligent Data-Centric Systems, pp. 399–
417. Academic Press (2017). https://doi.org/10.1016/B978-0-12-803801-7.00025-0
19. Popov, S.: The Tangle. Tangle White Paper (2018). https://assets.ctfassets.net/r1dr6vzfxhev/
4i3OM9JTleiE8M6Y04Ii28/d58bc5bb71cebe4adc18fadea1a79037/Tangle White Paper
v1.4.2.pdf
20. Lundkvist, D.C., Lilic, J.: An Introduction to IPFS (2016). https://medium.com/
@ConsenSys/an-introduction-to-ipfs-9bba4860abd0
21. Foundation, I.: mam.js (2021). https://github.com/iotaledger/mam.js
22. Foundation, I.: IOTA Streams (2021). https://www.iota.org/solutions/streams
23. Palmieri, A., Vilei, A., Castanier, F., Vesco, A., Carelli, A.: Enabling secure data exchange
through the IOTA Tangle for IoT constrained devices. Sensors 22, 1384 (2022). https://doi.
org/10.3390/s22041384
24. Saarinen, M.J., Aumasson, J.P.: The BLAKE2 Cryptographic Hash and Message Authenti-
cation Code (MAC) (2015). https://www.rfc-editor.org/rfc/rfc7693.html
25. Srivastava, D.: Plant Health (2020). https://www.kaggle.com/datasets/divumarcus/plant-
health
Sana Solo: An Intelligent Approach to Measure
Soil Fertility
Abstract. Worm castings (Worm Excretion) are one the richest natural fertil-
izers on earth, making earthworms a very important and applicable soil health
indicator. According to an article published in the Polish journal of Environmen-
tal studies, the most important chemical components of worm castings are pH,
total organic carbon (TOC), total nitrogen (N), plant available phosphorus (P),
plant available potassium (K), and calcium water soluble (Ca). These chemical
components of worm castings, paired with soil temperature, humidity and elec-
tric conductivity, are all measurable values that can indicate the overall health and
fertility of soil. Furthermore, these physical-chemical properties can also be mea-
sured and analyzed to estimate worm populations in soil, making traditional man-
ual extraction techniques obsolete. The proposed project, Sana Solo, is a device
that uses machine learning to estimate worm populations based on the quantities
of the physical-chemical properties listed above. Being able to estimate earth-
worm populations in a timely manner, without the use of extraction techniques,
can be used in farms and gardens to evaluate soil fertility.
1 Introduction
Soil is considered as the upper layer of earth in which plants grow. It is a mixture of
organic remains, clay and rocky particles and has black or brown color. Soil serves as a
medium for filtration of wastes, serves as the reservoir to hold water and nutrients.
The properties of soils exhibit significant variations due to differences in geology
and climate over both space and time. Even a basic attribute like soil thickness can vary
greatly, ranging from a few centimeters to several meters. This variability is influenced
by factors such as the intensity and duration of weathering, soil deposition and erosion
events, as well as the patterns of landscape changes. Despite these differences, soils
possess a distinct structural feature that sets them apart from ordinary earth materials
and forms the foundation for their classification: a vertical arrangement of layers formed
through the combined effects of water percolation and the actions of living organisms
[3]. The thematic representation of the proposed Sana solo is represented in Fig. 1.
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 395–404, 2023.
https://doi.org/10.1007/978-3-031-45878-1_27
396 L. Rachakonda and S. Stasiewicz
The actions of flowing water, wind, ice, and gravity constantly disturb soil profiles.
These erosive processes remove soil particles from the topsoil and expose underlying
horizons to the process of weathering. As a result, there is a loss of essential compo-
nents such as humus, plant nutrients, and beneficial soil organisms. These losses are
particularly crucial for agriculture and forestry. Furthermore, the removal, movement,
and subsequent deposition of soil can have significant economic implications [2].
In the past 150 years, approximately 50% of the Earth’s topsoil has been depleted
or eroded. According to a report by the Intergovernmental Panel on Climate Change
(IPCC), soil erosion is occurring at a rate up to 100 times faster than the natural process
of soil formation when cultivated without conservation practices. This alarming finding
highlights the current imbalance in soil health. Moreover, the risk of erosion is expected
to increase further in the future due to temperature changes driven by emissions. This
can lead to detrimental consequences such as reduced agricultural production, declining
land value, and negative impacts on human health [4].
Soil erosion has implications beyond environmental concerns; it also results in sig-
nificant economic losses. The global economic losses due to soil erosion are estimated
to be approximately $8 billion. These losses stem from reduced soil fertility, decreased
crop yields, and increased water consumption [5]. Soil erosion accounts for a 2% reduc-
tion in total agricultural GDP, considering both direct losses faced by farmers and down-
stream losses affecting others [6]. Another study revealed that in Sleman, a district in
Java, soil erosion costs an average farmer 17% of their net income per hectare of agri-
cultural land [7].
Each year, soil erosion inflicts significant economic losses on the agricultural sec-
tor. In the United States, the impact amounts to approximately $44 billion, encom-
passing reduced productivity as well as the adverse effects of sedimentation and water
pollution [8]. This erosion-induced loss extends further to an estimated $100 million
in farm income annually. European countries face substantial agricultural productivity
Sana Solo: An Intelligent Approach to Measure Soil Fertility 397
losses totaling $1.38 billion per year, alongside a decrease of $171 million in their gross
domestic product [9]. Similarly, South Asia experiences a staggering annual cost of $10
billion due to soil erosion [10].
2.1 Motivation
As mentioned, most of the solutions support plant growth but the focus on soil health
is very minimal. The proposed Sana Solo project focuses more on soil health and its
fertility. Monitoring the health of soil by macro-fauna and determining their relationship
with respect to soil fertility is the main objective. There are a few devices and products
which monitor soil fertility using macro-fauna as mentioned in Table 1.
devices can be anything from everyday objects like household appliances and vehicles
to industrial equipment and infrastructure components. The IoT allows these objects
to communicate and interact with each other, as well as with users or systems, creat-
ing a vast ecosystem of interconnected devices. This network of devices enables data
gathering, automation, remote monitoring, and control, leading to increased efficiency,
convenience, and potential innovation across various industries and domains [32].
Sana Solo: An Intelligent Approach to Measure Soil Fertility 399
Concepts of IoT have been used in Agricultural fields in Sana Solo. Here, a preci-
sion farming techniques are being implemented to monitor the soil health alongside the
health of macro-fauna present in the soil as represented in Fig. 3.
Application Layer
Applications
Network Layer
Gateways
Sensing Layer
Sensors
Alongside IoT, Edge computing is performed in Sana Solo to make the system effi-
cient and robust. Edge computing refers to the decentralized processing and storage of
data at or near the source of its generation, rather than sending it to a centralized cloud
or data center for processing. In edge computing, data is processed locally on devices or
edge servers, situated closer to where the data is produced, such as IoT devices, sensors,
or edge gateways [33] as represented in Fig. 4.
400 L. Rachakonda and S. Stasiewicz
Soil conductivity
Soil Moisture
Soil Temperature
pH Microprocessor Data Analyses Data Transfer User Interface
Nitrogen
Phosphorus
Potassium
For the future research, this system will be placed in multiple test beds. These test
beds will be with various configurations - with and without castings to monitor the
growth of the plants. The system will also be configured in a way that it can commu-
nicate with the devices in the network to share and transfer information to the farmer
using an user interface.
Acknowledgements. This version of the project is funded by the College of Arts and Sciences
in the University of North Carolina Wilmington.
References
1. Rachakonda, L.: ETS: a smart and enhanced topsoil health monitoring and control system at
edge using IoT. In 2022 IEEE International Symposium on Smart Electronic Systems (iSES),
Warangal, India, pp. 689-693 (2022). https://doi.org/10.1109/iSES54909.2022.00153.
2. Gregory, A.S., et al.: A review of the impacts of degradation threats on soil properties in the
UK. Soil Use Manag. 31(Suppl. 1), 1–15 (2015). Epub 12 October 2015. PMID: 27667890;
PMCID: PMC5014291. https://doi.org/10.1111/sum.12212
3. Liu, H., Li, B., Ren, T.: Soil profile characteristics of high-productivity alluvial cambisols in
the North China Plain. J. Integr. Agric. 14(4), 765–773 (2015). ISSN 2095-3119, https://doi.
org/10.1016/S2095-3119(14)60789-9
4. Dede, S., Thomas, W.: The Causes and Effects of Soil Erosion, and How to Prevent It. World
Resources Institute, February 2000. https://www.wri.org/insights/causes-and-effects-soil-
erosion-and-how-prevent-it#:~:text=A%20report%20from%20the%20Intergovernmental,
times%20quicker%20than%20it’s%20forming
5. Martina, S., et al.: A linkage between the biophysical and the economic: assessing the global
market impacts of soil erosion. Land Use Policy 86, 299–312 (2019). ISSN 0264–8377,
https://doi.org/10.1016/j.landusepol.2019.05.014
6. Strutt, A.: Trade Liberalisation and Soil Degradation in Indonesia. Indonesia in a Reform-
ing World Economy: Effects on Agriculture, Trade and the Environment, edited by Kym
Anderson et al., pp. 40–60. University of Adelaide Press. JSTOR (2009). http://www.jstor.
org/stable/10.20851/j.ctt1sq5w4j.10 Accessed 21 June 2023
7. Möller, A., Ranke, U.: Estimation of the on-farm-costs of soil erosion in Sleman, Indonesia.
WIT Trans. Ecol. Environ. 89, 43–52 (2006)
8. Shepard, K.C.: Oklahoma Farm Report. https://www.oklahomafarmreport.com/okfr/2023/
04/. Accessed May 2023
9. Panagos, P., Standardi, G., Borrelli, P., Lugato, E., Montanarella, L., Bosello, F.: Cost of
agricultural productivity loss due to soil erosion in the European Union: from direct cost
evaluation approaches to the use of macroeconomic models. Land Degrad. Dev. 29, 471–484
(2018). https://doi.org/10.1002/ldr.2879
10. Third World Network Berhad. https://twn.my/title/land-ch.htm#:~:text=Its%20shocking
%20conclusion%20was%20that,losses%20resulting%20from%20land%20degradation
11. Kumar, P., Bhagat, K., Lata, K., Jhingran, S.: Crop recommendation using machine learning
algorithms. In: International Conference on Disruptive Technologies (ICDT), Greater Noida,
India, pp. 100–103 (2023). https://doi.org/10.1109/ICDT57929.2023.10151325
12. Saha, P., Kumar, V., Kathuria, S., Gehlot, A., Pachouri, V., Duggal, A.S.: Precision agricul-
ture using Internet of Things and Wireless Sensor Networks. In: International Conference on
Disruptive Technologies (ICDT), Greater Noida, India, pp. 519–522 (2023). https://doi.org/
10.1109/ICDT57929.2023.10150678
Sana Solo: An Intelligent Approach to Measure Soil Fertility 403
13. Prabha, C., Pathak, A.: Enabling technologies in smart agriculture: a way forward towards
future fields. In: International Conference on Advancement in Computation and Computer
Technologies (InCACCT), Gharuan, India, pp. 821–826 (2023). https://doi.org/10.1109/
InCACCT57535.2023.10141722
14. Kumar, A., Savaridassan, P.: Monitoring and accelerating plant growth using IoT and Hydro-
ponics. In: International Conference on Computer Communication and Informatics (ICCCI),
Coimbatore, India, pp. 1–6 (2023). https://doi.org/10.1109/ICCCI56745.2023.10128383
15. Fleming, K., Gardner, A., Nagel, P., Miao, Y., Mizuta, K.: Hyperspectral sensing for soil
health. In: IEEE Conference on Technologies for Sustainability (SusTech), Portland, OR,
USA, pp. 1–5 (2023). https://doi.org/10.1109/SusTech57309.2023.10129629
16. Irene Monica, N., Pooja, S.R., Rithiga, S., Madhumathi, R.: Soil NPK prediction using
enhanced genetic algorithm. In: 9th International Conference on Advanced Computing and
Communication Systems (ICACCS), Coimbatore, India, pp. 2014–2018 (2023). https://doi.
org/10.1109/ICACCS57279.2023.10113121
17. Anuradha, B., Pradeep, R., Ahino, E., Dhanabal, A., Gokul, R.J., Lingeshwaran, S.: Vertical
farming algorithm using hydroponics for smart agriculture. In: International Conference on
Intelligent Systems for Communication, IoT and Security (ICISCoIS), Coimbatore, India,
pp. 432–437 (2023). https://doi.org/10.1109/ICISCoIS56541.2023.10100527
18. Rohini, V., Meghana, K., Sowmya, R.K., Krishna, K.S., Srikrishna, B.: Application of SMAP
images in predicting Crops by using Decision Tree and Random Forest. In: International
Conference on Artificial Intelligence and Knowledge Discovery in Concurrent Engineering
(ICECONF), Chennai, India, pp. 1–6 (2023). https://doi.org/10.1109/ICECONF57129.2023.
10083570
19. Verma, M., Kumar, A., Garg, M., Juneja, S.: Environment quality assessment web appli-
cation. In: International Conference on Artificial Intelligence and Smart Communication
(AISC), Greater Noida, India, pp. 1339–1342 (2023). https://doi.org/10.1109/AISC56616.
2023.10085252
20. Bhowmik, A., Sannigrahi, M., Dutta, P.K., Bandyopadhyay, S.: Using edge computing
framework with the Internet of Things for intelligent vertical gardening. In: 1st International
Conference on Advanced Innovations in Smart Cities (ICAISC), Jeddah, Saudi Arabia, pp.
1–6 (2023). https://doi.org/10.1109/ICAISC56366.2023.10085507
21. Nehra, V., Sharma, M., Sharma, V.: IoT based smart plant monitoring system. In: 13th
International Conference on Cloud Computing, Data Science and Engineering (Confluence),
Noida, India, pp. 60–65 (2023). https://doi.org/10.1109/Confluence56041.2023.10048792
22. Pallavi, C.V., Usha, S.: IoT based site specific nutrient management system for soil health
monitoring. In: International Conference on Smart and Sustainable Technologies in Energy
and Power Sectors (SSTEPS), Mahendragarh, India, pp. 166–170 (2022). https://doi.org/10.
1109/SSTEPS57475.2022.00050
23. Sui, X., Lin, C., Zhou, S.: Spatial decision analysis on soil erosion control measures research
based on GIS: taking Changting country as an example. In: Third World Congress on Soft-
ware Engineering, Wuhan, China, pp. 119–122 (2012). https://doi.org/10.1109/WCSE.2012.
29
24. Naik, T.R., et al.: Environmental testing methodology for real-time soil health monitoring
system. In: IEEE Applied Sensing Conference (APSCON), Bengaluru, India, pp. 1–3 (2023).
https://doi.org/10.1109/APSCON56343.2023.10101082
25. Microbiometer. https://microbiometer.com/. Accessed March 2023
26. Weyers, S.L., Schomberg, H.H., Hendrix, P.F., Spokas, K.A., Endale, D.M.: Construction
of an electrical device for sampling earthworm populations in the field. Appl. Eng. Agric.
24(3), 391–397 (2008). https://doi.org/10.13031/2013.24492
27. Kempson Extractor. https://www.ecotech.de/en/product/kempson_extractor_1. Accessed
May 2023
404 L. Rachakonda and S. Stasiewicz
28. Ismayilov, A., Feyziyev, F., Elton, M., Maharram, B.: Soil organic carbon prediction by Vis-
NIR spectroscopy: case study the Kur-Aras Plain, Azerbaijan. Commun. Soil Sci. Plant Anal.
51(6), 726–734 (2020)
29. Tauro, T.P., Mtambanengwe, F., Mpepereki, S., Mapfumo, P.: Soil macrofauna response
to integrated soil fertility management under maize monocropping in Zimbabwe. Heliyon
7(12), e08567 (2021). PMID: 34917826; PMCID: PMC8666646. https://doi.org/10.1016/j.
heliyon.2021.e08567
30. Adomako, M.O., Xue, W., Roiloa, S., Zhang, Q., Du, D.L., Yu, F.H.: Earthworms modulate
impacts of soil heterogeneity on plant growth at different spatial scales. Front Plant Sci.
23(12), 735495 (2021). PMID: 35003149; PMCID: PMC8732864. https://doi.org/10.3389/
fpls.2021.735495
31. Huang, M., et al.: Rice yield and the fate of fertilizer nitrogen as affected by addition of earth-
worm casts collected from oilseed rape fields: a pot experiment. PLoS ONE 11, e0167152
(2021). https://doi.org/10.1371/journal.pone.0167152
32. Rachakonda, L., Bapatla, A.K., Mohanty, S.P., et al.: BACTmobile: a smart blood alco-
hol concentration tracking mechanism for smart vehicles in healthcare CPS framework. SN
Comput. Sci. 3, 236 (2022). https://doi.org/10.1007/s42979-022-01142-9
33. Rachakonda, L.: Agri-Aid: an automated and continuous farmer health monitoring system
using IoMT. In: Camarinha-Matos, L.M., Ribeiro, L., Strous, L. (eds.) Internet of Things.
IoT Through a Multi-disciplinary Perspective. IFIPIoT 2022. IFIP Advances in Information
and Communication Technology, vol. 665. Springer, Cham (2022). https://doi.org/10.1007/
978-3-031-18872-54
Smart Agriculture – Demystified
1 Introduction
Throughout history, agriculture has been crucial to human survival, and it continues
to be the backbone of the economies of many countries today. Agriculture’s signifi-
cance has grown alongside the global population and economy. It now encompasses
not just farming but also livestock, poultry, forestry, fisheries, food supply chain, and so
on. Unprecedented population growth, climate change, depletion of natural resources,
urbanization, over-farming, and deforestation are the crucial factors that are affecting
crop yield, disrupting the food supply chain, and threatening human civilization with
food scarcity and high prices.
The food and agricultural industries embrace technological advancements, giving
birth to “Agriculture 4.0,” a green and smart revolution. Conventional agriculture is
transforming into “smart agriculture” and becoming more productive and sustainable
by optimizing human labor and natural resources. As a result, crop yield and food pro-
duction are increasing. Figure 1 shows the various areas of “smart agriculture.” In this
article, we highlighted the key factors of “smart agriculture.”
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 405–411, 2023.
https://doi.org/10.1007/978-3-031-45878-1_28
406 A. Mitra et al.
“Smart agriculture” differs from “precision agriculture” in that it does not priori-
tize metric precision. Instead, “smart farming” relies on data collection and analysis
enabled by modern computing systems to improve the predictability and efficiency of
agricultural processes. Both “smart agriculture” and “precision agriculture” together are
the two branches of “digital farming” with different focuses. The evolution of “digital
farming” also defines the fourth stage of the agricultural revolution, “Agriculture 4.0.”
In this context, a new hybrid system, Cyber-Physical System (CPS), which origi-
nated from the IoT deployment in physical systems, is gaining popularity. CPSs connect
physical things and infrastructure to the internet as well as to each other by integrating
sensing, processing, and networking into these physical objects and infrastructure. The
National Science Foundation (NSF) is a pioneer in fostering advancements in the foun-
dational knowledge and technologies necessary to bring cyber-physical systems into
existence [4]. Figure 2 shows the three parts of an A-CPS: physical systems, cyber sys-
tems, and network fabric. CPSs enable precision and improve functionality, scalability,
resilience, safety, security, and usability over simple embedded systems [5]. “Agricul-
ture Cyber-Physical Systems (ACPSs)” can collect meteorological, soil, and crop data
to improve agricultural management. ACPSs may monitor water, humidity, and plant
health and employ actuators and infrastructure to control temperature and humidity.
Another important and relevant term is “climate smart agriculture (CSA)” [6]. As
climate change has already been started, efforts to overcome the adverse effects of cli-
mate change are being included in agriculture for sustainability. Smart agriculture has
started to transform to climate-smart agriculture to fight against the aftermath of climate
change.
Smart Agriculture – Demystified 407
1 2 3 O2 O2 O2
CO2
CO2
CH4
Increase Yield
by adopting
Smart Adapt and Make Robust Solutions Greenhouse gas
Agricultural against Climate Change emission reduction
Technologies
Various efforts are being proposed to limit the causes of climate change; e.g., the
AgSTAR program [9] has been introduced to help the agricultural industry cut down on
methane emissions from livestock manure. Producers concerned with soil health should
reduce soil disturbance while increasing cover crops, biodiversity, and the number of
plants with roots in the ground. These practices work together to lessen the impact on
the environment by decreasing emissions and increasing carbon sequestration. They
also benefit the environment by decreasing soil erosion, decreasing the need for costly
inputs like fertilizer, increasing water infiltration, boosting nutrient cycling, and con-
structing more resilient soils over time [10].
All the efforts for sustainable agriculture are possible because of the rapid growth in
technologies, especially in the hardware and IC industries, Graphic Processing Units
(GPU) and Tensor Processing Units (TPU), computing platforms, and last but not
least, Information and Communication Technology (ICT). Industries in different sec-
tors are eagerly embracing digital, smart, green, and sustainable ecosystems to meet
the challenges of climate change. Because of this, the relationship between “man” and
“machine” is being rethought. Changes are happening in the agricultural sector. “Agri-
culture 5.0” [11] is knocking at the door.
Artificial Intelligence (AI), Machine Learning (ML), and IoT are playing a major
role here, along with UAVs and robotics, as they provide decision-making automation.
Smart Agriculture – Demystified 409
Remote sensing through satellite monitoring and cloud computing are two established
advanced technologies used for data gathering and decision-making. Different new con-
cepts, like edge computing in agriculture and Software as a Service (SaaS), are emerg-
ing. Distributed ledger technology is showing promise and can play an important role
in the agricultural industry because of its ability to store immutable data.
Farms are being equipped with sensors and actuators. These IoT sensors and actua-
tors generate huge amounts of data, or big data,” which demands a new stream of data
analysis, “big data analysis, in data science.
Farmers can now monitor how far along their crops are in their distinct growth
cycles thanks to drone technology. In addition, growers can use UAVs to provide treat-
ments for infected plants. The concept of urban farming, like hydroponics, aeroponics,
aquaponics, vertical farming, smart greenhouses, and livestock monitoring, is revolu-
tionizing today’s agriculture and ensuring sustainable agriculture.
– As the field size varies from small-holder farms to large farms, scalability of solu-
tions is needed. It optimizes all the efforts. Along with scalability, the reliability of
the solutions will optimize the number of devices. A smaller number of redundant
devices that replace faulty devices will minimize the cost.
– To modernize agriculture, one of the biggest challenges is the communication gap
between the research community and stakeholder farmers. The issues the farmers
need to address do not always reach the researchers, and the agricultural industry
cannot fully utilize the benefits of modern technologies.
7 Conclusions
Today, we live in a world where we cannot deny irreversible climate change. Techno-
logical progress and the rapid development of ICT have already boosted the digitiza-
tion and modernization of agriculture, which results in an increase in agricultural pro-
ductivity and yields, a decrease in ecological footprints, improved water conservation,
increased climate smart efforts, and a decrease in operational costs. Overall, agriculture
advances in quality and quantity. However, more climate-smart efforts are needed. In the
United States, $19.5 billion has been sanctioned via the Inflation Reduction Act to sup-
port climate change alleviation efforts from 2023 to 2027 [12]. Common Agricultural
Policy 2023-2027 of the European Commission aims to form a sustainable, resilient,
and contemporary European agriculture economy. It also has a focus on efforts for cli-
mate change mitigation [13]. In 2011, India launched National Innovations in Climate
Resilient Agriculture (NICRA) with $42.7 million to make Indian agriculture-crops,
livestock, and fisheries-more resilient to climate change and unpredictability.
References
1. Farooq, M.U., Waseem, M., Mazhar, S., Khairi, A., Kamal, T.:A review on internet of things
(Iot). Int. J. Comput. Appl. 113(1), 1–7 (2015)
Smart Agriculture – Demystified 411
2. Winston, P.H.: Artificial intelligence. Addison-Wesley Longman Publishing Co., Inc (1984)
3. Mitra, A., et al.: Everything You Wanted To Know About Smart Agriculture. CoRR,
abs/2201.04754 (2022)
4. Cyber-Physical Systems: Enabling a Smart and Connected World. https://www.nsf.gov/
news/special reports/cyber-physical/. Accessed May 25 2023
5. An, W., Wu, D., Ci, S., Luo, H., Adamchuk, V., Xu, Z.: Chapter 25 - agriculture cyber-
physical systems. In: Song, H., Rawat, D.B., Jeschke, S., Brecher, C. (eds.) Cyber-Physical
Systems. Intelligent Data-Centric Systems, pp. 399–417. Academic Press, Boston (2017)
6. Lipper, L., et al.: Climate-smart agriculture for food security. Nature Climate Change 4(12),
1068–1072 (2014)
7. Mitra, A., Mohanty, S.P., Kougianos, E., et al.: aGROdet: A Novel Framework for Plant
Disease Detection and Leaf Damage Estimation. In: Proceedings of the 5th IFIP International
Internet of Things Conference (IFIP-IoT), pp. 3–22 (2022)
8. The Effects of Climate Change. https://climate.nasa.gov/effects/. Accessed May 25 2023
9. AgSTAR Accomplishments. https://www.epa.gov/agstar/agstar-accomplishments Accessed
June 22 2023
10. NRCS Climate-Smart Mitigation Activities. https://www.nrcs.usda.gov/conservation-basics/
natural-resource-concerns/climate/climate-smart-mitigation-activities. Accessed June 23
2023
11. Saiz-Rubio, V., Rovira-Más, F.: From smart farming towards agriculture 5.0: A review on
crop data management. Agronomy 10(2), 207 (2020)
12. Climate-Smart Agriculture and Forestry. https://www.farmers.gov/conservation/climate-
smart Accessed June 23 2023
13. Key policy objectives of the CAP 2023–27. https://agriculture.ec.europa.eu/common-
agricultural-policy/cap-overview/cap-2023-27/key-policy-objectives-cap-2023-27 en
Accessed June 2023
Student Research Forum (SRF)
WeedOut: An Autonomous Weed Sprayer
in Smart Agriculture Framework Using
Semi-Supervised Non-CNN Annotation
1 Introduction
Agriculture is the primary source of food for all the human beings across the world. Var-
ious factors like rapid growth in human population, reduction of farmland, depletion of
natural resources and advances in Internet-of-Agro-Things (IoAT) [1] paved path to
new paradigm in agriculture named “Smart Agriculture” [2] to automate agriculture
routines with help of Artificial Intelligence (AI). Weeds are unwanted plants that grow
along with the crop being cultivated and compete with primary crops for resources like
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 415–423, 2023.
https://doi.org/10.1007/978-3-031-45878-1_29
416 K. K. Kethineni et al.
sunlight, water, nutrients, space. Weeds can also serve as a habitat for pests and diseases
that can infect crops, provide shade which promotes the growth of fungi. These factors
present weed management as a significant part of cultivation in agriculture [3]. Man-
ually spraying herbicide to suppress weeds is easier when farm area is small. In cases
of large-scale farming where area of farmland ranges from tens of acres to hundreds of
acres manual weeding needs lot of physical labor and inappropriate usage of herbicide
can have several negative effects on environment [4]. To reduce manpower and use right
amount of herbicide there ought to be a system which can scout through farm to identify
weeds and spray at locations of weeds so that weed growth is suppressed without affect-
ing primary crop. Such systems where multiple IoAT devices and Al technologies like
Computer Vision are deployed in agriculture infrastructure to automate a specific task
are referred to as Agriculture Cyber-Physical System (A-CPS). Current article Weed-
Out is a weed management A-CPS that follows a semi-supervised approach as depicted
in Fig. 1 to identify weeds and suppress their growth.
The Rest of the paper is organized as follows: Sect. 2 presents novel contributions of
this article followed by discussion on related works in Sect. 3. Section 4 demonstrates
the working of proposed solution while experimental results are discussed in Sect. 5.
Section 6 concludes the article.
cases [5]. In such cases solutions that consider other properties like shape of the crop,
area occupied by the crop, patterns in sowing the crop can help us distinguish between
primary crop and weed.
Proposed WeedOut tries to differentiate crops by their shape (profile plots) and clus-
ters similar crops (similar shapes) together. This approach makes proposed method un
effected by differences in appearances/features of the crop due to geographical and
aging factors. Knowledge of farmer is utilized in identifying the clusters that represent
primary crop to classify crops as weeds and primary crop in semi-supervised fashion.
There have been multiple solutions that do not use CNN for identifying weeds in farm-
land like [6] which detect rows of plantation, row orientation to know crop margins and
label crops outside of crop margins and with lower NDVI as weeds. Whereas, in [7]
crop rows are detected by help of depth data and crops lying between crop margins are
clustered to 2 clusters by their geometric properties and KNN algorithm. Assuming the
number of weeds is greater than primary crop, the smaller cluster is marked as primary
crop. In [8] authors proposed a method where crop lines are derived and super pixels
(obtained by SLIC) that are in contact with crop lines will be classified as crops, super
pixels that are not in contact with crop lines are classified by comparing with neighbors.
In contrast to the above solutions that rely on practice of cultivating in rows, some
solutions classify crops by the area they occupy. Authors of [9, 10] proposed methods
where area covered (number of pixel occupied) by individual crop is computed and the
one whose area is below a threshold is classified as weed while the one whose area
is above the threshold is classified as primary crop by assuming individual primary
crop occupies more area than individual weed crop. But in [11, 12] the classification
is performed the other way assuming individual primary crop occupies less area than
individual weed crop. Article [13] proposes use of Active Shape Models (ASM) for
418 K. K. Kethineni et al.
classification, which calculates shapes of crops present in the image and compares them
to shapes of primary crops in memory (training data) to know if its a primary crop or
weed. A brief summary of these works are presented in Table 1.
Unlike the above approaches, current approach makes no assumptions on pattern
in cultivation or differences in area occupied by individual crops. Instead, WeedOut
utilizes shapes of the crops similar to [13] to cluster similar crops and farmers inputs to
classify them.
Algorithm of crop detection proceeds by processing one image at a time from the set
images in the sequence they have been captured. In-order to classify crops, first task is
separation of crops from soil by eliminating background. So, image is transformed into
HUE color space which represents colors based on hue, saturation and value parameters.
Thresholding is performed on image with prior defined limits for green color to detect
objects that are in green color (crops) [15]. Image is then resized to 250×250 for ease
of computing and converted to binary image as in Fig. 3.
In order to classify crops in the image as primary crops and weeds, individual crops
in the image have to be identified and labeled uniquely. Two-pass Connected Compo-
nent Labeling is a Computer Vision algorithm, which essentially identifies and uniquely
labels all the objects in a image by just passing over the image twice. When ever a binary
image is presented, the algorithm starts to process each pixel of the image column after
column in each row. In the first pass, whenever it reads a pixel that is bright it looks
for its neighbor pixels that are bright. If there are any neighbors available, highest of
their labels would be assigned to the current pixel. If no neighbors are found a unique
label is assigned to the pixel and equivalence between neighboring labels is saved. This
process of labeling continues till all the pixels in current image are assigned a label.
Second pass identifies various labels assigned to a single object and replaces them with
label that is unique to every object as represented in Fig. 3.
Every crop essentially differs with others in properties like length of leaves, width of
leaves, number of leaves, orientation of leaves. All these features effect how the whole
crop looks and how width of crop changes with its length from tip to tip. A plot describ-
ing variation in width of a plant with length is termed as Profile Plot, Fig. 4 shows profile
plots of two crops demonstrating how profile plots can help differentiating crops.
After computing profile plots for all crops identified in the image, they are extrap-
olated to length of 250 for ease of visualization. All these profile plots are compared
with one another by Dynamic Time Warping (DTW). Dynamic Time Warping of two
signals is finding best alignment between them by stretching and compressing one of
them along time axis while distance between corresponding points is being minimized
as in Fig. 4. DTW distance is the minimum distance required to align the signals. In
simple terms, signals those are highly similar would have low DTW distance and thus,
can be used as similarity measure. This helps in finding pairs of crops that are similar
and all pairs that have a common element are merged to form clusters in iteration till
no clusters have a common element. DTW is performed between identified crops for
couple images at the initial stage to detect all kinds of crops present. Later on instead
of performing DTW between crops identified in a image, DTW is performed between
each identified crop and identified crop types to group with similar ones.
5 Experimental Results
Proposed solution was implemented with python and a Computer Vision library
OpenCV on a data set from kaggle [16]. To create an image of a farmland multiple
images were combined and results of one of such image are discussed below. Thresh-
olding and Connected Component Algorithm are performed on inputted image to iden-
tify individual crops, profile plots are plotted for 8 individual crops identified shown in
different shade of gray in Fig. 5. DTW is then performed to detect and group similar
crops to 3 clusters in Fig. 6.
A sample from each cluster is now presented with percentage of contribution of that
cluster to the total vegetation to farmer as in Fig. 7. In this experiment farmer selected
cluster 1 as his primary crop. All other clusters except cluster 1 are marked weeds and
colored red while primary crops are colored green. Final results are presented in Fig. 7
422 K. K. Kethineni et al.
along with computed weed pressure. and can be used as an input to autonomous weed
sprayer to spray herbicide at weed location.
The same algorithm is fed with 20 of such images to simulate a small sized farmland
and calculate its performance metrics. The proposed clustering method showed an accu-
racy of 93% while F1 score for primary crops and weeds were 0.80, 0.95 respectively
indicating that the proposed method was particularly effective at identifying weeds.
6 Conclusion
Current article proposed a novel methodology for an A-CPS delegated with weed man-
agement utilizing shape of crops and domain knowledge of farmer to detect weeds in the
farmland instead of CNN methods which depend on visual features of crops. WeedOut
identifies various crops in the image using Connected Component Labeling Algorithm
which checks if any pixel has a directed connection or connected path to other pixel of
a object to decide if it belongs to same object or not. This assumption leads to two crops
with some overlap be considered as single crop. which means this solution only works
for non-overlapping crops in farmland. Proposed method classifies crops by their shape
which poses chances of misclassification if two crops have similar profile. Methods to
distinguish crops even in cases of overlap with help of edge detection and considering
some additional geometrical features that help in more accurate identification can be
explored as future works for the proposed solution.
WeedOut: An Autonomous Weed Sprayer in Smart Agriculture Framework 423
References
1. Mohanty, S.P.: Internet-of-agro-things (IoAT) makes smart agriculture. IEEE Consum. Elec-
tron. Mag. 10(4), 4–5 (2021)
2. Mitra, A., et al.: Everything you wanted to know about smart agriculture (2022)
3. Ekwealor, K.U., Echereme, C.B., Ofobeze, T.N., Okereke, C.N.: Economic importance of
weeds: a review. Asian J. Plant Sci. 3, 1–11 (2019)
4. Kudsk, P., Streibig, J.C.: Herbicides-a two-edged sword. Weed Res. 43(2), 90–102 (2003)
5. Slaughter, D.C., Giles, D.K., Downey, D.: Autonomous robotic weed control systems: a
review. Comput. Electron. Agric. 61(1), 63–78 (2008)
6. Louargant, M., et al.: Unsupervised classification algorithm for early weed detection in row-
crops by combining spatial and spectral information. Remote Sens. 10(5), 761 2018
7. Ota, K., Kasahara, J.L.Y., Yamashita, A., Asama, H.: Weed and crop detection by combining
crop row detection and k-means clustering in weed infested agricultural fields. In: Proceed-
ings IEEE/SICE International Symposium on System Integration (SII), pp. 985–990 (2022)
8. Bah, M.D., Hafiane, A., Canals, R.: Weeds detection in UAV imagery using SLIC and the
hough transform. In: Proceedings Seventh International Conference on Image Processing
Theory, Tools and Applications (IPTA), pp. 1–6 (2017)
9. Rani, K.A.A., Supriya, P., Sarath, T.V.: Computer vision based segregation of carrot and
curry leaf plants with weed identification in carrot field. In: Proceedings International Con-
ference on Computing Methodologies and Communication (ICCMC), pp. 185–188 (2017)
10. Tejeda, A.J.I., Castro, R.C.: Algorithm of weed detection in crops by computational vision.
In: Proceedings International Conference on Electronics, Communications and Computers
(CONIELECOMP), pp. 124–128 (2019)
11. Aravind, R., Daman, M., Kariyappa, B.S.: Design and development of automatic weed detec-
tion and smart herbicide sprayer robot. In: Proceedings of IEEE Recent Advances in Intelli-
gent Computational Systems (RAICS), pp. 257–261 (2045)
12. Siddiqi, M.H., Ahmad, I., Sulaiman, S.B.: Weed recognition based on erosion and dilation
segmentation algorithm. In: Proceedings of International Conference on Education Technol-
ogy and Computer, pp. 224–228 (2009)
13. Persson, M., Åstrand, B.: Classification of crops and weeds extracted by active shape models.
Biosys. Eng. 100(4), 484–497 (2008)
14. Mitra, A., Singhal, A., Mohanty, S.P., Kougianos, E., Ray, C.: eCrop: a novel framework for
automatic crop damage estimation in smart agriculture. SN Comput. Sci. 3(4), 319 (2022)
15. Mitra, A., Mohanty, S.P., Kougianos, E.: aGROdet: a novel framework for plant disease
detection and leaf damage estimation. In: Proceedings of the 5th IFIP International Internet
of Things Conference (IFIP-IoT), pp. 3–22 (2022)
16. Dabhi, R., Makwana, D.: Crop and weed detection data with bounding boxes (2020). https://
www.kaggle.com/datasets/ravirajsinh45/crop-and-weed-detection-data-with-bounding-
boxes. Accessed 20 Jan 2023
ALBA: Novel Anomaly Location-Based
Authentication in IoMT Environment Using
Unsupervised ML
1 Introduction
The growth of IoT embedded systems and biosensors, has introduced the IoMT as a
branch that integrates medical devices, applications, and networks to enhance the effi-
ciency of healthcare system [1, 2]. The rapid advancements in mobile technology have
enabled smartphones to become an important component of the IoMT network and
a source of information due to the increasing complexity of software and hardware
components and multiple interfaces in medical devices [3]. However, smartphones also
introduce new security challenges due to the sensitive nature of medical data that they
collect, making them a valuable target for cybersecurity threats [4, 5]. Therefore, ensur-
ing the security of IoMT is crucial to mitigate risks and enhance the sustainability of
healthcare.
Artificial Intelligence (AI) technologies have been advanced significantly and can
be used to monitor and predict the behavior of entities within an IoT environment. How-
ever, data quality is crucial in machine learning (ML) for achieving accurate results.
c IFIP International Federation for Information Processing 2023
Published by Springer Nature Switzerland AG 2023
D. Puthal et al. (Eds.): IFIPIoT 2023, IFIP AICT 683, pp. 424–432, 2023.
https://doi.org/10.1007/978-3-031-45878-1_30
ALBA: Anomaly Location-Based Authentication in IoMT 425
The paper is organized as follows: Section 2 presents the literature review, while
Sect. 3 introduces the novel contributions of this paper. The proposed method is pre-
sented in Sect. 4, and data preprocessing is detailed in Sect. 5. The ML model used in
our method is described in Sect. 6. Experimental results are provided in Sect. 7, and the
conclusion with future work are in Sect. 8.
Existing research on behavioral authentication has provided valuable insights using var-
ious techniques and sensors, such as Keystroke Dynamics (KD) [6, 7], Touch Gestures
(TG) [8, 9], and Gait Behavior [10, 11]. However, some studies may not have fully met
expectations in terms of accuracy and stability of authentication data, nor considered
the holistic security needs covering diverse environmental, device, and user conditions.
This limits their accuracy, suitability and effectiveness for the IoT devices, especially
smartphones.
426 F. J. Alruwaili et al.
The limitations of these techniques are mainly due to internal and external factors,
such as the variety of devices, where smartphones have touchscreens or keyboards with
different shapes, layouts, and sizes [12]. Also, the specific language in which tech-
nique is applied affects the tested interval time between touches, where the user may be
unfamiliar with some vocabularies. Additionally, there are external factors that affect
these techniques, such as, environment, clothing, sickness, injuries, fatigue, emotional
or mental status, and smartphone position. These limitations make the extracted fea-
tures insufficient for behavioral-based authentication. Based on the above discussions,
we conclude that existing approaches to behavioral authentication in IoMT are still
lacking and have limitations. Therefore, ALBA method aims to address these limita-
tions and improve authentication data stability to be more accurate and usable in IoT
devices.
3 Novel Contributions
3.1 Problem Addressed and Proposed Solution
suitable for device technologies and more comprehensive in terms of security consid-
erations in an IoMT environment. To the best of our knowledge, we have proposed the
first behavioral authentication method integrating GPS and unsupervised ML technolo-
gies for IoMT security.
When user credentials are validated, authentication is based on comparing the current
location with historical locations stored in the database within a given time frame. If the
behavior of current location matches the behavior of the historical locations, it will be
considered normal location. Otherwise is anomaly.
Figure 2 illustrates the proposed authentication method, where the mobile device
collects data from different resources (user, medical devices, and GPS satellites) and
transmits it to the medical server’s IoMT system for verification. Specifically, when
users connects their medical devices to the server, the IoMT system prompts the user
for username and password. The GPS data are then verified and analyzed using ML
algorithms to detect whether the current location is anomalous or not. If the location is
normal, the verification process is successful, and user’s medical devices will be con-
nected to the medical server, allowing secure transmission of medical data for doctor
diagnosis. The user also will be able to access health record. Historical location data
is pre-processed before being stored to reduce computational time and resource con-
sumption, which positively impacts power consumption during future authentication
processes. The result of data analysis determines whether to continue monitoring user’s
current location or take appropriate action in case of any deviation from the expected
behavior, such as limiting system functionality until additional authentication is pro-
vided or sending alerts through other channels.
428 F. J. Alruwaili et al.
5 Data Pre-processing
In our experiment, location features, such as latitude, longitude, and timestamp fea-
tures are selected from the original dataset, while others are redundant. Day and hour
features were extracted from timestamps to improve anomaly detection by identifying
deviations during specific times or days. In addtion, latitude and longitude features are
standardized to be on comparable scales. These features are then combined using PCA
to create a single location feature, enabling easier analysis and visualization of location
patterns and anomalies.
ALBA: Anomaly Location-Based Authentication in IoMT 429
7 Experimental Results
In our experiment, the iForest model was trained on location data in the real-word
dataset. The results showed that the model successfully calculated anomaly scores as
shown in Fig. 4 (a), identifying 11 anomaly scores represented in red dots as negative
values, while positive vlaues (blue dots) represent the normal scores. The farther from
0, the more anomalous (or normal) a location is. The model’s prediction is shown in
Fig. 4 (b) as a binary series of -1 for anomalies and 1 for normal points.
The model’s sensitivity was controlled by utilizing a anomaly threshold of –0.05
after conducting multiple experiments and evaluating performance, leading the model
to identify 6 real anomalous locations with a significant deviation indicated by dark
red in Fig. 4 (c). For contextual anomalies, they occurred during certain hours or days,
which makes them different from other locations. However, the model considered them
as normal based on the determined threshold as they have slight deviation. For more
insight ont he spatial distribution of anomalous and normal locations, they were pro-
jected on the map depicted in Fig. 4 (d).
430 F. J. Alruwaili et al.
The model was trained on the public dataset, and all anomaly scores were calculated
successfully as shown in Fig. 5 (a). There were 121 anomalous scores (red dots), where
58 of them were above the utilized threshold of –0.015, representing real anomalous
locations as shown in Fig. 5 (b) represented by dark red dots.
We can notice that some of dark red locations have a significant deviation from
the behavior of other locations, while the red locations have a slight deviation, and
which cosidered normal based on the determined threshold. However, the public dataset
locations are shown in a clear daily pattern with some significant deviation, especially
on the right side. Most of the deviations occur at the beginning or end of the week or
during the weekends, which is a reasonable pattern.
ALBA: Anomaly Location-Based Authentication in IoMT 431
In the virtual dataset, there were 7 anomaly scores as depicted in Fig. 6 (a). Based on
the calculated anomaly scores, the model successfuly identified 7 anomalous locations
as illustarated in Fig. 6 (b)
The model effectively detected all 5 known anomalous locations, which are repre-
sented by dark red dots. However, there were 2 false positives identified by the model,
which were due to the adjustment of the model’s parameters to optimize accuracy for
the virtual dataset. Despite this, by adjusting the threshold value to 0.018, the model
effectively reclassified these false positives as normal locations (red dots), demonstrat-
ing its ability to adapt and perform well on the given dataset.
References
1. Mohanty, S.P., Choppali, U., Kougianos, E.: Everything you wanted to know about smart
cities: the internet of things is the backbone. IEEE Consum. Electron. Mag. 5(3), 60–70
(2016)
2. Ghubaish, A., Salman, T., Zolanvari, M., Unal, D., Al-Ali, A., Jain, R.: Recent advances
in the internet-of-medical-things (IoMT) systems security. IEEE Internet Things J. 8(11),
8707–8718 (2021)
432 F. J. Alruwaili et al.
3. Mutyara, A.G., Farras, B.A., Sari, L.P., Achmad, S., Sutoyo, R.: The influence of smartphone
applications on human healthcare. In: International Conference on Informatics Electrical and
Electronics (ICIEE), pp. 1–6 (2022)
4. ur Rahman, G.M.E., Chowdhury, R.I., Dinh, A., Wahid, K.A.: A smart sensor node with
smartphone based IoMT. In IEEE International Conference on Consumer Electronics - Asia
(ICCE-Asia), pp. 92–95 (2019)
5. Iqbal, W., Abbas, H., Daneshmand, M., Rauf, B., Bangash, Y.A.: An in-depth analysis of IoT
security requirements, challenges, and their countermeasures via software-defined security.
IEEE Internet Things J. 7(10), 10250–10276 (2020)
6. T. L. Lin and Y. S. Chen. A chinese continuous keystroke authentication method using cog-
nitive factors. In IEEE International Conference on Consumer Electronics - Taiwan (ICCE-
TW), pp. 1–2 (2019)
7. Herath, H.M.C.K.B., Dulanga, K.G.C., Tharindu, N.V.D., Ganegoda, G.U.: Continuous user
authentication using keystroke dynamics for touch devices. In: 2nd International Conference
on Image Processing and Robotics (ICIPRob), pp. 1–6 (2022)
8. Ouadjer, Y., Adnane, M., Bouadjenek, N.: Feature importance evaluation of smartphone
touch gestures for biometric authentication. In: 2nd International Workshop on Human-
Centric Smart Environments for Health and Well-being (IHSH), pp. 103–107 (2021)
9. Suharsono, A., Liang, D.: Hand stability based features for touch behavior smartphone
authentication. In: 3rd IEEE International Conference on Knowledge Innovation and Inven-
tion (ICKII), pp. 167–170 (2020)
10. Musale, P., Baek, D., Choi, B.J.: Lightweight gait based authentication technique for IoT
using subconscious level activities. In: IEEE 4th World Forum on Internet of Things (WF-
IoT), pp. 564–567 (2018)
11. He, L., Ma, C., Tu, C., Zhang, Y.: Gait2vec: Continuous authentication of smartphone users
based on gait behavior. In: IEEE 25th International Conference on Computer Supported
Cooperative Work in Design (CSCWD), pp. 280–285 (2022)
12. Roy, S., et al.: A systematic literature review on latest keystroke dynamics based models.
IEEE Access 10, 92192–92236 (2022)
13. SimonD, J.: Mobile location history of 10/2014. Kaggle (2017). Accessed 28 Dec 2022
14. Zhang, L., Liu, L.: Data anomaly detection based on isolation forest algorithm. In: Interna-
tional Conference on Computation, Big-Data and Engineering (ICCBE), pp. 87–89, Yunlin,
Taiwan (2022)
15. Marcelli, E., Barbariol, T., Savarino, V., Beghi, A., Susto, G.A.: A revised isolation for-
est procedure for anomaly detection with high number of data points. In: IEEE 23rd Latin
American Test Symposium (LATS), pp. 1–5, Montevideo, Uruguay (2022)
A Configurable Activation Function
for Variable Bit-Precision DNN Hardware
Accelerators
x1 w1 sigmoid sign
MUX f(R)=AFout (N-1)-q bit q bits
inputs
Fig. 1. Neuron architecture with MAC and multi-AFs, utilizing fixed-point computa-
tion format <N, q>.
architecture has been proposed in [9]. This design efficiently realizes sigmoid and
tanh AFs using the same logic resources. However, it has two main drawbacks:
low accuracy and high LUT utilization for bit-precision ≤ 8. To overcome these
limitations, we propose a new approach that combines the Cordic algorithm for
high bit-precision AF and ROM for low bit-precision AF (≤ 8). The distinct
contributions of this paper are as follows:
The scaling factor coshθi = 0.8281 is factored out from Eq. 2 as the pseudo-
1
rotation scaling factor, while coshθi
= 1.2075 is applied at Pi as an offset. To
derive Eq. 3, θi is substituted with di · Ei , where di ∈ {−1, 1} determines the
negative (−1) or positive (1) rotational direction and Ei represents the memory
element (Lookup Table) for the ith iteration (shown in Fig. 2). The values of
Ei and mode m depend on the type of coordinate system (linear, circular, or
hyperbolic) being employed [5]: Ei ∈ {2−i , tan−1 (2−i ), tanh−1 (2−i )}, and m ∈
436 S. Vishwakarma et al.
{0, 1, −1}, respectively. Therefore, for hyperbolic mode, Ei = tanh−1 (2−i ), and
m = −1.
The core of the configurable AF, as depicted in Fig. 3, consists of the ROM/
Cordic Block (Fig. 2). The ROM/ Cordic Block incorporates adder/subtractors,
shifters, and memory elements. The most significant bit (MSB) of Rin [N-1] (sign
bit) generates the directional signal di [5], determining whether addition or sub-
traction is performed such that Rout converges to 0. Here di ∈ {0, 1} as the sign
bit Rin [N-1] ∈ {0, 1}. Equation 3 has been modified to the equations presented
in algorithm 1 for hardware realization. The Add/Sub block utilizes the 2's
complement form for subtraction operations. The 1:2 DeMux, controlled by the
precision_select signal, selects the input Rin [N-1:0] for either ROM or Cordic
operation. In the ROM-based approach (Fig. 2), the value at the Rin address
is accessed as ROM[Rin ]. Depending on the implementation of the ROM, the
AF’s output for sigmoid or tanh is obtained. Thus, Rout = ROM[Rin ] for the
ROM-based approach, while Rout converges to 0 for the Cordic-based approach
A Configurable Activation Function for Variable Bit-Precision 437
(Fig. 2). Additionally, the state machine in Fig. 2 generates control signals for
input and feedback based on the clock and reset signals. The output of the
Cordic block in Fig. 2 produces the values cosh(Rin ) and sinh(Rin ) at Pout
and Qout , respectively. These outputs are used for exponential calculation, as
described in Eq. 4.
Shift-reg Shift-reg
constant
Fig. 2. ROM/Cordic Block. The precision_select line allows for the selection of
either the ROM[Rin ] output (sigmoid or tanh) or the Cordic-based output i.e.,
sinh(Rin ) and cosh(Rin )
438 S. Vishwakarma et al.
AF_select
ADDER1
1.0
ROM[Rin] / 0
e^Rin
ADDER2
1+e^Rin
MUX1 MUX2
DIVISION Block
Cordic[Rin]
MUX3
To evaluate the accuracy of the AF, LeNET and VGG-16 [1] DNN models,
along with the MNIST, CIFAR-10, and CIFAR-100 datasets are utilized. The
inference accuracy results for Tensor (T) and proposed AF (P) using varying-bit
A Configurable Activation Function for Variable Bit-Precision 439
precision (4, 8, 16, and 32-bit) are presented in Table 2. This paper uses the
terms ‘fixed <9, 6>’ and 8-bit precision interchangeably, as the 9th bit (MSB) is
the sign bit. The same applies for all bit precisions presented in this paper. The
ROM-based approach demonstrates superior accuracy for low-bit precision com-
putation (≤ 8), while the Cordic-based approach offers considerable accuracy
improvements for higher-bit precision. Across all bit widths, the proposed con-
figurable AF achieves accuracy levels comparable to the tensor-based model [1],
with an accuracy loss of less than 1.5%.
A comparative analysis of resource utilization is presented in Table 3 for
ROM, Cordic, and BRAM-based approaches across different bit-precisions. For
4-bit precision, the ROM-based design employs 6 LUTs, while the Cordic-based
design utilizes 45 LUTs and 37 flip-flops (FFs). This results in a notable LUT
saving of 86.66% for the ROM-based design compared to Cordic. Similarly, in the
case of 8-bit precision, the ROM-based design requires 16 LUTs, whereas Cordic
utilizes 84 LUTs and 72 FFs, achieving a LUT saving of 80.95%. Remarkably,
the ROM-based design exclusively relies on LUTs and does not require any FFs.
However, as the precision increases to 16-bit, the ROM-based design demands
a substantial number of 2111 LUTs, in contrast to the Cordic-based design’s
requirement of 140 LUTs and 126 FFs. Furthermore, it is observed that imple-
menting a 32-bit ROM-based design on smaller FPGAs is infeasible due to the
exponential increase in resource requirements (2N memory elements). In con-
trast, Cordic requires 257 LUTs and 221 FFs for 32-bit precision. Additionally,
the evaluation includes the BRAM-based approach, which reveals a significant
increase in BRAM utilization as precision increases. Precisely, for 4, 8, and 16-bit
precisions, the corresponding BRAM requirements are 0.5, 0.5, and 17 BRAMs,
440 S. Vishwakarma et al.
References
1. Abadi, M., et al.: TensorFlow: Large-scale machine learning on heterogeneous sys-
tems. Software available from tensorflow.org (2015)
2. Aggarwal, S., Meher, P.K., Khare, K.: Concept, design, and implementation of
reconfigurable CORDIC. IEEE Trans. Very Large Scale Integr. (VLSI) Syst. 24(4),
1588–1592 (2015)
A Configurable Activation Function for Variable Bit-Precision 441
3. Lee, J., et al.: Unpu: an energy-efficient deep neural network accelerator with fully
variable weight bit precision. IEEE J. Solid-State Circuits 54(1), 173–185 (2018)
4. Lin, C.H., Wu, A.Y.: Mixed-scaling-rotation CORDIC (MSR-CORDIC) algorithm
and architecture for high-performance vector rotational DSP applications. IEEE
Trans. Circuits Syst. I Regul. Pap. 52(11), 2385–2396 (2005)
5. Mehra, S., Raut, G., Das, R., Vishvakarma, S.K., Biasizzo, A.: An empirical eval-
uation of enhanced performance softmax function in deep learning. IEEE Access
11, 34912–34924 (2023)
6. Mohamed, S.M., et al.: FPGA implementation of reconfigurable CORDIC algo-
rithm and a memristive chaotic system with transcendental nonlinearities. IEEE
Trans. Circuits Syst. I Regul. Pap. 69(7), 2885–2892 (2022)
7. Prashanth, H., Rao, M.: Somalib: Library of exact and approximate activation
functions for hardware-efficient neural network accelerators. In: 2022 IEEE 40th
International Conference on Computer Design (ICCD), pp. 746–753. IEEE (2022)
8. Raut, G., Karkun, S., Vishvakarma, S.K.: An empirical approach to enhance perfor-
mance for scalable cordic-based deep neural networks. ACM Trans. Reconfigurable
Technol. Syst. (2023)
9. Raut, G., Rai, S., Vishvakarma, S.K., Kumar, A.: RECON: resource-efficient
CORDIC-based neuron architecture. IEEE Open J. Circuits Syst. 2, 170–181
(2021)
10. Raut, G., et al.: Data multiplexed and hardware reused architecture for deep neural
network accelerator. Neurocomputing 486, 147–159 (2022)
11. Sakr, C., Kim, Y., Shanbhag, N.: Analytical guarantees on numerical precision
of deep neural networks. In: International Conference on Machine Learning, pp.
3007–3016. PMLR (2017)
Authentication and Authorization of IoT
Edge Devices Using Artificial Intelligence
1 Introduction
rate of 14.4% from 2017 to 2021 [3]. This increased consumption of internet
resources further emphasizes the need for robust security measures to protect
IoT infrastructure from potential attacks.
Since the late 1990 s, IoT devices have been manufactured with little emphasis
on security [4]. The limited resources and constrained nature of end devices
present a significant obstacle to implementing robust security measures in IoT
[5]. Security and privacy remain ongoing challenges in the IoT landscape [5].
Various solutions have been proposed to address these issues, however, most
of them involve adding some computational power and storage in IoT devices,
which are not practical. To tackle these challenges, fog computing has been
introduced as a potential solution [6]. However, authentication and authorization
continue to be key security issues, especially due to the resource limitations of end
devices [6]. Use of Artificial Intelligence (AI) to address these challenges under
fog environment has not been proposed so far. This paper focuses on addressing
these authentication and authorization challenges using AI technology within
a fog computing model. By leveraging AI and machine learning techniques, a
proposed solution aims to enhance the security of IoT devices in a resource-
efficient manner. Authors have already proposed the model in [7]. However, in
this paper, specific AI techniques to be used in proposed model, have been
suggested, thus taking the model a step further.
The primary contribution of this paper is the utilization of AI for authenti-
cation and authorization of edge devices, without requiring additional compu-
tational capacity, storage, or power from the edge devices. The structure of the
paper is as follows: Sect. 2 presents a comprehensive literature review, discussing
the research on the use of AI for authentication, authorization, and enhanc-
ing security in IoT. In Sect. 3, an AI-based framework for the authentication
and authorization of end devices is proposed. In Sect. 4, implementation of AI
techniques for authentication of IoT devices has been discussed. Finally, Sect. 5
concludes the paper.
2 Literature Review
Computational, storage, and end-device power limitations in the Internet of
Things pose serious challenges in implementing workable solutions that address
security concerns in the Internet of Things. This section describes some of the
latest research in addressing end device security issues.
Roman et al. We evaluate authentication and access control issues in research
and analysis of security threats and challenges for mobile edge computing [8].
They recommend studying the applicability of other edge network distributed
authentication mechanisms for IoT. It has also been concluded that edge device
security is in its infancy.
444 M. S. Zareen et al.
The proposed solution in this paper is focused on the smart home scenario, where
the limited computational capacity, storage, and power of edge devices make it
challenging to incorporate authentication and authorization directly into them.
As mentioned in Sect. 2, adding extra computation or storage to edge devices is
not practical. To overcome this limitation, the paper suggests adding an addi-
tional hardware component at the fog layer in the user’s premises for authenti-
cation and authorization purposes. An AI-based solution is proposed to run on
this additional hardware to authenticate and authorize edge devices. Clustering
techniques of unsupervised learning and classification techniques of supervised
learning are recommended to be used for this purpose. The proposed solution
effectively addresses the long-standing issues of authentication and authorization
of edge devices in IoT. The modified architecture of Fog Computing, along with
the addition of the proposed AI-based computer, is depicted in Fig. 1.
In order to utilize the IoT devices of a specific manufacturer, the user is required
to download and install a user application on their computer or smartphone.
This application allows the user to create an account with the device manufac-
turer. When adding a new smart device to the smart home, the user needs to
configure it for communication with its cloud-based application server and to
enable remote management by the owner. The user completes this setup process
through the user application. During the installation process, the smart device is
configured and assigned a password to access the cloud-based application server
managed by the manufacturer through a fog gateway router. The device’s unique
ID is then linked to the user account on the application server, allowing the user
to access it using their login credentials from their computer or smart device.
Occasionally, IoT devices can also be accessed through a web browser; however,
the user must still provide login credentials. Each type and manufacturer of IoT
device has a unique format and data structure for exchanging data.
446 M. S. Zareen et al.
Smart devices have two-way communication, where the owner can communi-
cate with the device and the device can communicate with its owner. However,
the format and structure of data exchange vary for each edge device, depending
on its functionality. The data exchange process is illustrated in Fig. 2.
Smart devices rely on two-way communication between the device owner
and the device itself. The specific type of data exchange varies depending on
the functionality of each individual edge device. As illustrated in Fig. 2, this
exchange typically involves the owner sending update requests to the device and
the device responding with status notifications and other relevant information
via a fog router. To understand the data exchange mechanism, few examples are
discussed in succeeding paras.
For example, a smart bulb may receive update requests from its owner regard-
ing its status, color settings, and other customizable features. In response, the
bulb sends notifications to its owner through the fog router indicating its current
status (on or off) and color setting (in the case of an RGB bulb), as well as other
relevant information. Upon receiving these updates, the owner can then provide
instructions to the bulb to modify its status, color, or other settings as needed.
Smart devices have distinct types of communication: from the owner to the
smart device and from the smart device to its owner. The data exchange pro-
cedure is unique for each edge device in IoT, depending on its functionality. To
illustrate the data exchange process, several examples are provided below.
Authentication and Authorization of IoT 447
For example, communication from the owner to a smart bulb involves requests
for status updates, color settings, and other relevant information, according to
the manufacturer’s offerings. The smart bulb communicates its status notifi-
cation (on or off) and color settings (in the case of an RGB bulb), and other
relevant information to its owner through a fog router. After receiving an update
about the bulb, the owner can give instructions to modify the status, color, or
related information/settings.
Communication from the owner to a smart water heating system involves
update requests on status, current temperature, desired water temperature, and
other relevant information, as per the manufacturer’s offerings (general settings
triggering its auto on/off, record of its previous activities, etc.). The smart water
heating system communicates its status notification, current water temperature,
desired water temperature, and other relevant information to its owner. Sub-
sequently, the owner may give instructions regarding changing the status and
related settings.
Communication from the owner to a smart coffee maker involves update
requests on status, supplies needed to make coffee, and other information, as
per the manufacturer’s offerings (general settings of auto-making coffee, etc.).
The smart coffee maker communicates its status notification, the state of water,
coffee beans, and other necessary supplies needed to make coffee, and other
related information to its owner. The owner may give instructions on preparing
the coffee or changing related settings.
448 M. S. Zareen et al.
In summary, each smart edge device has its unique set of communications
and data exchange procedures depending on its type and the services offered by
the device manufacturer.
– Data Preprocessing. The first step in the data preprocessing phase is to clean
the raw data generated by the edge device [18]. This involves filling in missing
data, smoothing out noise, removing outliers and inconsistencies, and other
techniques to ensure that the data is accurate and consistent. However, data
integration is not necessary in the proposed model, as it is assumed that the
same type of IoT device (e.g., a smart bulb) from the same manufacturer
Authentication and Authorization of IoT 449
– Data Preprocessing. In the data preprocessing phase, the raw data generated
by the edge device is refined for use in authentication with AI techniques
[18]. The first step is to clean the data, which may contain errors due to wire-
less transmission or inconsistency in the edge device’s performance. Clean-
ing involves filling in missing data, smoothing noise, removing outliers and
inconsistencies. Data integration is not required in the proposed model, as
we assume that devices of the same type from the same manufacturer have
the same communication and data exchange patterns. Data reduction is also
not performed, as the feature space of each IoT type is not large enough to
require reduction. The cleaned data is then transformed or consolidated into
the appropriate form of mining, referred to as refined data, which is the most
efficient for the AI algorithm used in subsequent steps.
– Division of Refined Data. The refined data is divided into two parts: 70% for
training the learning algorithm and 30% for validating/testing the trained
model. After training and validation, live input of the refined dataset will be
provided to the trained model for the authentication of edge devices.
The device owner typically uses a smartphone or laptop to access their smart
devices, and these devices have ample processing capability, storage, and power.
Therefore, incorporating authentication and authorization of device owners in
450 M. S. Zareen et al.
IoT is not a significant challenge. Several solutions have been proposed in various
research papers, as discussed in Sect. 2. This paper does not propose a new
solution for device owner authentication and authorization, as existing solutions
can be used for this purpose.
– Data Cleaning. Data cleaning will be carried out in first phase. Cleaning is
carried out for data which becomes dirty in shape of outliers, inconsistencies,
noise and missing values. In context of smart home, wireless communication
media is assumed to be main cause of making the data dirty. As data is
coming live, hence “ignoring the tuple” technique will be adopted for dirty as
it will have no affect on volume of data.
– Data Transformation and Discretization. Data discretization will be used to
transform continuous attributes to categorical attributes to speed up the
implementation of AI techniques during authentication and authorization
phases. Some of the examples of continuous attributes are colour settings
Authentication and Authorization of IoT 451
Recall will return the percentage of correctly identified items of a type from
total population of that item in the dataset.
Accuracy will return the percentage of total correctly classified items from
complete dataset.
5 Conclusion
References
1. Iot connected devices to reach 20.4 billion by 2020, says gartner. https://which-
50.com/iot-connected-devices-reach-20-4-billion-2020-says-gartner/
2. Columbus, L.: (2017) Internet of things market to reach $267b by 2020. https://
www.forbes.com/sites/louiscolumbus/2017/01/29/internet-of-things-market-to-
reach-267b-by-2020/#1a073d14609b
3. Framingham, M.: Idc forecasts worldwide spending on the internet of
things to reach $772 billion in 2018 (2017). https://www.idc.com/getdoc.jsp?
containerId=prUS43295217
4. Zareen, M.S., Tariq, M.: Internet of things (IoT): the next paradigm shift but
whats the delay? In: 17th IEEE International Multi Topic Conference 2014, pp.
143–148, December 2014
5. El-Mougy, A., Al-Shiab, I., Ibnkahla, M.: Scalable personalized IoT networks. Proc.
IEEE 107(4), 695–710 (2019)
6. Bonomi, F., Milito, R.: Fog computing and its role in the internet of things. In:
Proceedings of the MCC Workshop on Mobile Cloud Computing, August 2012
7. Zareen, M.S., Tahir, S., Akhlaq, M., Aslam, B.: Artificial intelligence/machine
learning in IoT for authentication and authorization of edge devices. In: 2019
International Conference on Applied and Engineering Mathematics (ICAEM), pp.
220–224. IEEE (2019)
8. Roman, R., Lopez, J., Mambo, M., Mobile edge computing, fog et al.: A
survey and analysis of security threats and challenges. Future Gener. Com-
put. Syst. 78, 680–698 (2018). https://www.sciencedirect.com/science/article/pii/
S0167739X16305635
9. Iorga, M., Feldman, L., Barton, R., Martin, M.J., Goren, N., Mahmoudi, C.: Nist
special publication 500–325 - fog computing conceptual model (2018) . https://
doi.org/10.6028/NIST.SP.500-325
10. Ibrahim, M.: Octopus: An edge-fog mutual authentication scheme. Int. J. Netw.
Secur. 18, 1089–1101 (2016)
Authentication and Authorization of IoT 453
11. Jia, X., He, D., Kumar, N., Choo, K.K.R.: A provably secure and efficient identity-
based anonymous authentication scheme for mobile edge computing. IEEE Syst.
J. 14(1), 560–571 (2019)
12. Xiong, L., Peng, D., Peng, T., Liang, H.: An enhanced privacy-aware authentication
scheme for distributed mobile cloud computing services. KSII Trans. Internet Inf.
Syst. 11, 6169–6187 (2017)
13. Chiang, M., Zhang, T.: Fog and IoT: an overview of research opportunities. IEEE
Internet Things J. 3(6), 854–864 (2016)
14. Sezer, O.B., Dogdu, E., Ozbayoglu, A.M.: Context-aware computing, learning, and
big data in internet of things: a survey. IEEE Internet Things J. 5(1), 1–27 (2018)
15. Wan, J., Jun, Y., Zhongren, W., Qingsong, H.: Artificial intelligence for cloud-
assisted smart factory. IEEE Access 6, 55419–55430 (2018)
16. Chin, W.S., soo Kim, H., Heo, Y.J., Jang, J.W.: A context-based future network
infrastructure for IoT services. In: Procedia Computer Science, vol. 56, pp. 266–270,
2015, the 10th International Conference on Future Networks and Communications
(FNC 2015)/The 12th International Conference on Mobile Systems and Perva-
sive Computing (MobiSPC 2015) Affiliated Workshops. https://www.sciencedirect.
com/science/article/pii/S1877050915016889
17. Blazek, P., Krejcar, O., Jun, D., Kuca, K.: Device security implementation model
based on internet of things for a laboratory environment. IFAC-PapersOnLine
49(25), 419–424 (2016). 14th IFAC Conference on Programmable Devices and
Embedded Systems PDES 2016. https://www.sciencedirect.com/science/article/
pii/S2405896316327240
18. Han, J., Kamber, M., Pei, J.: Data Mining: Concepts and Techniques, 3rd ed. 225
Wyman Street, Waltham, MA 02451, Elsevier, USA (2012)
Secure Dynamic PUF for IoT Security
1 Introduction
2 Background
2.1 PUF Designs
Arbiter PUF (APUF) is the most researched PUF because of its easy imple-
mentation and ability to produce more CRPs. APUF is derived from racing
conditions between two identical paths. Depending on the time the Multiplexer
chain takes, the flipflop used as arbiter at the end triggers 1 or 0. The APUF
design demands a completely identical path for race conditions to avoid biased
responses due to delay differences added by wires. The Arbiter PUF generates
a random response that is not predictable, and due to manufacturing variation,
the adversary can’t clone the APUF. This property of PUF makes it secure
against side-channel attacks. However, several modeling attacks have been pro-
posed earlier to predict the PUF response [13]. Several complex architectures of
PUFs, such as XOR PUF, Double Arbiter PUF(DA PUF), and Anderson PUFs,
have been proposed. Despite the complex architecture, machine learning-based
modeling attacks can clone the behavior of PUF.
To enhance the resilience of the arbiter PUF, Suh and Devadas proposed
using an XOR PUF [14]. This approach involves merging several arbiter PUFs,
with the response from each arbiter being XORed to produce the output response
of the XOR PUF. While this introduces nonlinearity to the design, it also
456 S. Rajput and J. Dofe
3 Experimental Analysis
3.1 Experimental Setup
The APUF architecture was implemented on the Xilinx Artix 7 100T board
using the Xilinx Vivado tool, and all challenges are captured at room tempera-
ture. Thousands of CRPs are required to obtain performance metrics. As shown
in Fig. 1, a control unit is designed to generate random challenge bits using Lin-
ear Feedback Shift Register (LFSR) and pass them to the APUF. The control
unit consists of three main modules: LFSR, PUF, and RAM, which generate
a random response bit based on their delay characteristics. Xilinx’s Integrated
Logic Analyzer (ILA) tool captures approximately 131K randomly generated
responses from the PUF and LFSR-generated challenge bits in one round. The
ILA tool made debugging the APUF, real-time designing, and recording CRPs
easier. Floor planning is performed using the Xilinx Vivado tool to ensure accu-
rate response pairs for the APUF. The Vivado tool automatically redesigns the
synthesis according to behavioral logic. However, the PUF functionality cannot
be distinguished by behavioral logic as, ideally, the circuit output should be the
same for the identical path. It is necessary to ensure that the auto-synthesis
tool of Vivado does not change the path design of PUF according to behavioral
logic. To avoid auto-optimization by Vivado DONT TOUCH attribute is used.
The APUF and its variants can generate a higher number of CRPs due to the
switching of the cross-coupled and parallel paths due to the select line of the
multiplexer. The implementation of this multiplexer is depicted in Fig. 2 more
precisely.
Secure Dynamic PUF for IoT Security 457
be XORed with path 2 for multiple response bits to alter the pattern of mul-
tiple responses. Assuming the randomness of the PUF in generating an equal
distribution of 50% 1’s and 0’s for both the intermediate and main arbiter PUF,
approximately 12–15% of the final response will be biased, resulting in a pattern
that thwarts various modeling attacks.
First, we evaluate the resilience of the proposed DPUF against the most suc-
cessful modeling attacks on PUF functions–MLP and LR and compare it with
APUF. Table 1 shows the prediction accuracy of these attacks on the existing
APUF and proposed DPUF design for 10k, 100k, 500k, and 1 Million CRPs. The
accuracy is determined by applying the formula: (Number of correct predictions
/ Total number of predictions). We observed that the maximum accuracy for
the MLP and LR attack on DPUF was 81.11% and 77.97%, respectively, for 1
Million CRPs, whereas that of APUF was over 98%.
Table 1. Prediction of Proposed DPUF with APUF on Test Data for MLP and LR
Attack Models.
CRPs MLP LR
APUF DPUF APUF DPUF
10k 72.90 % 64.00 % 88.00 % 51.00 %
100k 88.40 % 79.99 % 94.61 % 75.20 %
500K 98.28 % 81.09 % 95.22 % 75.02 %
1M 98.99 % 81.11 % 98.36 % 77.97 %
Secure Dynamic PUF for IoT Security 459
Next, we analyze the impact of the MLP attack on the DPUF for 10k, 100k,
and 1 Million CRPs of 4-XOR, 5-XOR, and 6-XOR with DPUF, as depicted in
Fig. 4. The figure shows that the prediction accuracy for all PUF designs, except
6-XOR PUF, was around 60% for 10k challenges, with 6-XOR PUF having
a prediction accuracy of around 50%. When trained on 100k CRPs, the model
predicted the 4 & 5-XOR variants with over 95% accuracy, DPUF at around 78%,
while 6-XOR remained close to 55%. However, when the MLP model was trained
with 1 Million CRPs, it successfully predicted the response of all mentioned XOR
variants with over 95% accuracy. Nonetheless, DPUF’s accuracy remained close
to 80%, even with one Million CRPs. This demonstrates that DPUF poses a
more significant challenge to MLP-based algorithms, making it more difficult
to predict response bits than XOR PUFs. Finally, we compare the proposed
PUF’s resilience with n-XOR PUF against the LR attack in Fig. 5. The results
reveal that DPUF outperforms 4-XOR PUF in terms of resistance to LR attack.
However, 5 and 6-XOR PUFs exhibited greater immunity toward LR attack than
the proposed dynamic PUF.
Fig. 4. Comparison of Prediction Accuracy on XOR and Proposed PUF using MLP
Attack.
460 S. Rajput and J. Dofe
The analysis clearly showed that the proposed DPUF offers significantly
higher resilience against both types of modeling attacks compared to the APUF
and 4-XOR PUF. Additionally, it is worth mentioning that the proposed PUF
requires significantly fewer resources compared to previously suggested complex
PUF architectures.
References
1. Rajput, S., Dofe, J.: Counteracting modeling attacks using hardware-based
dynamic physical unclonable function. In: 2023 IEEE International Conference on
Cyber Security and Resilience (CSR), Venice, Italy, pp. 586–591 (2023). https://
doi.org/10.1109/CSR57506.2023.10224914
2. Anderson, R., Kuhn, M.: Tamper resistance - a cautionary note new. In:
2nd USENIX Workshop on Electronic Commerce (EC 96). Oakland, CA,
USENIX Association, November 1996. https://www.usenix.org/conference/2nd-
usenix-workshop-electronic-commerce/tamper-resistance-cautionary-note
3. Gassend, B., Clarke, D., van Dijk, M., Devadas, S.: Silicon physical random func-
tions. In: Proceedings of the 9th ACM Conference on Computer and Communica-
tions Security, Ser. CCS 2002. New York, NY, USA, Association for Computing
Machinery, pp. 148–160 (2002). https://doi.org/10.1145/586110.586132
4. Yu, M.-D., Sowell, R., Singh, A., M’Raı̈hi, D., Devadas, S.: Performance metrics
and empirical results of a PUF cryptographic key generation ASIC. IEEE Int.
Symp. Hardware-Oriented Secur. Trust 2012, 108–115 (2012)
5. Khalfaoui, S., Leneutre, J., Villard, A., Gazeau, I., Ma, J., Urien, P.: Security
analysis of machine learning-based PUR enrollment protocols: a review. Sensors
21, 8415 (2021)
6. Strieder, E., Frisch, C., Pehl, M.: Machine learning of physical unclonable functions
using helper data - revealing a pitfall in the fuzzy commitment scheme, Cryptology
ePrint Archive, Paper 2020/888 (2020). https://eprint.iacr.org/2020/888
7. Tobisch, J., Aghaie, A., Becker, G.T.: Combining optimization objectives: new
modeling attacks on strong PUFs. IACR Trans. Cryptographic Hardware Embed-
ded Syst. 2021(2), 357–389 (2021). https://tches.iacr.org/index.php/TCHES/
article/view/8798
8. Santikellur, P., Bhattacharyay, A., Chakraborty, R.S.: Deep learning based model
building attacks on arbiter PUF compositions. IACR Cryptol. ePrint Arch. 2019,
566 (2019)
9. Wisiol, N., et al.: Splitting the interpose PUF: a novel modeling attack strategy.
Cryptology ePrint Archive, Paper 2019/1473, 2019, https://eprint.iacr.org/2019/
1473
10. Li, G., Mursi, K.T.: A subspace pre-learning strategy to break the interpose PUF.
Electronics 11(7) (2022). https://www.mdpi.com/2079-9292/11/7/1049
11. Zhuang, H., Xi, X., Sun, N., Orshansky, M.: A strong subthreshold current array
PUF resilient to machine learning attacks. IEEE Trans. Circuits Syst. I Regul.
Pap. 67(1), 135–144 (2020)
12. Xiong, W., Schaller, A., Katzenbeisser, S., Szefer, J.: Dynamic physically unclon-
able functions. In: Proceedings of the 2019 on Great Lakes Symposium on VLSI,
ser. GLSVLSI 2019. New York, NY, USA, Association for Computing Machinery,
pp. 311–314 (2019). https://doi.org/10.1145/3299874.3318025
13. Rührmair, U., et al.: PUF modeling attacks on simulated and silicon data. IEEE
Trans. Inf. Forensics Secur. 8(11), 1876–1891 (2013)
462 S. Rajput and J. Dofe
14. Suh, G.E., Devadas, S.: Physical unclonable functions for device authentication and
secret key generation. In: 2007 44th ACM/IEEE Design Automation Conference,
pp. 9–14 (2007)
15. Tobisch, J., Becker, G.T.: On the scaling of machine learning attacks on PUFs with
application to noise bifurcation. In: Mangard, S., Schaumont, P. (eds.) RFIDSec
2015. LNCS, vol. 9440, pp. 17–31. Springer, Cham (2015). https://doi.org/10.1007/
978-3-319-24837-0 2
Author Index
K P
Kalimuthu, Sivakumar I-293, I-306, I-316 Panda, Satyajit II-36
Kalyanam, Lakshmi Kavya I-34, I-76 Panigarhi, Narayan II-402
Karabiyik, Umit I-337 Pati, Umesh Chandra I-229
Karam, Robert I-199, II-410 Patra, Jayaprakash II-36
Karthikeyan, S. I-362 Patra, Ramapati I-261
Kashef, Mohamed I-182 Perry, Nicholas I-272
Katkoori, Srinivas I-34, I-76, I-168 Perumal, Thinagaran I-316
Keller, Myles I-199 Pradhan, Ashok Kumar II-114, II-133
Kethineni, Kiran Kumar I-415 Puthal, Deepak II-212, II-227
Khatik, Sadhvi II-212
Kougianos, Elias I-375, I-405, I-415, I-424,
Q
II-283, II-371, II-381
Qi, Xiaowen I-182
L
Lakshmanan, An Sakthi I-306 R
Liu, Ying I-3, II-14, II-179, II-333 Rachakonda, Laavanya I-395, II-302, II-390
Rajesh, M. A. II-402
M Rajput, Shailesh I-454
Mahapatra, Kamalakanta I-229 Ramanujam, E. I-316
Mahmud, Md Ishtyaq II-246 Raut, Gopal I-433
Mahmud, Shakil I-199, II-410 Roy, Swapnoneel I-104, II-64
Maji, Prasenjit I-261
Mandke, Pranav I-104 S
Mishra, Alekha Kumar II-212 Sadhu, Pintu Kumar II-246, II-262
Mitra, Alakananda I-405, I-415, II-371 Sahith, Chitumalla II-3
Mohanty, Prajnyajit I-229 Sankaran, Sriram I-114
Mohanty, Saraju P. I-375, I-405, I-415, Shah, Iqra Amin II-198
I-424, II-283, II-371, II-381 Shah, Sneh II-64
Mondal, Hemanta Kumar I-261 Shill, Ponkoj Chandra I-62
Mondal, Rishabh II-74 Shillingford, Nadine II-309
Monteiro, Maxwell Eduardo I-18 Shivam, Pant II-49
Morshed, Bashir I. I-135 Sibi Chakkaravarthy, S. I-362
Mukhopadhyay, Debajyoti I-104, II-64 Siddiqui, Arish II-87, II-101
Muresan, Anca O. I-353 Singh, Aniket I-104
Musale, Vinayak I-104 Singh, Harshdeep I-97
Singh, Rohit II-3
N Somesula, Raaga Sai I-168
Naga Nithin, G. II-133, II-160 Song, Sejun I-48
Naik, Varsha II-64 Stasiewicz, Samuel I-395
Narayan, Panigrahi II-49 Swain, Gandharba II-133
Author Index 465