Advances in Computational Intelligence - Security and Internet of Things 2019 PDF
Advances in Computational Intelligence - Security and Internet of Things 2019 PDF
Saha
Nirmalya Kar
Suman Deb (Eds.)
Advances in Computational
Intelligence, Security and
Internet of Things
Second International Conference, ICCISIoT 2019
Agartala, India, December 13–14, 2019
Proceedings
Communications
in Computer and Information Science 1192
Commenced Publication in 2007
Founding and Former Series Editors:
Phoebe Chen, Alfredo Cuzzocrea, Xiaoyong Du, Orhun Kara, Ting Liu,
Krishna M. Sivalingam, Dominik Ślęzak, Takashi Washio, Xiaokang Yang,
and Junsong Yuan
Advances in Computational
Intelligence, Security and
Internet of Things
Second International Conference, ICCISIoT 2019
Agartala, India, December 13–14, 2019
Proceedings
123
Editors
Ashim Saha Nirmalya Kar
National Institute of Technology Agartala National Institute of Technology Agartala
Agartala, India Agartala, India
Suman Deb
National Institute of Technology Agartala
Agartala, India
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Preface
It is a matter of great privilege to have been tasked with the writing of this preface for
the proceedings of Second International Conference on Computational Intelligence,
Security & IoT (ICCISIoT 2019) organized by the Department of Computer Science
and Engineering of the National Institute of Technology Agartala, India, during
December 13–14, 2019. This mega event covered the major aspects of computational
intelligence, security, and the Internet of Things (IoT) where the scope was not limited
only to various engineering disciplines such as computer science and electronics
engineering researchers but also included researchers from allied communities working
in bioinformatics, computational intelligence for smart power and energy management,
artificial intelligence and cognitive science, modeling and simulation, smart factories
and Industry 4.0, etc. This conference series was initiated with a focus of bringing
together researchers, academics, scientists, and industry experts from around the globe
to share thoughts and promote research activities to derive benefits from the advances
of next-generation computing technologies in the listed areas. It was our wish and hope
that the people participating in the conference would find common ground on which
they could learn from each other, and that the conference would engender future fruitful
scientific activities. Such collaborative endeavors are necessary for the development
of the cutting-edge solutions embodied within our society and its fabrics.
The Program Committee of ICCISIoT 2019 is extremely grateful to the authors who
showed an overwhelming response to the call for papers, submitting over 153 papers in
four tracks. The entire review team (Technical Program Committee members along
with external experts as reviewers) expended tremendous effort to ensure fairness and
consistency during the selection process resulting in the best quality papers being
selected for presentation and publication. It was ensured that every paper received at
least three, and in most cases four, reviews. Checking of similarities and overlaps was
also done based on the international norms and standards. We are very thankful to our
reviewers for their efforts in finalizing the high-quality papers and finally, based on the
recommendation, 37 papers were accepted for publication. The proceedings of the
conference are published as one volume in the Communications in Computer and
Information Science (CCIS) series by Springer, and are also indexed by ISI Proceed-
ings, DBLP, Ulrich’s, EI-Compendex, SCOPUS, Zentralblatt Math, MetaPress, and
Springerlink. We, in our capacity as volume editors, convey our sincere gratitude to
Springer for providing the opportunity to publish the proceedings of ICCISIoT 2019 in
their CCIS series.
The conference program was enriched by a series of keynote presentations and the
keynote speakers included: Prof. Anu A. Gokhale (Illinois State University, USA);
Dr. Ahcene Bounceur (University of Brest, France); Prof. Manoj Sachan (Sant
Longowal Institute of Engineering and Technology, Punjab, India); etc. We are grateful
to them for sharing their insights on their latest research with us.
vi Preface
Honorary Chairs
Harish Kumar Sharma NIT Agartala, India
Tokunbo Ogunfunmi Santa Clara University, USA
Organizing Chairs
Ashim Saha NIT Agartala, India
Nirmalya Kar NIT Agartala, India
Suman Deb NIT Agartala, India
Additional Reviewers
Computational Intelligence
Security
Role of Open Hardware and IoT in Home Automation: A Case Study . . . . . . 251
Anurag De, Ashim Saha, and Praveen Kumar
6LoWSD: A Protocol Platform for Interoperability Between SDN and IoT. . . 273
Wanbanker Khongbuh and Goutam Saha
Contents xiii
Extended Track
1 Introduction
Human body is a very complex system and brain control it by secreting various
types of hormone, pancreas is one which secret insulin and its very important
for the body as it restrains the sugar levels in the body. In diabetic condition
body produces less insulin then the required quantity which trigger the damage
to various body parts and the first to get hit is eye. According to a study number
TIT.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 3–15, 2020.
https://doi.org/10.1007/978-981-15-3666-3_1
4 A. R. Choudhury et al.
of people across the globe suffering from diabetic will rise to double in coming
years and India is already on the top of the list having the largest population of
diabetic patients in the world [1]. Also, the dominance of this ailment has already
grown two times in the precedent 30 years [2] and is only expected to enhance
in the Asian region [3]. Among the diabetic population, about 33% are likely to
be spotted with diabetic retinopathy (DR), an ailment of eye which is chronic
and can step forward to irreversible loss in vision [4]. After duration of 15 years
10% of diabetic patients face visual impairment including diabetic retinopathy
[5]. According to UN report, India was the 140th country in happiness index
[6] and 15% of the population is going through any sort of mental illness [7].
India is one the top 3 country where corporate going people are not happy
about their working hours and do not play much sport. All this makes up a
perfect recipe for diabetic boom. Diabetic retinopathy is a condition in which
a person’s eyes sight gets affected because of internal damage of blood vessels
inside retina. In diabetic retinopathy the main reason for the damage of eye
sight is leaking of blood and other waxy substances inside retina which obstruct
the light from entering. Another phenomenon is fat concentration around fovea,
all these lead to obstruction of light and refection of light from optical disk to
macula. When a person is being a diabetic patient should go for regular eye
checkups since late detection can lead to even complete blindness. According to
the international monitoring agencies the problem of eye sight deterioration is
very common among the population of both developing and developed world.
The main problem is inadequate health care infrastructure in country like India
where a large number of people are not even entitled to basic healthcare. In the
young population the risk is high because of modern life style and negligence.
To ensure good quality of life and ensured economic growth the population
must be provided with good health care facilities, so it is mandatory to develop
systems and put up adequate infrastructure to deal with this situation in future.
As a developing nation India must achieve all these with limited resources and
frugal innovation. Diabetic retinopathy detection system proposed in this paper
is one of such initiative that will be backed by image processing and deep learn-
ing using existing CNN architecture with crucial modifications. The pro-posed
system will be trained up to a stage where it will be capable of identifying good
and affected retina. The concept that has been proposed for the detection of
diabetic retinopathy detection can also be replicated in other eye related prob-
lems. The proposed methodology is applied on EyePACS [8] LLC dataset and
obtained 83.32% on training set and 74% classification accuracy on testing set in
least possible time that shows the robustness of the methodology. The remain-
ing paper is organized as follows: Sect. 2 demonstrates the previous work, while
Sect. 3 provides necessary details of data used and methodology. Experimen-
tal details with evaluations are shown in Sect. 4 and finally Sect. 5 presents the
concluding remarks.
An Integrated IP and DL Approach for DRC 5
2 Literature Survey
There are presently many developed systems to detect and diagnose diabetic
retinopathy and most of these existing systems which are based on supervised
learning classifier are discussed in this section:
Previous works on image processing techniques have achieved good results in
identification of one specific feature used in the detection of this ailment [9–11].
For other features, algorithms involving KNN, [12,13] support vector machines
(SVM) [14], and ensemble-based system [15] successfully achieved sensitivity and
specificity of 90% range, but involves huge preprocessing.
Previous CNN studies also achieved promising results in terms of sensitiv-
ity and specificity for binary classification on private datasets [16,17]. Rahim
et al. [18] utilized binary decision tree, K nearest algorithm, Support vector
and random forest, all these are self teaching system and they had done feature
extraction using MATLAB and SPSS. These systems are incapable of process-
ing big size and complex data thus bringing down the accuracy. The proposed
system does not require these steps as CNN classifier is used which follows rein-
forcement learning process and robust enough to handle big size and complex
data. Verma et al. [19] implemented technique of hybrid classification which is
designed using neural network and fuzzy logical system. In the proposed tech-
nique, texture, shape and statistical feature extraction was carried out which is
time consuming.
Fig. 1. Number of image used for classification of retinal images with symptoms of
diabetic retinopathy and healthy
The data set containing 956 images is loaded in the system for classification. In
this system, an image will be either classified as normal or infected. The dataset
on load is shown in Fig. 3.
An Integrated IP and DL Approach for DRC 7
Fig. 3. Original fundus image of retina (a), (b) with no symptoms of diabetic retinopa-
thy damage, (c), (d) with symptoms of diabetic retinopathy damage
All the images used for the training and test cases were of higher resolutions and
because of RGB configuration the complexity was worse so all got converted to
grey scale format as shown in Fig. 4 for better efficiency. These images were also
put to normalization for keeping the range of the images from 0.0–1.0. using the
formulae,
X = X/255.0 (1)
As the images in the EyePACS data set are of higher dimensions, it is resized and
downsampled to 244 × 244 pixel that significantly reduced the required memory
size preserving the intricate features of the image. The downsampled image is
represented in Fig. 5.
The existing CNN architecture with certain modification as a part of deep learn-
ing method has been utilized and there are ten layers in proposed modified CNN
architecture and shown in Fig. 6.
Architecture of CNN
The model is a simple Sequential model and the window size is 3 × 3. The differ-
ent layers such as Dense, Dropout, Activation, Flatten, Conv2D, MaxPooling2D
are used. At the end of all the layers a fully connected Dense layer is used. Flat-
ten layer is used before feeding data through the final Dense layer because the
Dense layer accepts one dimensional data. The Conv2D layer used in the model
8 A. R. Choudhury et al.
is of 64 units. The Activation layer used is rectified linear (relu). Pool size of the
MaxPooling layer used is 2 × 2. Batch size is determined based on the number
of inputs as the numbers of images are 956 so, a batch size of 32 is suitable. The
detailed architecture of the model is shown in Fig. 7. Over fitting is one of the
most common issues in neural networks and it can be solved by reducing the
complexity of the neural network, i.e., by changing the values of weights and the
number of weights.
Epochs
Epoch in neural network expressions is one forward pass and one backward pass
of all the training set. Total epochs describe the number of times the algorithm
will work through the complete dataset of training. In every epoch the training
samples has a prospect to revise the parameters of model.
Training
Initially, the CNN was trained on 60% of total dataset to facilitate the train-
ing to reach a significant level. The training is essential for achieving a fast
classification outcome with no wastage of training time. The entire dataset was
repeatedly trained using 2, 5, 10, 15, 20, 25 epochs. The learning rate was kept
low for 2 and 5 epochs, later it was increased after the weights were stabilized.
Epochs while training the model were quite fast as the images were converted
to grayscale and resized to 224 × 224. One epoch took approximately 30 s with a
time of 51 ms/step which is very fast when compared to a RGB image. The total
training time for 15 epochs was roughly 10 min. In the beginning, 413 images
were taken which gave accuracy less than 60%. So the number of images were
increased which took the accuracy over 70%.
Testing
For testing 40% of the total number of images are taken in the validation set
which has given pretty good results. The images can be tested by loading this
saved model (model.h2). Saving the model is necessary as the weights and the
topology of the model need to be saved for future testing. The model is not ‘json
serializable’ it can be made ‘json serializable’ according to the desire of the user.
A single image can also be tested by the help of a prediction function which
takes the file path as an argument and returns a Boolean value i.e., TRUE or
FALSE. It also prints the category of the image. It can also print the index value
(0/1). It is completely based on the choice of the user.
10 A. R. Choudhury et al.
4 Results
This section shows the illustration of the proposed Diabetic Retinopathy Detec-
tion system that will examine the images and give prognosis of the eye. And
the whole prediction process will be based on the diabetic retinopathy detection
module which is the most important module that comes after data resizing of
various fundus images. The system will predict severity of the eye with at least
1000 to 2000 or more fundus images. For the prediction process the image data
sets were taken from EyePACS where the data sets are available with the severity
level numbers. We have segregated the image data set into two sub image data
set namely healthy and infected. Convolution Neural Network classifier tech-
nique has been used for classification of the image data. After analyzing of the
images it has been observed that the system gives different accuracy for different
epoch. The highest accuracy given by our proposed system is all most 83.32%
for 25 epoch during training phase and during testing phase of the machine it
has shown accuracy rate near to 74% for 25 epoch. The number of epoch here
means how many times the system is trained or tested with a train dataset and
validation dataset respectively. The details of Epoch are shown in Fig. 8.
12 A. R. Choudhury et al.
4.1 Prediction
The model is to be loaded before testing it. Testing can be performed on a single
or multiple images. In this figure testing of a single image is shown. An image
can be tested by calling the function which requires the file path of the image
directory (for multiple images) and for any single image it takes the name of the
image as an argument. It prints the output as integer value or categorical value
as shown in the code below:
The proposed system was implemented on Windows GPU environment with intel
CPU (i5, 2.2 GHz) using Python, Keras and Tensorflow. The average training
time for 956 images for each epoch is 30 s.
An Integrated IP and DL Approach for DRC 13
Fig. 9. (a) Epoch vs accuracy (b) Epoch vs loss of final model (model.h2)
5 Conclusions
Development in medical sciences is very important for tackling the problems of
human health issues and for this Information technology had played an impor-
tant role in last few decades for providing better procedures and innovative
ideas for mitigation of these problems. Currently most of the advancements in
the field of medical science because of advancements in the technologies of image
processing, machine learning and artificial intelligence. If we look into our past
the situation in the medieval period and the 20th century was tough in predicting
damages of internal organs. Prediction of diseases used to be done manually with
very less use of Holistic data but with advancement in data collection electronics
and more better computer programs which help us in gathering more and more
brut information about human internal organs, that helps in predicting more
accurately about any future health condition and also help us in segregating
and finding out the main cause which help in dealing with it more effectively.
While doing this work many project papers are explored and observed that a lot
of research is already going on across the globe on medical image analysis with
the help of artificial intelligence and machine learning and advanced electronic
devices. Doctors are using 3D modelling for orthopaedic surgeries and finding
out tumours and other anomalies in human body.
In future the existing system can be used to classify the input fundus images
in four classes with different stage of damage and healthy with better accuracy.
There can be a web based portal where any person who have fundus images of
retina can check for signs of diabetic retinopathy no matter from which corner
of world. This system can be used as an application on any device it can be a
pc or a smart phone with the help of a fundus image capture device can work
as any other bulky system.
In this paper, full concentration is on a very small topic diabetic retinopathy,
whereas there are many more eye related issues other than DR which may happen
because of diabetic condition. The goal is to automate the process with a slick
system in order to make it more cost effective and less complex to integrate.
The system would assist ophthalmologists to perform suitable classification by
monitoring the requisite result for early diagnosis.
References
1. http://www.ncbi.nlm.nih.gov/pmc/articles/PMC3920109
2. WHO. Global report on diabetes (2016)
3. Chan, J.C., et al.: Diabetes in Asia: epidemiology, risk factors, and pathophysiology.
JAMA 301(20), 2129–2140 (2009)
4. Congdon, N.G., Friedman, D.S., Lietman, T.: Important causes of visual impair-
ment in the world today. JAMA 290(15), 2057–2060 (2003)
An Integrated IP and DL Approach for DRC 15
5. Verma, K., Deep, P., Ramakrishnan, A.G.: Detection and classification of diabetic
retinopathy using retinal images. In: 2011 Annual IEEE India Conference, Hyder-
abad, India. https://doi.org/10.1109/INDCON.2011.6139346
6. https://m.hindustantimes.com/india-news/indian-ranked-140th-among-lowest-
20nations-on-un-world-happiness-report/
7. National Mental Health Survey of India, 2015–2016, Supported by Ministry of
Health and Family Welfare Government of India. http://www.nimhans.ac.in
8. Cuadros, J., Bresnick, G.: EyePACS: an adaptable telemedicine system for diabetic
reti-nopathy screening. J. Diab. Sci. Technol. 3(3), 509–516 (2009)
9. Mookiah, M., Acharya, U., Chua, C., Lim, C., Ng, E., Laude, A.: Computer-aided
diagnosis of diabetic retinopathy: a review. Comput. Biolo. Med. 43(12), 2136–
2155 (2013)
10. Niemeijer, M., et al.: Retinopathy online challenge: auto-matic detection of microa-
neurysms in digital color fundus photographs. IEEE Trans. Med. Imaging 29(1),
185–195 (2010)
11. Wang, S., Tang, H.L., Hu, Y., Sanei, S., Saleh, G.M., Peto, T., et al.: Localizing
microaneurysms in fundus images through singular spectrum analysis. IEEE Trans.
Biomed. Eng. 64(5), 990–1002 (2017)
12. Abramoff, M.D., et al.: Automated early detection of diabetic retinopathy. Oph-
thalmology 117(6), 1147–1154 (2010)
13. Quellec, G., Lamard, M., Josselin, P.M., Cazuguel, G., Cochener, B., Roux, C.:
Optimal wavelet transform for the detection of microaneurysms in retina pho-
tographs. IEEE Trans. Med. Imaging 27(9), 1230–1241 (2008)
14. Noronha, K., Acharya, U.R.: Decision support system for diabetic retinopathy
using discrete wavelet transform. Proc. Inst. Mech. Eng. Part H J. Eng. Med.
227(3), 251–261 (2013)
15. Antal, B., Hajdu, A.: An ensemble-based system for microaneurysm detection and
dia-betic retinopathy grading. IEEE Trans. Biomed. Eng. 59(6), 1720–1726 (2012)
16. Gulshan, V., et al.: Development and validation of a deep learning algorithm for
detection of diabetic retinopathy in retinal fundus photographs. JAMA 316(22),
2402–2410 (2016)
17. Gargeya, R., Leng, T.: Automated identification of diabetic retinopathy using deep
learning. Ophthalmology 124(7), 962–969 (2017)
18. Rahim, S.S., Palade, V., Shuttleworth, J., Jayne, C.: Automatic screening and
classification of diabetic retinopathy fundus images. In: Mladenov, V., Jayne, C.,
Iliadis, L. (eds.) EANN 2014. CCIS, vol. 459, pp. 113–122. Springer, Cham (2014).
https://doi.org/10.1007/978-3-319-11071-4 11
19. Verma, K., Deep, P., Ramakrishnan, A.G.: Detection and Classification of Diabetic
Retinopathy Using Retinal Images
Early Prediction of Brain Tumor
Classification Using Convolution
Neural Networks
1 Introduction
The report released by the World Health Organization (WHO) classifies the
tumor grade by four types where grade I to grade IV and tumor types are benign
and malignant. The class III and IV are high-level grade tumors which grow very
fast, and it spreads to other parts of the brain and spinal cords while grade I and
II are low-level grade tumors. Detecting the tumors and classifying at an early
stage is the challenging task in medical science, which helps to give treatments
the for patients at the beginning stage and decides for suitable therapy or surgery
can be determined.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 16–25, 2020.
https://doi.org/10.1007/978-981-15-3666-3_2
Early Prediction of Brain Tumor 17
2 Related Work
The MRI image is initially pre-processed to remove signal noise and balance for
variations in intensity brain tissues [1]. The pre-processed image is then separated
into different regions of interest (ROI) - areas with a likely tumor. The segmented
image extracts various types of features. The number of extracted features is
large, which in the following stage increases processing time and storage needs.
Therefore, it retains only the most useful features [2]. The characteristics will
then be sent to the classification process to determine if they contain a tumor
or not. Classification performance is then evaluated for accuracy and processing
time over all these images.
18 J. Arunnehru et al.
The segmentation techniques like Fuzzy C Means (FCM) and Active Contour
Model are adopted as there are no image sets with ground truths [3]. The spatial
FCM algorithm [4] is used due to its sensitivity. Among the other available
segmentation methods, the mathematical morphological process is the mainly
efficient method for extracting tumor area more accurately with less processing
time [5]. Different studies use Gabor wavelet features, but these features are
partial to capturing only local MRI image structures such as regularity, location,
and direction. Similarly, many authors suggested using statistical features of
first-order, co-occurrence gray level matrix features, and matrix features of the
gray level run length [6].
In recent deep learning methods, CNNs are actively used in object clas-
sification challenges, which increase their popularity among computer vision
researchers [7]. In connection to conventional classification approaches, where
manually extracted features are fed into the classifier for training and testing,
CNNs learn low, mid, and high-level features automatically from the previous
layers. In our work, CNNs architecture design mainly focus on brain tumor
identification rather than conventional feature extraction using image process-
ing approaches. In [8], 3D CNN architecture proposed for classifying the multi-
model MRI glioma region detection. Initially, the input image is divided into 3D
patches and are called voxels cuboid, obtained from different MRI brain sam-
ples, where 3D sections consist of spatial intensity information. The methods
like multi-model properties (T1, T2, and Flair) working on pixel classification
[9] by semi-supervised, supervised, and unsupervised.
In general, the unsupervised approach applies to cluster for finding the simi-
larity distance measure between features and voxels. However, these approaches
do not consider the spatial correlation. In addition to that, semi-supervised
approaches like Markov Random Fields (MRF) give province to this problem
by diminishing the voxel noises [10,11]. Gering et al. [12] present a framework to
detect the brain tumor using multi-layer MRF, whereas, in this framework, each
layer targets the intensity information on inter-structure properties and relation-
ships for justifying precinct coherence. In [13], authors present Convolutional
Neural Networks (CNN) architecture for tumor segmentation using supervise
pixel classification [14] that can extract more non-linear features without para-
metric distribution hypothesis. In addition to that, the authors [7] uses voxel
patch extracted from the image with multi-channel intensity to CNN for classi-
fication. Havaei et al. [15] designed a double stream layered networks, to obtain
more contextual features from the input patches, which output feature maps
connected to the second network simultaneously.
3 Proposed Methods
In the initial step for the tumor classification, the input MRI image preprocessed
with a median filter of 3 × 3. After preprocessing the fuzzy c-means clustering,
edge detection, and spectral residual applied for tumor region segmentation.
Finally, we fuse all three regions with three channels of the RGB image into
Early Prediction of Brain Tumor 19
one representation for CNN training and testing on a large dataset called T1-
weighted CE-MRI. The following subsections explain about fuzzy c-means clus-
tering algorithm, edge detection, and spectral residual for saliency mapping and
CNN for classification. Figure 1 gives an overview of the proposed approach.
The simplification of the output takes place in the pooling layer, thereby the
number of parameters can be reduced. After the operation of the feature learning
process, the classification takes place. After the output layer, the next layer
obtains the output of the number of classes that are being present.
The CNN architecture contains the final layer as the classification layer to
provide the classification output. The CNN training from scratch or pre-trained
model can be performed using the Matlab. The process of training the system
from scratch requires the number of layers and filters, along with the necessary
parameters. The training to be accurate it should contain a large dataset, and
it consumes more time.
CNN derived from the visual cortex, which contains the arrangement of cells,
both the simple and complex cells. The working of the cell depends on the sub-
regions(receptive field). Similarly, in the case of CNN, the neurons are connected
with the sub-regions of the layer. The neurons become unresponsive once it is
out of the sub-regions in the image. Only CNN can produce spatially correlated
output. The neurons arranged in 3D manner for every cell. For the 3D model
as the input, the Convolution layer produces the 3D output, whereas, in the
hidden layer, it learns the non-linear combinations of the original data. This
process is known as feature extraction. Figure 2 gives a detailed overview, and
Table 1 contains the CNN model parameters for training the CNN.
4 Experimental Results
The proposed model was trained and tested on a system Intel i7 processor with
8GB RAM and NVIDIA 1050 Ti GPU with 4 GB of memory. MRI dataset con-
sists of T1 - 1416, T2 - 1426, and T3 - 1483 images were collected from internet
sources. The CNN feature extraction is the fastest way to use the representa-
tional power of trained deep networks. In our proposed CNN network has four
convolutional layers and one fully connected layer, as shown in Fig. 2 and Table 1
contains the CNN model parameters for training the CNN. In our experiment,
the input MRI image dataset divided into 70% training and 30% for test data.
Early Prediction of Brain Tumor 23
Epoch Iteration Time elapsed Mini-batch Validation Mini-batch Validation Base learning
(hh:mm:ss) accuracy accuracy loss loss rate
1 1 0:00:14 35.16% 39.85% 1.2186 4.0171 0.01
2 30 0:01:57 68.75% 74.94% 1.2068 0.9769 0.01
3 50 0:03:08 83.59% 77.28% 0.4608 0.6521 0.01
4 70 0:04:18 87.50% 81.51% 0.3302 0.4417 0.01
5 110 0:06:38 91.41% 86.57% 0.2069 0.3172 0.01
6 130 0:07:47 88.28% 86.19% 0.2568 0.3256 0.01
7 150 0:08:54 86.72% 85.74% 0.2995 0.3401 0.01
8 170 0:10:00 89.84% 89.13% 0.2387 0.2665 0.01
9 190 0:11:08 95.31% 87.92% 0.1976 0.3092 0.01
10 210 0:12:17 94.53% 89.58% 0.1793 0.2569 0.01
11 240 0:13:57 92.97% 90.04% 0.1505 0.2539 0.01
12 260 0:15:05 96.09% 90.79% 0.1359 0.2324 0.01
13 280 0:16:11 92.97% 90.72% 0.1945 0.2247 0.01
14 300 0:17:18 97.66% 90.34% 0.1051 0.2291 0.01
15 330 0:18:59 97.66% 91.85% 0.1259 0.2137 0.01
16 350 0:20:06 99.22% 91.92% 0.0672 0.2102 0.01
17 370 0:21:13 96.09% 91.09% 0.1456 0.2619 0.01
18 400 0:22:55 97.66% 91.85% 0.0828 0.2044 0.01
19 420 0:24:01 98.44% 92.53% 0.0827 0.1905 0.01
20 440 0:25:08 96.88% 91.40% 0.0951 0.2179 0.01
24 J. Arunnehru et al.
Class T1 T2 T3
T1 94.65 4.94 0.4
T2 5.91 93.79 0.3
T3 0.58 0.1 99.33
The CNN learns image features from MRI dataset, and use those features to
train an image classifier. The CNN requires input images of size 128 × 128 × 3.
Images are normalized. Table 2 shows the performance values obtained at each
epoch. The total time taken by the training phase of the MRI tumor classifica-
tion system is 25 min 8 s. We had a total of 20 epochs with 22 iterations per
epoch. The total numbers of iterations were 440 iterations with a learning rate
of 0.01, which was constant. This experiment resulted in a validation accuracy
of 91.40% and the corresponding confusion matrix of the CNN shown in Table 3
where correct predictions defined in the diagonal and the tumor classes like T3 is
recognized well and reached an accuracy of 99%. Also, T2 and T1 was a misclas-
sification rate of 4% to 5%. The experimental results show that CNN performs
well in tumor classification.
5 Conclusion
In this research work, we have used brain MR images, classified into brain tumors
types like meningioma (T1), glioma (T2), and pituitary tumor (T3). Next, we
have used the fuzzy c-means algorithm for finding cancer surrounding tissues
followed by canny edge detection and spectral residual for saliency map for
tumor region identification. Convolution Neural Networks (CNNs) used for the
classification of tumors from brain MRI images. From the experimental results,
it clearly expressed that the classification accuracy reported as 91.40% of brain
tumor types and also shows that our proposed approach is fast and accurate
when compared to the manual detection carried out by clinical experts.
References
1. Aslam, A., Khan, E., Beg, M.S.: Improved edge detection algorithm for brain tumor
segmentation. Procedia Comput. Sci. 58, 430–437 (2015)
2. Shanthakumar, P., Ganesh Kumar, P.: Computer aided brain tumor detection
system using watershed segmentation techniques. Int. J. Imaging Syst. Technol.
25(4), 297–301 (2015)
3. Natteshan, N.V.S., Angel Arul Jothi, J.: Automatic classification of brain MRI
images using SVM and neural network classifiers. In: El-Alfy, E.-S.M., Thampi,
S.M., Takagi, H., Piramuthu, S., Hanne, T. (eds.) Advances in Intelligent Infor-
matics. AISC, vol. 320, pp. 19–30. Springer, Cham (2015). https://doi.org/10.
1007/978-3-319-11218-3 3
Early Prediction of Brain Tumor 25
4. Dandıl, E., Çakıroğlu, M., Ekşi, Z.: Computer-aided diagnosis of malign and benign
brain tumors on MR images. In: Bogdanova, A.M., Gjorgjevikj, D. (eds.) ICT
Innovations 2014. AISC, vol. 311, pp. 157–166. Springer, Cham (2015). https://
doi.org/10.1007/978-3-319-09879-1 16
5. Yazdani, S., Yusof, R., Karimian, A., Riazi, A.H., Bennamoun, M.: A unified frame-
work for brain segmentation in MR images. Comput. Math. Methods Med. 2015,
829893 (2015)
6. Nabizadeh, N., Kubat, M.: Brain tumors detection and segmentation in MR
images: Gabor wavelet vs. statistical features. Comput. Electr. Eng. 45, 286–301
(2015)
7. Zikic, D., Ioannou, Y., Brown, M. and Criminisi, A.: Segmentation of brain tumor
tissues with convolutional neural networks. In: Proceedings MICCAI-BRATS, pp.
36–39 (2014)
8. Urban, G., Bendszus, M., Hamprecht, F., Kleesiek, J.: Multi-modal brain tumor
segmentation using deep convolutional neural networks. MICCAI BraTS (Brain
Tumor Segmentation) Challenge. Proceedings, winning contribution, pp. 31–35
(2014)
9. Zhang, L., Han, Y., Yang, Y., Song, M., Yan, S., Tian, Q.: Discovering discrimina-
tive graphlets for aerial image categories recognition. IEEE Trans. Image Process.
22(12), 5071–5084 (2013)
10. Han, J., Zhang, D., Cheng, G., Liu, N., Xu, D.: Advanced deep-learning techniques
for salient and category-specific object detection: a survey. IEEE Signal Process.
Mag. 35(1), 84–100 (2018)
11. Menze, B.H., van Leemput, K., Lashkari, D., Weber, M.-A., Ayache, N., Golland,
P.: A generative model for brain tumor segmentation in multi-modal images. In:
Jiang, T., Navab, N., Pluim, J.P.W., Viergever, M.A. (eds.) MICCAI 2010, Part
II. LNCS, vol. 6362, pp. 151–159. Springer, Heidelberg (2010). https://doi.org/10.
1007/978-3-642-15745-5 19
12. Gering, D.T., Grimson, W.E.L., Kikinis, R.: Recognizing deviations from normalcy
for brain tumor segmentation. In: Dohi, T., Kikinis, R. (eds.) MICCAI 2002, Part
I. LNCS, vol. 2488, pp. 388–395. Springer, Heidelberg (2002). https://doi.org/10.
1007/3-540-45786-0 48
13. LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to
document recognition. Proc. IEEE 86(11), 2278–2324 (1998)
14. Bengio, Y.: Practical recommendations for gradient-based training of deep archi-
tectures. In: Montavon, G., Orr, G.B., Müller, K.-R. (eds.) Neural Networks: Tricks
of the Trade. LNCS, vol. 7700, pp. 437–478. Springer, Heidelberg (2012). https://
doi.org/10.1007/978-3-642-35289-8 26
15. Havaei, M., et al.: Brain tumor segmentation with deep neural networks. Med.
Image Anal. 35, 18–31 (2017)
16. Zhang, L., Tong, M.H., Marks, T.K., Shan, H., Cottrell, G.W.: SUN: a Bayesian
framework for saliency using natural statistics. J. Vis. 8(7), 32–32 (2008)
17. Rajkumar, R., Arunnehru, J.: A study on convolutional neural networks with active
video tubelets for object detection and classification. In: Wang, J., Reddy, G.R.M.,
Prasad, V.K., Reddy, V.S. (eds.) Soft Computing and Signal Processing. AISC, vol.
898, pp. 107–115. Springer, Singapore (2019). https://doi.org/10.1007/978-981-13-
3393-4 12
18. Arunnehru, J., Chamundeeswari, G., Bharathi, S.P.: Human action recognition
using 3D convolutional neural networks with 3D motion cuboids in surveillance
videos. Procedia Comput. Sci. 133, 471–477 (2018)
Routine Statistical Framework
to Speculate Kannada Lip Reading
1 Introduction
Artificial Intelligence plays a very important role in many core areas like robotics,
machine learning, as it involves significant aspects of imparting the information
to systems. Autonomous systems is another core area that includes driving of
vehicles from source destination with the help of speech recognition without any
assistance of drivers [2,3,5,12]. These core areas defines the need of artificially
intelligent system that does the task of reading of lips. The spoken language
may be English or Spanish or Dutch or any Indian language. These languages
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 26–38, 2020.
https://doi.org/10.1007/978-981-15-3666-3_3
Routine Statistical Framework to Speculate Kannada Lip Reading 27
2 Related Work
Typically, there are some challenges with automating lip-reading process [20],
including the co-articulation effect, visual features diversity and speaker-
dependency of features. The history of these issues along with some proposed
solutions for them could be found in the literature. Regarding to the co-
articulation effect and visual feature diversity, the following studies have reached
to noticeable solutions. In [14,16,21] have demonstrated the effects of inter-
speaker variability of features could be reduced by applying the feature improve-
ment of per speaker z-score normalization and the hierarchical LDA (HiLDA)
techniques.Using HMM recognizer [18,19,22], the best viseme accuracies of 45.
In [1,15] the best viewing angle for an automated lip-reading system is stud-
ied, using a proposed built audio-visual speech database LiLiR. The database
contains multi-camera, multi-angle recordings of a speaker reciting 200 sentences
from the Resource Management Corpus, with the vocabulary size of 1000 words.
Visual features extracted in this work are based on the Active Appearance
Models, which includes the shape, appearance, a simple concatenation of the
shape and appearance parameters, and an application of PCA [13,17] over con-
catenation and HiLDA variations.
3 Datasets
The videos of a dataset is considered for features extraction by exercising various
simple statistical parameters resulting in appropriate accuracy. The statistical
methods have been used to essentially obtain the features like analysis of variance
and other strategies that are essentially required for calculating the features from
various tilted and frontal faces of small deaf and dumb child.
The dataset has been chosen with all challenging issues to address in frames
of a dataset. The frames of a video is extracted and analyzed to detect the facial
features like lip, mouth and teeth as a way of reference to extract statistical
featured information from every frames of a video as shown in Fig. 2.
The challenging issues like frontal face, tilted face are used to check the effi-
cacy of the erected method with various statistical features that are required for
features extraction and features classification. So that the assessment of effort
with different statistical models shall be checked. Some of the statistical cri-
teria explored are Analysis of variance, chi-square test, factor analysis, Mann
Whitney-U, mean square weighted deviation, and Pearson product moment cor-
relation coefficient, regression analysis. The spearman rank correlation coeffi-
cients are some of the statistics used to measure the efficacy of the present effort
with respect to features useful for achieving the desired task and calculating the
features classification.
4 Proposed Methods
The effortless statistical feature based representation learning towards recog-
nition of lip shapes has led system determine the words spoken by a person.
Routine Statistical Framework to Speculate Kannada Lip Reading 29
Fig. 2. Data set used with statistical approach after detecting the mouth
Analysis of Variance. The statistical models are collected together for analysis of
variance is based on certain principles like finding the mean of n different values
and determining the variance. We have noticed that the determined variance is
partitioned into different sources of variation. Further, it provides a statistical
test of whether two models and its population means are equal and generalizes
the test those two models are beyond two means. It employs data decomposition.
In other words, its sum of squares indicate the variance of each component of
the decomposition. It allows testing of nested sequence of models to achieve
comparison of mean squares.
with the pattern that would be expected. The predicted values of variables
against the existing one together forms the system predict the characteristic
features of independent variables.
(f0 − fc )2
x2 = (1)
fe
In Eq. 1 shows chi-square statistic based on predicted versus existence or ground
truth values. The independent variables together constitutes that the proposed
system is relying on completely independent variable also shows the similarity
in values of one variable with respect to the other.
Factor Analysis. The process of expressing the observed against the existing val-
ues and explored to reduce the large number of independent variables into fewer
number of factors. It extracts maximum variance among common independent
variables and keeps them into a form of score. A part of general linear model,
predicts several observations like linear relationships and multi co-linearity, and
includes relevant variables into analysis. If there is a true co-relation between
variables then, several factors are available but principal component analysis is
used as a reference to determine the relations among independent variables.
Where the term μ the mean of all items of features, and LF corresponds to lower
bound of the features set, ε indicates the + or − values of bounding errors.
The process also involves certain principles of converting the several similar
values of grouped variables into few variable with common features eliminated
and keeping few factors that are responsible for estimation of factors with data
to be predicted.
Mean Square Weighted Deviation (MSWD). The method relies on three different
values between <1 and >1.
N N
W wi .(xi − x̄i )2
M SW D = N i=1 N (6)
i=1 − i=1 Wi i=1
2 (σxi )
By referring to Eq. 2, MSWD = 1, the data fits into univariate normal dis-
tribution, MSWD <1, the data is said to be undispersed, MSWD >1 indicates
as over dispersed. The arithmetic mean is first determined and then standard
deviation calculated to perform the task of values that ranges between −1 and
>1. The mean of the values of considered from all data variables are squared to
obtain the weighted deviation that fall into the values of deviation between <1
and >1.
The PPMCC also relies on the values of variables that may supersede the sys-
tem with correlation existence with other system coefficients. Since, the system
makes use of certain principles like determining the values of qualitative analysis
and its effectiveness towards achieving the strength of correlation coefficient.
Regression Analysis. The powerful statistical routine has able to determine the
relationship between two or more variables that may be of interest. Here the
dependent variables are under the influence of one or more independent variables.
These independent variables are more interested towards achieving the values of
dependent values of variables formed from it. The practice makes use of certain
32 M. S. Nandini et al.
Y = θ1 + θ2 .X (8)
The statistical perusal of results of various parameters shall be verified with the
presence of different numerical study that is more powerful towards determining
the effectiveness of independent and dependent variables. These variables are
more likely to be used for analysis of true versus false positive and various other
metrics are too considered for evaluation of the system.
Results of Precision versus Recall. The precision is obtaining the true positive
of predicted results with respect to sum of True positive and false positive rate.
TPR
P recission = (10)
TPR + FPR
TPR
Recall = (11)
TPR + FPR
Routine Statistical Framework to Speculate Kannada Lip Reading 33
Table 1. Indicates the precision versus recall results produced from different statistical
manner.
It is clear from the Eqs. (10) and (11) that the statistical enumerations
are assessed with discrete modus operandi. Further based on true positive rate
yielded from predicted results with reference to false prediction results from
enlarged statistical framework (Table 1).
In addition to other, the statistical parameter have not yielded much opti-
mized results.
It is evident from Fig. 3, the precision and recall of different statistical pro-
cesses have shown their best of ability to detect the shape of lips or extract
much detailed information from lips in term of true positive and false positive
rate. These metrics together with other false negative rate also defines the recall
strategy.
Results of Sensitivity versus Specificity. The sensitivity and specificity are two
additional metrics plays a very important role in accurately determining the fea-
tures to be used for prediction of Lip reading. The performance of the statistical
approach has been assessed with metrics enforced visibility are shown in Table 2.
34 M. S. Nandini et al.
By employing the sensitivity and specificity mentioned in Eqs. (12) and (13)
respectively.
TPR
Sensitivity = (12)
TPR + FPR
The sensitivity and specificity defines the metric used to assess the proposed
statistical methods. Overall, the Statistical methods have given better informa-
tion as to how the frames is to be processed. The result of processing of informa-
tion using different statistical methods yielded better outcome while mapping
the lip shapes with annotated values at the time of testing.
TNR
Specif icity = (13)
TNR + FPR
It is evident from the Eqs. (11) and (12) respectively that the sensitivity is a
metric defines the statistical features in terms of true positive rate versus sum of
true positive rate and false negative rate. Further, the specificity is determined
by employing Eq. (13) defined by true negative rate versus sum of true negative
rate and false positive rate. The graphical representation is shown in Fig. 4.
The Eq. (14) determines the accuracy of Lip Reading strategy. The perfor-
mance has been verified with accuracy based on Eq. (14) and the results of it
are shown in Table 3.
36 M. S. Nandini et al.
The accuracy is another metric used to determine the most accurate values of
predicting the Lip Reading using sundry statistical machine learning strategies
(Fig. 5).
6 Conclusion
The research work has attempted to contribute towards predicting the lip move-
ment by analyzing the features of simple statistical boundaries and imposed
certain metric as a rough yardsticks towards predicting the correctness of var-
ious statistical method that are very much helpful in predicting the language
spoken by a person. The sensitivity and specificity of the proposed statistical
methods have yielded 41.21% and 46.19% respectively. The accuracy of these
statistical methods have been presented with 40.18%. As statistical methods
are computationally incapable to produce good performance, hence the outcome
accuracy is 40.18%. The performance can be improved with hybrid techniques.
The effort made here to signify the sophomore rate of outcomes. Further, the
attempt made with the effortless and manageable parameters as a rehearsal to
reveal lip reading accomplishment.
Routine Statistical Framework to Speculate Kannada Lip Reading 37
References
1. Abadi, M., et al.: TensorFlow: large-scale machine learning on heterogeneous dis-
tributed systems. arXiv preprint arXiv:1603.04467 (2016)
2. Assael, Y.M., Shillingford, B., Whitestone, S., de Freitas, N.: LipNet: sentence-level
lipreading. arXiv:1611.01599 (2016)
3. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning
to align and translate. In: Proceedings of ICLR (2015)
4. Bengio, S., Vinyals, O., Jaitly, N., Shazeer, N.: Scheduled sampling for sequence
prediction with recurrent neural networks. In: Advances in Neural Information
Processing Systems, pp. 1171–1179 (2015)
5. Chan, W., Jaitly, N., Le, Q.V., Vinyals, O.: Listen, attend and spell. arXiv preprint
arXiv:1508.01211 (2015)
6. Chatfield, K., Simonyan, K., Vedaldi, A., Zisserman, A.: Return of the devil in the
details: delving deep into convolutional nets. In: Proceedings of BMVC (2014)
7. Chorowski, J., Bahdanau, D., Cho, K., Bengio, Y.: End-to-end continuous speech
recognition using attention-based recurrent NN: first results. ArXiv Preprint
arXiv: 1412.1602 (2014)
8. Chorowski, J.K., Bahdanau, D., Serdyuk, D., Cho, K., Bengio, Y.: Attention-based
models for speech recognition. In: Advances in Neural Information Processing Sys-
tems, pp. 577–585 (2015)
9. Chung, J.S., Zisserman, A.: Lip reading in the wild. In: Lai, S.-H., Lepetit, V.,
Nishino, K., Sato, Y. (eds.) ACCV 2016. LNCS, vol. 10112, pp. 87–103. Springer,
Cham (2017). https://doi.org/10.1007/978-3-319-54184-6 6
10. Chung, J.S., Zisserman, A.: Out of time: automated lip sync in the wild. In:
Chen, C.-S., Lu, J., Ma, K.-K. (eds.) ACCV 2016. LNCS, vol. 10117, pp. 251–
263. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-54427-4 19
11. Cooke, M., Barker, J., Cunningham, S., Shao, X.: An audio-visual corpus for speech
perception and automatic speech recognition. J. Acoust. Soc. Am. 120(5), 2421–
2424 (2006)
12. Czyzewski, A., Kostek, B., Bratoszewski, P., Kotus, J., Szykulski, M.: An Audio-
visual corpus for multimodal auto-matic speech recognition. J. Intell. Inform. Syst.
49, 167–192 (2017)
13. Feichtenhofer, C., Pinz, A., Zisserman, A.: Convolutional two-stream network
fusion for video action recognition. In: Proceedings of CVPR (2016)
14. Galatas, G., Potamianos, G., Makedon, F.: Audio-visual speech recognition incor-
porating facial depth information captured by the Kinect. In: 2012 Proceedings
of the 20th European Signal Processing Conference (EUSIPCO), pp. 2714–2717.
IEEE (2012)
15. Graves, A., Fernández, S., Gomez, F., Schmidhuber, J.: Connectionist temporal
classification: labelling unsegmented sequence data with recurrent neural networks.
In: Proceedings of the 23rd International Conference on Machine Learning, pp.
369–376. ACM (2006)
16. Graves, A., Jaitly, N.: Towards end-to-end speech recognition with recurrent neu-
ral networks. In: Proceedings of the 31st International Conference on Machine
Learning (ICML-14), pp. 1764–1772 (2014)
17. Graves, A., Jaitly, N., Mohamed, A.-R.: Hybrid speech recognition with deep bidi-
rectional LSTM. In: 2013 IEEE Workshop on Automatic Speech Recognition and
Understanding (ASRU), pp. 273–278. IEEE (2013)
38 M. S. Nandini et al.
18. Hermansky, H.: Perceptual linear predictive (PLP) analysis of speech. J. Acoust.
Soc. Am. 87(4), 1738–1752 (1990)
19. Kazemi, V., Sullivan, J.: One millisecond face alignment with an ensemble of regres-
sion trees. In: Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 1867–1874 (2014)
20. King, D.E.: Dlib-ml: a machine learning toolkit. J. Mach. Learn. Res. 10, 1755–
1758 (2009)
21. Koller, O., Ney, H., Bowden, R.: Deep learning of mouth shapes for sign language.
In: Proceedings of the IEEE International Conference on Computer Vision Work-
shops, pp. 85–91 (2015)
22. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep con-
volutional neural networks. In: NIPS, pp. 1106–1114 (2012)
Incepting on Language Structures
with Phonological and Corpus Analysis
Using Multilingual Computing
1 Introduction
Languages spoken by the human beings are extinguishing at rates and it is affect-
ing the global bio-diversity as the disappearance of languages also dissociate us
from the dimensions of those cultures which are associated with those lost lan-
guages. According to the world report, there are about 7000 languages in the
world are in use today and it is expecting to have 3000 languages in next 100
years. The most challenging thing is that the 48% of the world’s population are
first language speakers of one of the world’s top 10 most spoken languages: Stan-
dard Chinese, English, German, Spanish, Portuguese, Russian, Bengali, Hindi,
Arabic and Japanese. The most alarming situation of the language extinguishing
is that out of 10000 languages before 100 years already 3000 languages are extin-
guished as it is forgotten by people. Now out of 7000 languages 3500 languages
are spoken by a little group of people consisting of around 10000 and around
2100 languages are spoken by a mere 1000 people [4]. Therefore it is our utmost
duty to make some way to preserve such languages. This study focuses on dis-
cussing some parameters of language computing and also gives the theoretical
idea on grammar for the major languages spoken in the world which will give
for further researches on this language preservation.
1.2 Purpose
This work deals with the location based study and an attempt is made to find the
variations of spoken languages of the different regions spread over a range of lon-
gitude and latitude of the world. In our study we had taken languages: along the
longitude-India (initial), Uzbekistan (immediate up from India), Mongolia (next to
Uzbekistan), Russia (next to Mongolia) and Sri-Lanka down from India. The lan-
guages along the latitude are Saudi Arabia left of India and China, right of India.
(i) Indian Languages: All Indian languages have marker as they are rooted
to Sanskrit. Those languages that have influence of Chinese or Tibetan lan-
guage are usually behaved like no-marked languages.
(a) Arunachal Pradesh, a state situated in the north-east part of India and
bordered by neighboring country China. So its language is influenced by
Chinese language. Having 30 mostly popular and 50 distinct languages
Arunachal Pradesh uses the language named Tani that belongs to Sino-
Tibetan family. The peculiarity of this particular language is the missing
of cognate and overload of reconstructed roots. In respect of vocabulary
the phonemic strategy of this language is completely depended on its
phonetic structure.
(b) Assam, a state is situated in the south east of Arunachal Pradesh and
bordered by Bangladesh. The language Assamese is the official native lan-
guage of this state and it is highly influenced by Silet dialect of northern
region of Bangladesh and Mising, Tani languages of Arunachal Pradesh.
With the effect of isolated phonology and morphology the Assamese lan-
guage has more the 50% single-lettered suffixes. Such as, ė, o, ni etc.
42 B. Choudhury and A. Das
the verb represents the past tense then the suffix has been inflecting with
the verb. In the past tense suffix –u is inflected by replacing the stem
vowel “a” of the present finite verb. For example baeluvā means looked
instead of balanna. Denna in present form of verb root “de” means give.
But in the past form suffix -nā is inflected with the verb root and makes
new word Dunnā.
(e) Characteristics of Chinese Language: Chinese language is the only exam-
ple of root or isolating classification of language. In such languages word
has been decoding by its position in a sentence. Thus a word can be sub-
ject, predicate as well as verb also. The intension of the word is depended
on the tonal verification of the speaker and its usage in a particular sen-
tence. And so it is called analytic language. In Chinese language there is
no separate formation or conjugation of verb according to the tense. The
only way to intend past, present, future is to add “today”, “yesterday”
and “tomorrow” before the verb. For example, “eat” is used as “chi” in
all the forms of tenses. Zuótiān wǒ chı̄le Ork means in English Yesterday
I ate pork. Pitch intonation is therefore a primary equipment to deci-
pher the intension of morphemes. There are five major tone structures in
Standard Chinese language. Every Chinese character has different tone.
Amongst them first four are basic pitch tones and fifth is neutral tone.
It is as followed as first dā is high and level, second dá starts medium
in tone, then rises to the top, third dǎ starts low, dips to the bottom,
then rises toward the top, fourth dà – starts at the top, then falls sharp
and strong to the bottom, and neutral da – flat, with no emphasis. Exam-
ple, ‘tāng’ with a high tone means ‘soup’, but ‘táng’ with a rising tone
means ‘sugar’. ‘Āi’ with a high tone means ‘hey’, but ‘ǎi’ with curvy low
to rising high tone means ‘short and ‘ài’ with sharp fall tone from top to
bottom means ‘love’. Similarly ‘bā’ means ‘eight’, ‘lā’ means ‘pull’, ‘bǎwò’
means ‘grasp’ and ‘Fùqı̄n ’ means ‘father’.
The languages of different parts of the world are different in acoustic features
with the variation of phonetics, phonemes etc. The proposed model LCFG is
designed with the following features: phonological analysis, corpus generation
and text-mining. It measures the following metrics like: phonetic distance, phono-
logical distance, combined distance and also generate the corpus of the languages.
The phonetic analysis is done by applying the phonetic distance, phonological
distance, combined distance, the corpus generation is done by applying an easy
mechanism with the help of the huge repository of web. Finally find some per-
plexity of some corpus of the languages (Sanskrit, English, and Russian): The
below diagram shows the flow of the work of LCFG.
Incepting on Language Structures with Phonological and Corpus Analysis 45
Fig. 3. The wave format of the EL, RL and SL of the sentence I EAT RICE three
times.
This phase of the model works out for finding the phonetic distances as vectors,
phonological distances of characters (with two categories: vowels and consonants)
and finally blends the two to find the resultant effect in the repository. Let us
take an example to explain how we derive the phonetic and phonological analysis
from a sentence. The sentence in English “I eat rice” is spoken in Sanskrit
and Russian as: “aham annang khaadaami” and “ya yem ris”. The sentences
are now taken as the strings by eliminating the white spaces in between the
words. Now the sentences in the three languages will be looked as the following
string: “ieatrice”, “ahamannangkhaadaami” and “yayemris”. Now we may get
more than one phonetic strings in this process but for more precision we will
take the larger strings. Thus finally we take the strings as: “(ai)eatr(ai)ce”,
“ahamannangkhaadaami” and “y(aa)y(a)mr(ii)s”. The following table gives the
details (Fig. 3).
Sentence Corpus: In general definition, the corpus is a collection of written or
spoken material stored in a computer and they are used to find the usage of a
language. Thus they are the units stored in a computer to use for a language [6].
The corpus may be stored for a single language or two or more languages. If
it stores for one language it is termed as monolingual and for two or more
it is known as multilingual corpus. Multilingual corpora are used for side-by-
side comparison which is called aligned parallel corpora. The corpora can be
prepared for linguistic research by applying the annotation process like: PoS
Incepting on Language Structures with Phonological and Corpus Analysis 47
Table 2. languages with its vowel and consonants used in the strings
Table 3. The Uni-Gram (UG) and Bi-Gram (BG) of the vowels and consonants
(part of speech) tagging, lemma (base), interlinear glossing etc. (Figs. 4, 5 and
Tables 1, 2, 3, 4).
Incepting on Language Structures with Phonological and Corpus Analysis 49
Table 4. The weights calculated of UG and BG of vowels and consonants of the strings
N
N
N
P (S) = (P (s1s2.s3, ....swm ) − 1/N ) = p(si) 1/N
= 2log2 p(si)1/N
i=1 i=1 i=1
N
m
= 2 − 1/N log2 p(si) = 2−1/N p(si )log2 p(x) (6)
i=1 x
50 B. Choudhury and A. Das
Table 5. The frequency wise occurrences of words in EL, SL and RL from randomly
selected text
10% in EL, SL and RL. More than frequency 2, there is no word found in SL,
whereas in EL and RL some words are present. Interestingly there are two words
in EL, viz. “to” and “the” the frequencies are 10 and 21 respectively (Table 5
and Fig. 7).
The generation corpus of a language is really a tedious task. But here one
easy mechanism is shown for generating the corpus of any language. The steps
are:
(i) By using AI-application web-crawler the searching is done from www, and
it returns the list of all possible websites which contains the desired texts,
for our test we searched Sanskrit, English and Russian. The application
downloads the information of each language into a specific method.
(ii) The text collected from step-1 is stored according to fonts. The information
will be stored as per the different folder names to get an easy access for fur-
ther process. Next all the language texts are converted to its corresponding
UNICODE by applying the converter.
(iii) We can use Google search engine for searching by setting the keywords
from the test set languages like: Sanskrit, English, and Russian etc. The
searched documents are then stored in the selected folders local drives of
the computer.
(iv) Then the proper storing in the machine is done for generating the corpus
of the language. Thus the corpus of the language is prepared.
3 Findings
From the above study the following observations are found:
(i) In phonetic and phonological distances, though the Sanskrit language has
less number of vowels still this language produces maximum number of BG
(Bi-Gram) strings. Thus Sanskrit is more phonetically and phonologically
strong language than the other two languages.
(ii) Sanskrit Language is free from ordering of words.
(iii) In Sanskrit, the utterance and the phonetic and phonological strings are
same as that of the actual string of the sentence, but in case of other two
languages some changes are observed.
(iv) A minute observation of UG and BG would show that if any language
is much closer to its proto or primitive family language then its Mora of
morpheme would behave likewise. Here East Slavic existing language Rus-
sian is much closer to Indo-European primitive language Sanskrit. Thus
the audio graph of this language shows some similarities in morphological
Mora with Sanskrit rather than English. Because as per antiquity English
is much younger than Russian and Sanskrit.
(v) Features of characteristics of languages those discussed in this paper would
denote the facts that not only borrow words from different languages
would change the structural periphery of a particular language but char-
acter of neighbouring language also makes some noticeable variations in
that language.
52 B. Choudhury and A. Das
References
1. Katre, S.M.: Astadhyayi of Panini. Motilal Banarasidas, Delhi (2015). (2nd Edition
of 1st edition 1989)
2. Chatterjee, S.K.: The Origin and Development of the Bengali Language. Rupa Pub-
lication India Pvt. Ltd., New Delhi (2011). 4th Impression of, 1st edn. 1926
3. Das, A., Choudhury, B.: A study on the process of Supra segmental preservation
of IE Language and computational data analysis of acoustic phonetics. J. Emerg.
Technol. Innov. Res. 7(3), 898–903 (2019). E-ISSN 2347–2693
4. Das, A., Choudhury, B.: A study on language computations by preserving them as
audio documents. Int. J. Comput. Sci. Eng. 5(8), 209–214 (2018). ISSN 2349-5162
5. Das, B., Mandal, S., Mitra, P.: Bengali speech corpus for continuous speech recog-
nition. IEEE (2011)
6. Sarma, S.K., et al.: A structured approach for building Assamese corpus: insights
applications and challenges. In: Proceedings of the 10th Workshop on Asian Lan-
guage Resources, COLING 2012, Mumbai, December 2012 (2012)
7. Baruah, B.K.: History of Assamese Literature, 1st edn. Sahitya Akademy Publica-
tion, New Delhi (1984)
Morphotactics of Manipuri Verbs:
A Finite State Approach
1 Introduction
Manipuri is a Tibeto-Burman language as well as a Scheduled Language of India
mainly spoken in the state of Manipur and other states like Assam, Tripura,
Mizoram, Nagaland and other countries like Bangladesh and Myanmar. Accord-
ing to [1] in the classification of T.B. languages Manipuri language comes under
the Kuki-Chin subgroup of the Burmese division. Shafer [2] places Manipuri in
the Meitei branch of Kukish section while [3] puts Manipuri in Kuki-Naga of
Kukish section. It is the medium of instruction in schools, colleges, universities,
governmental and non-governmental organizations and institutes. However, the
status of the language in the field of Natural Language Processing is not so
advanced as compared to other languages like Hindi, Bengali, Tamil and Kan-
nada. This is because of the fact that the language is highly agglutinating and
morphologically complex. A chain of string can be concatenated to derive a new
word which creates ambiguity in determining the Part of Speech of the given
word. Manipuri has extensive verb morphology and extensive suffixes with lim-
ited prefixation [4]. There are some works done in Manipuri verb but the setting
of morphotactics rule is not discussed so far or is not utilized. The paper is
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 53–63, 2020.
https://doi.org/10.1007/978-981-15-3666-3_5
54 H. K. Devi et al.
the introductory step towards because it deals with the designation of various
automata setting from the various rules. Testing the result is not covered but
saved for later work.
1.1 Objective
2 Morphotactics
The word morphotactics is formed by the combination of two words; morph and
tactics which means the possible combinations of morphemes. In other word, it
can be defined as the permissible and non-permissible combination of morpheme.
Morphotactics is one of the most useful techniques used by the linguists as well
as the computer scientists for developing the morphological analyzer of the word.
As for instance, the suffix -ity produces a noun from an adjective (capable >
capability), and -able creates adjective from verbs (eat > eatable). The reversed
order may violate the morphotactics of the language. All the languages of the
world have morphotactic rules or constraints. For Manipuri, it has two negative
allomorph suffixes /-te/ and /-de/. /-te/ is used after voiceless ending verbal root
(k@k ‘cut’+ te ‘NEG’ > k@kte ‘not cut’) and /-de/ after voiced ending sound such
as ca ‘eat’+ de ‘NEG’ > cade ‘didnt eat’.
Manipuri verbs cover 80% of vocabulary and they are bound in nature. As stated
by [7] that words in Manipuri can consists of stems or bound root with suffixes
from one to ten. For instance, the word ca-rm–d-b-ni-ko-ne, where ca is the
Morphotactics of Manipuri Verbs: A Finite State Approach 55
bound root and the rest are the suffixes. It is mandatory to identify all the
verbal roots available in the language along with the affixes. David [8] defines a
root as the base form of a word which cannot be further analysed without total
loss of identity. Negations are also derived from the verbs. As for instance, nt
(Not)+ -te (NEG) > nt.te (not true) and na (sick)+ -de (NEG) > na.de (not
ill). -te∼-de are negative marker added to the verbal roots to form negation. -te
is used after /t/, /k/ and /p/ sounds which are all voiceless and -de is used after
the following verbal roots ending with N, n, m, a, a, i, u sounds.
Rule 1: VR + -te∼-de > Negative
For instance,
jeN+de > jeNde ‘didnt see’
mun+de > munde ‘not ripe’
ph @m+de > ph @mde ‘not sit’
na+de > nade ‘not ill’
irué@+de > irué@de ‘not taken bath’
i+de > ide ‘not writen’
u+de > ude ‘not seen’
Q0 Q1
Q2
Start Root word -te ∼ -de
The Fig. 1 shows the Transition diagrams of Rule 1. This Rule is consisted
of 3 states. State Q0 is the starting state(intial state), on getting root word as a
input the FSA move from Q0 to Q1 state. On getting the input ∼ -te,∼ -de as
a input the FSA accept the string and identify as the Verb. Q2 is the final state
in the FSA (Tables 1 and 2).
Exception: There are only three verbal roots which take –te to make the
negative even though the verbal roots are ended with voiced sounds such as hau,
l@i, h@i as shown below: verbal roots ended with voiced sound + -te suffix to
form the negation.
56 H. K. Devi et al.
Manipuri verbs are not inflected for number, person, gender and tense (or aspect
for Manipuri).
For instance,
The above three examples have been excerpted from the thesis done by Pre-
mabati [9]. The verb ca-i remains in all the three sentences.
Rule 2: If ∼p@ ∼b@ are added to these above dynamic verbal roots, it becomes a
derived noun. The Fig. 2 shows the Transition diagram of the Rule 2. This Rule
is Transition diagram using the 3 states. State Q0 is the starting state(intial
state), on getting root word as a input the FSA move from Q0 to Q1 state. On
getting the input p@ b@as a input the FSA accept the string and identify as the
Noun. Q2 is the final state. For example:
Q0 Q1
Q2
Start Root word
Rule 3: But if we want to derive noun again from the negative verb, -te -de
changes into -t@∼ d@with the addition of a nominalizer -b@as shown below: The
Fig. 3 shows the Transition diagram of the Rule 3. This Rule is demonstrated
using the 4 states. State 0 is the starting state(intial state), on getting verbal
root as a input the FSA move from Q0 to Q1 state. On getting the input t@ d@the
FSA move from state Q1 to the state 2(Example 2). On getting nominalizer as
input from state Q2 or from state Q1 the FSA move to the final Step. This
accept the string which is derived noun (Tables 3 and 4).
Nominalizer(NZM)
Q0 Q1 Q2 Q3
Start NZM
Verbal Root
The various Stative bound roots are given below: Cao-, waN,-nem, -saN, -pak,
ph @, -mot, -nan, -mu, -sin, -ka etc. Nouns can be derived from these verbal roots
by suffixing nominalizer -p@/-b@as shown below:
Verbal root + Nominalizer > Derived Noun
Cao + -b@> caob@ ‘big/ name of a person’
waN + -b@> waN b@ ‘tall’
saN + -b@> saNb@ ‘long’ pak + -p@> pakp@ ‘broad’
ph @+ -b@> ph @b@ ‘good’
Nominalizer(NZM)
Q0 Q1 Q2 Q3 Q4
Start NZM
Verbal Root
Rule 4: From these derived nouns as shown above adjectives are derived by
prefixing @-. This is for the case where the verbal root is monosyllabic in nature.
The Fig. 4 shows the Transition diagram of the Rule 4. This Rule is represented
using the 5 states. State Q0 is the starting state, on getting @∼ as a input the
FSA move from Q0 to Q1 state. when followed by the verbal root the FSA look
for the rule 3. the final state in the FSA is Q4. By going through the state Q0-
Q1-Q2-Q4 (Rule 4.1), Q0-Q1-Q2-Q3-Q4 (Rule 4.2) (Table 5).
For instance,
N.B. Adjectives in Manipuri are mostly derived from stative verbs than
dynamic verbs.
Rule: 4.1 If the verbal root is disyllabic, only nominalizer is added followed by
a noun to modify the quality of the noun such as
nuNsi (2 VR) + b@(NMZ) is@i > nuNsib@is@i (sweet song)
So the syntax is Compound + NMZ > Adj
Rule 4.2: However, there are exceptional cases which is after suffixing the neg-
ative marker t@∼ d@ to the verbal root followed by the nominalizer -p@∼-b@, the
prefixation of @- is no more required to form the adjectives. For example,
*@+ca (VR)+d@(NM)+ b@(NMZ) > @cad@b@is not a valid word.
Rule 4.3: Again the prefix /@/ is not required if the verbal root is immediately
followed by an aspect marker and a nominalizer as shown below
Ca (VR) + ri (SAM) + b@(NMZ) cak > carib@cak (the food which is eating)
Rule 5: Again adverbs are derived from verbs with the addition of suffix -n@to
the verbal root as shown below:
Morphotactics of Manipuri Verbs: A Finite State Approach 61
Q0 Q1 Q2
Start Verbal Root
The Fig. 5 shows the Transition diagram of the Rule 5. This Rule is imple-
mented using the 3 states. State Q0 is the starting state(intial state), on getting
verb as a input the FSA move from Q0 to Q1 state. On getting ∼n@as input the
FSA accept the string and identify as the Adverb. VR + -n@(suffix) > Adverb
i.e. t@m (slow) + n@> t@mn@ ‘slowly’
ca (to eat) + n@> can@ ‘by eating’
Rule 5.1: But in order to express the adverb of manner, the verb will be redu-
plicated. There are two ways to express it. One is the complete reduplication
and the other is the partial reduplication. That is (i) t@m (VR) +n@+ t@m (VR)
+ n@> t@mn@t@mn@ ‘slowly slowly’
And, (ii) t@m (VR) + t@m (VR) + n@> t@m t@mn@ ‘slowly’ (Table 6).
The first one is the case for complete reduplication and the second one for
partial reduplication.
Now, the finite-State Machine (FSM) is constructed to represent the con-
struction of the negation, adjective, adverb and noun to identify each path that
takes the machine to a final state from its initial state as shown below: The
Fig. 6 shows the combination of all above 5 Rules. This FSA accept Verb, noun,
Q0 Q1 Q2 Q3 Q4
Start NZM
Verbal Root
Fig. 6. Left to Right Transition diagram of FSA of Negation, adjective, adverb & noun
from verb
62 H. K. Devi et al.
adjective and adverb. State Q4 represent the final state(Verb), adjective, derived
noun and the adverb. There is two transition from Q0 to Q1 that is via epislon
or ∼@.
The morphological rules for the derivation of negation, adjective, adverb and
derived nouns are presented by using finite- state techniques. A hand crafted
rule automaton has been developed to implement the system in order to verify
the validity of verbs in Manipuri. Since verb covers the major part of Manipuri
Vocabularies. Hence, it is not possible to list down and identify all the verbs
available in the language. Again all the verbs in Manipuri are bound in nature.
Further it can be studied into two ways - Structurally and semantically [9]. For
this a huge amount of corpus is required. The present paper is restricted to the
formation of negation, adjective, adverb and derived noun by means of affixation.
References
1. Grierson, G.A.: Linguistic Survey of India, Part III, vol. III, p. 8. Motilal Banarasi-
das, Delhi (1927). (Reprinted 1967)
2. Shafer, R.: Classification of the Sino-Tibetan languages. Word 11, 94–111 (1955)
3. Benedict, P.K.: Sino-Tibetan: A Conspectus, pp. 4–11. Cambridge University
Press, Cambridge (1972)
4. Singh, C.Y.: Manipuri Grammar, pp. 31–55. Rajesh Publications, New Delhi (2000)
5. Jurafsky, D., Martin, J.H.: Speech and Language Processing, p. 16. Pearson Edu-
cation, Delhi (2000)
6. Kaplan, R.M., Kay, M.: Regular models of phonological rule systems. Comput.
Linguist. 20(3), 331–337 (1994)
7. Chelliah, S.L.: A Grammar of Meithei, p. 26. Mouton de Gruyter, New York (1997)
8. Crystal, D.: A Dictionary of Linguistics and Phonology, p. 267. Basil Blackwell,
Oxford (1985)
9. Premabati. The Syntax of Manipuri Verbs. Unpublished Ph.D Thesis, Dept. Lin-
guistics, Manipur University, p. 105 (2009)
10. Singha, K.K.B., Singh, K.R., Purkayastha, B.S.: Morphotactics of Manipuri adjec-
tives: a finite state approach. Int. J. Inf. Technol. Comput. Sci., 96 (2013). https://
doi.org/10.5815/ijitcs.2013.09.10
11. Abbi, A.: Reduplication in South Asian Languages: An Areal, Typological and
Historical Study, p. 12. Allied Publishers Ltd., New Delhi (1992)
Proportional Analysis of Feature
Extraction for Tamil Handwritten
Characters Using Centroid Based Method
1 Introduction
issues limit the accuracy of OHR systems. The proposed method is based on
feature extraction technique to recognize offline handwritten Tamil characters.
It’s a Dravidian and official language of Tamil Nadu belongs to south India
and also widely used in Pudhucherry, Singapore and Srilanka [5]. The structure
of language is complex due to its cursive shapes and loops. The Tamil script
embodies 12 vowels, 18 consonants and 1 special character named ayutha and
an additional 216 characters its derived from vowels and consonants all together
is about 247 characters.
2 Literature Survey
Raj and Abirami [8] proposed 8 directional Chain Code and Zone based Features
approach, calculate their area is additional feature used Support Vector Machine
classifier to achieve 80% accuracy for Tamil Handwritten Character.
Raj and Abirami [7] proposed a Zone based Statistical Features approach
and Eight directional Chain Code for each zone Axis based, pixel location And
Point counts, Used Support Vector Machine classifier to achieve 89% accuracy
for Offline Tamil Handwritten Character.
Rajashekararadhya [10] presented a Zone and Distance metric based feature
extraction method for the recognition of Kannada and Telugu numerals. Feed
forward back propagation neural network is used for classification and recogni-
tion with a recognition rate of 98% for isolated handwritten Kannada numerals
and 96% for isolated handwritten Telugu numerals.
Sangame [4] proposed a system for unconstrained handwritten Kannada vow-
els recognition based upon invariant moments. A Euclidian distance criterion and
K-NN classifier is used for classification of the Kannada vowels. The system had
overall accuracy of 85.53%.
Rajashekararadhya [9] proposed an offline handwritten numeral recognition
technique for four south Indian languages, based on zone and image centroid used
two different classifiers nearest neighbor and back propagation neural network
to achieve 99% accuracy for Kannada and Telugu, 96% for Tamil and 95% for
Malayalam.
The existing methods are based on 8 directional chain code, distance metric and
zone based the accuracy found for offline Tamil Character from Literature survey
was 80% and 89%. In order to increase the accuracy the proposed algorithm is
implemented based on Euclidean distance measure, Manhattan distance measure
and Chess board distance measure for Tamil vowels character to extract feature,
the method uses 8 directional chain code for feature selection.
66 D. S. kala and M. P. I. Gandhi
the horizontal and vertical, distance between two point etc. In the proposed work
8 directional chain code is used for feature selection and centroid based distance
measure is considered to extract features, we use the methods such as Euclidean
distance measure, Manhattan distance measure, Chessboard distance measure
to extract features [6]. Finally training of extracted features for recognition of
characters is given as input to a neural network for further classification, the
algorithm to extract feature is given below:
Input: Characters
Output: Classification
This work is used to calculate the distance from the centroid to the pixels
in each direction using Euclidian distance formula, Manhattan Distance formula
and Chessboard Distance formula. The image is rotated in different angle 0◦ ,
90◦ , 180◦ and 270◦ the image is rotated to prove that it is invariant to rotation.
The above steps are repeated for each degree. Our feature vector consists of 16
features to recognize the character.
6 Experimental Results
The classification plays a special role in OHR the extracted feature are given
as an input to the neural network. An artificial neural network trained with
feed forward algorithm is used here to classify and recognize the handwritten
characters.
A multilayer neural network classifier comprises of three layer: hidden layer,
input and output layer. The hidden layer uses log sigmoid, bias is found both in
output and hidden layer. Input layer is connected to output layer through hidden
layer and output layer is connected to the output. They are connected by means
of weight whose output is always one. In this work the total of neurons in the
output layer and input layer is 3 and 16 respectively. Figure 2 shows Multilayer
neural network.
This work uses MATLAB R2015a as tool for implementing the algorithm
and designed using neural network, this data stored in Microsoft excel and fed
in MATLAB, where the data was loaded in Neural Network function to get the
accuracy rate for different distance measure. The experiment result and over-
all accuracy achieved for different character samples (100 * 3 = 300) are imple-
mented using Euclidean distance, Manhattan Distance and Chessboard distance
are respectively tabulated in the following Table 1.
A confusion matrix is often used to illustrate the performance of a classifier,
it is used to list correct classifier as true positives or true negatives, incorrect
classifier as false positive or false negatives. Error rate and accuracy are most
common predict, it also gives imminent into the types of errors made by classifier.
For the experiment a confusion matrix of 3 * 3 is considered.
Proportional Analysis of Feature Extraction 69
Table 1. Show the experimental results for different character and accuracy percentage
for different degree.
Accuracy (ACC): Total true Positive +true negative divide by Total positive
+Total negative.
T rue Positive Rate(TPR): Total true positive divide by Total true Positive
+ Total false negative.
F alse Negative Rate(FNR): Total false negative divide by Total false nega-
tive + Total true positive.
T rue Negative Rate (TNR): Total true negative divide by Total false Posi-
tive + Total true negative.
F alse Positive Rate (FPR): Total false positive divide by Total true negative
+ Total false positive.
70 D. S. kala and M. P. I. Gandhi
Confusion matrix shows the classification classes for the dataset, the value
given diagonally in 3 * 3 matrix is True positive rate for class 1, class 2, class 3
and the other rate (FN, FP, TN) are calculated based the Fig. 3. The average
of diagonal value gives the Accuracy of the each degree and Error Rate is (1-
Accuracy) the Fig. 4 shows the confusion matrix for different distance measure
and Rotation. The image is rotated at an angle of 90◦ to prove that algorithm
is invariant to rotation.
The TP rate is calculated for each class that is class 1, class 2 and class3
separately and the average TPR of all the classes is taken. Similarly the other
rates TNR, FPR and FNR are calculated. The best rate for TPR and TNR
is 1.0, whereas the worst rate is 0.0. The best rate for FNR and FPR is 0.0
and worst rate is 1.0. This means most positive classes are recognized correctly.
Table 2 show the Performance Rate of Euclidean, Manhattan and chess board
distance.
Proportional Analysis of Feature Extraction 71
Fig. 4. Shows the representation of confusion matrix for rotation for (a) Euclidean
Distance rotation of 0◦ , 90◦ , 180◦ , 270◦ respectively (b) Manhattan Distance rotation
0◦ , 90◦ , 180◦ , 270◦ respectively. (c) Chess Board Distance rotation 00 , 900 , 1800 , 2700
respectively in Vertical wise
72 D. S. kala and M. P. I. Gandhi
Table 2. Shows the TPR, TNR ,FNR and FPR rates for distance measure
7 Conclusion
The proposed features are found to be effective for Tamil offline handwritten
characters. This work has been implemented using centroid based distance mea-
sure by choosing 3 letters from Tamil real character dataset which was generated
by us and compared the experimental results for Euclidean distance, Manhattan
distance and Chessboard distance. It was found that results comparable to the
best result of Chess board distance was 95.3%, 93.3%, 96.0% and 96.3% when
the image was rotated to different angles of 90◦ such as 0◦ , 90◦ , 180◦ , and 270◦
respectively. It is observed that rotation accuracy vary in Euclidean and Man-
hattan distance measure and better in chess board distance measure. The TNR
and TPR value of chess board distance is better than other distance measure
The error rate is basically due to unusual writing style, which arise haziness
among similar shaped characters. Future work our aim is to consider all Tamil
characters for recognition. Also hoping to develop method to find relative error
in terms of distance measure.
References
1. Jain, A.K., et al.: Feature extraction methods for character recognition-a survey.
Pattern Recogn. 29(4), 641–662 (1996)
2. Kudo, M., Sklansky, J.: Comparison of algorithms that select features for pattern
classifiers. Pattern Recogn. 33(1), 25–41 (2000)
3. Jagadeesh Kannan, R., Prabhakar, R., Suresh, R.M.: Offline cursive Tamil hand-
written recognition. In: IEEE International Conference on Security Technology,
vol. 4, pp. 159–164 (2008)
Proportional Analysis of Feature Extraction 73
1 Introduction
In the present era where virtualization and inter connectivity have covered the
world in multi-layer interaction, which prevails more in virtual than in actual.
Digital devices have more latent to enhance the performance in the classroom,
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 74–86, 2020.
https://doi.org/10.1007/978-981-15-3666-3_7
Mobile Supported Interaction Modeling to Find Engagement of Acolyte 75
[5] increasing the ability of the students, making the classroom more interac-
tive, engaged students in exciting ways. In the area of the online environment,
Problem-solving capability is also prolonged. Wireless devices i.e., mobile devices
stimulate learners to learn and maintain their interest in learning. Nowadays
learning through mobile devices at different education levels is more popular.
Mobile learning [9,12] field also refers to the use of mobile devices for learning
purposes. Mobile devices facilitate students to instant interaction with real con-
texts and also allow them to join in different learning activities. Several studies
[1,5,12] suggest that mobile learning gives a positive result in concentration, atti-
tudes, behavior, and learning achievements. On the other hand, many researchers
[18–20] give negative responses about mobile learning and they also suggest
that there is also a need for further investigations. Maintaining the attention of
learners during the learning period through a mobile device is also a big issue.
This study mainly focuses on the learning attention of college students who use
m-learning [9] concept in the classroom and also entail that correlation exist
between the students score and engagement.
2 Literature Review
Mobile devices can support the flexible and real-time learning environment. The
traditional classroom can be transformed into a new form by adapting the mobile
devices in the classroom. Different studies [1–3] find the actual use of student
response system(SRS) in classroom. By using mobile devices students can eas-
ily response to the question-answer and also asked questions to their teach-
ers. Thus SRS can improve classroom interactivity, and also motivate students.
Mobile devices used in the classroom do not mean students understand every-
thing or fully rising the attention of the students. But mobile devices had that
much potential that they can promote the active learning technique in classroom
Mobile technology types which make sense for the classroom are e-readers: In
the traditional classroom, students use textbooks for study purpose. But the
main issue about this fact is they are so swiftly outdated regarding the subject
matter. But e-readers eliminate that issue because they allow real-time updates
for the learners. Individual mobile modules: Educational games and apps allow
the user for individual log in. Students can take extra time according to their
need and it also gives the chance at their own speed. Seamless cloud learning:
Mobile technology allow students to direct connection with the cloud. This facil-
ity allows them to transition from working, it may be the classroom, home or
anywhere else. Thus time is saved and skill also developed.
Portability: Mobile phones are portable devices. For these mobility characteris-
tics, it can be moved or carried anytime and anywhere.
Screen size: Smartphones have smaller screens, which makes them pocketable
and convenient. But a smaller screen size can be used as a disadvantage since
phone screens put up less number of contents compared to a personal computer
or notebook.
Ram and processing power: These two core components helps to store program
and run program easily.
Accessories: S-pen is under the accessories part, it helps the user to create much
work easily which are usually takes more time. By using this work will has done
a very quick and friendly manner.
Software factors: Mobile device allows much educational software, which allows
students to do some educational work.
– In the traditional classroom, if the teacher didn’t give equal attention to all
the learners in the classroom then it will be the cause of the destruction of
their self-confidence. As a single person, it can’t be possible to give equal
attention all the time.
This study aims to investigate the effect of mobile phones in the classroom and
also try to find out the learning concentration of acolyte who use mobile phone
in classroom for learning purpose.
Item Description
(1) I try to give my best in mobile learning
(2) I can fully concentrate on my learning contents by ignoring the surrounding noises
(3) During mobile learning I lose my concentration
(4) I can easily concentrate on my topics in case of mobile learning
(5) In the mobile learning activity my mind wandered
(6) During mobile learning it is hard to maintain concentration after 25 min
(7) In mobile learning I can easily grip everything
(8) In the mobile learning activity I easily forgot what I learn or listen
5 Method
This method consists of participants and instruments. All the participants were
belongs to NIT Agartala campus and instrument contains m-learning devices.
5.1 Participants
analyzing their responses we find that both group had same output. So, by con-
sidering their result we can say that there is no difference between both the
group. This was our pre test.
After this survey we took post test. Assigned same question again but that
time smartphone was allowed to one of the group, which is our treatment group.
Another group also did the same task, but were not allowed to use smartphone.
After analyzing the result of the post test found that control group result was
poor compare to the treatment group (Tables 2 and 3).
F df p
Marks 5.751e−32 1 1.000
Group N Mean SD SE
Marks 1 45 65.022 14.274 2.128
2 45 65.022 14.274 2.128
5.2 Instruments
By some m-learning questions, questions on learning concentration and perform-
ing a pre-test and post-test this quasi-experiment is designed.
Post Test: Post test was conducted between the 90 students. All the students
are divided into the two groups one group used mobile phones and others are
not. Table 5. Shows the result of the post test.
Figures 4 and 5: shows the histograms of the students marks for group1
(Control group) and in group 2(Treatment group). Visually, This diagrams rep-
resent that Students in treatment group means who were use mobile phone in
Mobile Supported Interaction Modeling to Find Engagement of Acolyte 81
F df p
Marks 0.707 1 0.403
Group N Mean SD SE
Marks 1 45 65.022 14.274 2.128
2 45 66.911 12.845 1.915
classroom for study purpose getting better marks on average compare to Control
group who were not use mobile phone in classroom (Tables 6 and 7).
This (Fig. 6) plot diagram shows the mean result of the students who are in
Control Group as well as Treatment group. Error bars shows 95 % confidence
intervals approximately the mean value. Visually, it shows that there is a differ-
ence between the groups.
6 Motion Detection
By using smartphone motion can be detected. Accelerometer, the rotation-
sensing gyroscope can be used as the main tool for collecting data. This study
also used the sensor lab app for collecting data. Nowadays every student is having
a mobile phone and they also bring them in the classroom. We have examined
the data by using the mobile phone for some classroom activities as well as by
keeping it on the table. After analysis, those data key taps reveal that in which
direction and how much a smartphone rotates. Here the value “X” shows the
movement along the horizontal and “Y” Shows the vertical movement. Table 8
shows the data when the mobile phone was used for different classroom activi-
ties during the classes and Table 9, shows the data when the mobile phone was
kept on the table during the classes. It is clearly shown that while the phone
was used for doing classroom activity maximum rotation has happened. After
analyzing those value we conduct a controlled experiment, we ask participant
(NIT Agartala student) to perform different study-related activities during class
for five minutes time stamp during the class. A sensor lab app is installed in
the smartphone which was provided by us for data collection. The app runs in
the background of the smartphone for collecting the values of the three sensors(
accelerometer, gyroscope, and magnetometer). After that, a small test was taken
from the topic that was covered during the class. In which he scored 6 out of
10. The same experiment was again done but this time that smartphone was
kept in the table. Again test was taken from the covered topic, now this time he
scored 8 out of 10. All the collected values are manually labeled and analyzed in
the JASP 0.10.2.0 statistical tool. By analyzing the collected data observed that
their some distinct patterns for both the experiment as shown in Figs. 7 and 8.
Table 8. Data captured while mobile phone was used for educational activity
x y z
Valid 15300 15300 15300
Missing 0 0 0
Mean 0.005 −0.769 9.777
Std. Deviation 0.470 0.586 0.285
Minimum −9.596 −10.285 3.486
Maximum 6.522 11.042 16.290
Mobile Supported Interaction Modeling to Find Engagement of Acolyte 83
When the smartphone was placed in the table there is less movement happened
so the density of the x,y,z axis also less as compare to the graph which shows
the smartphone used as an activity.
The above observation affirms that it is possible to identify that there is a
relationship present between the motion and attention.
x y z
Valid 387 387 387
Missing 0 0 0
Mean 2.818 3.185 8.303
Std. Deviation 2.518 2.074 0.919
Minimum −4.434 −1.015 2.662
Maximum 7.604 8.092 13.005
7 Limitation
There is a few limitation in mobile learning concepts
84 B. Debnath et al.
– Screen size: Mobile devices have smaller screen size where student is required
to strain their eyes for long time in the concept of mobile learning.
– Cost: For mobile learning mobile devices are needed. As we know nowadays
most of the people carry mobile phone but still there are so many poor people
exist they can’t effort mobile phone since it is relatively expensive.
– Limited storage: Memory storage of mobile devices is not enough for students
to store all the classroom courses, offline data.
– Unexpected distraction: In the Mobile learning concept distraction can hap-
pen in many ways e.g., in the form of notification, call, messages etc.
– Short battery life: Battery life of mobile devices is very short. Since many
apps are active all the time in the mobile devices so, they discharge quickly.
This is one of the big limitations in mo-learning concept.
– Limited connectivity: Mobile devices have limited bandwidth; it may be create
difficulties to access online contents.
– Technological Problem: Mobile devices have different operating system and
different hardware. This may create difficulties in learning concept.
– Safety: One of the major problem is that safety for users. Preventing users
from cyber-bullying, personal bio hack, preventing them from the access of
illegal information in classroom is one of the key challenge.
8 Conclusion
This study investigate the interaction effects of learning interest and improve-
ment in learning of all the students who use their mobile phones for learning
purposes in the classroom. The major finding of this paper as follows: essential
Mobile Supported Interaction Modeling to Find Engagement of Acolyte 85
correlations present between the interest and concentration. This paper also find
the good result for mobile learning concept in the classroom. First, the learning
progress of the acolyte also measured by conducting the pretest and posttest.
Second, the learning concentration of the students also measured by the result
of the test. Third, it also shows that during the m-learning concept there maybe
a relationship present between the movement of the mobile phone and the con-
centration of the students. Actually concentration during study may differ in
various factors like educational content, motivational speech, superiority of the
course design, gender, formative assessment technique etc.
References
1. The Interaction of Pedagogical Approach, Gender, Self-Regulation, and Goal Ori-
entation Using Student Response System Technology. J. Res. Technol. Educ.
(2019). https://www.tandfonline.com/doi/abs/1
2. Gauci, S.A., et al.: Promoting student-centered active learning in lectures with
a personal response system. - PubMed - NCBI. https://www.ncbi.nlm.nih.gov/
pubmed/192617
3. Internationalization testing for a multilingual based wireless response system: a
case study - University of Huddersfield repository (2019). Accessed 2 Sept 2019
4. Kydd, C.T., Ferry, D.L.: Managerial use of video conferencing. Inf. Manage. 27(6),
369–375 (1994)
5. Using mobile phones in college classroom settings: effects of presentation mode
and interest on concentration and achievement - PDF Free Download (2019).
kundoc.com. https://kundoc.com/pdf-using-mobile-phones-in-college-classroom-
settings-effects-of-presentation-mode-a.html. Accessed 2 September 2019
6. Parhi, P., Karlson, A., Bederson, B.: Target size study for one-handed thumb use
on small touchscreen devices. In: Proceedings of the 8th Conference on Human-
Computer Interaction with Mobile Devices and Services - Mobilehci 2006 (2006)
https://doi.org/10.1145/1152215.1152260
7. Sanchez, C.A., Branaghan, R.: Turning to learn: screen orientation and reasoning
with small devices. Comput. Hum. Behav. 27, 793–797 (2011). https://doi.org/10.
1016/j.chb.2010.11.004
8. Molina, N.A., Mahan, R.P., Illingworth, A.J.: Establishing the measurement equiv-
alence of online selection assessments delivered on mobile versus nonmobile devices.
Int. J. Sel. Assess. 22(2), 124–138 (2014)
9. Garcı́a-Cabot, A., De-Marcos, L., Garcı́a-López, E.: An empirical study on m-
learning adaptation: learning performance and learning contexts. Comput. Educ.
82, 450–459 (2015). https://doi.org/10.1016/j.compedu.2014.12.007
10. Lee, H., Parsons, D., Kwon, G., Kim, J., Petrova, K., Jeong, E., Ryu, H.: Coop-
eration begins: encouraging critical thinking skills through cooperative reciprocity
using a mobile learning game. Comput. Educ. 97, 97–115 (2016). https://doi.org/
10.1016/j.compedu.2016.03.006
11. Shu, H., Gu, X.: Determining the differences between online and face-to-face
student-group interactions in a blended learning course. Int. High. Educ. 39, 13–21
(2018). https://doi.org/10.1016/j.iheduc.2018.05.003
12. Pedro, L.F.M.G., Barbosa, C.M.M.O., Santos, C.M.N.: A critical review of mobile
learning integration in formal educational contexts. Int. J. Educ. Technol. High.
Educ. 15(1), 1–15 (2018). https://doi.org/10.1186/s41239-018-0091-4
86 B. Debnath et al.
13. Heflin, H., Shewmaker, J., Nguyen, J.: Impact of mobile technology on student
attitudes, engagement, and learning. Comput. Educ. 107, 91–99 (2017). https://
doi.org/10.1016/j.compedu.2017.01.006
14. Balta, N., Perera-Rodrı́guez, V.-H., Hervás-Gómez, C.: Using socrative as an online
homework platform to increase students’ exam scores. Education and Information
Technologies 23(2), 837–850 (2017). https://doi.org/10.1007/s10639-017-9638-6
15. A user-friendly voting and quiz system for classroom use with connected devices.
In: IEEE Conference Publication (2019). https://ieeexplore.ieee.org/document/7.
Accessed 22 July 2019
16. Influence of Digital Equipment on Interaction Quality in Technology-rich
Classroom. In: IEEE Conference Publication (2019). https://ieeexplore.ieee.org/
document/7757022. Accessed 22 July 2019
17. Sung, Y., Chang, K., Liu, T.: The effects of integrating mobile devices with teach-
ing and learning on students’ learning performance: a meta-analysis and research
synthesis. Comput. Educ. 94, 252–275 (2016). https://doi.org/10.1016/j.compedu.
2015.11.008
18. Jumoke, S., Blessing, O.: Analysis of mobile phone impact on student aca-
demic performance in tertiary institution (2015). https://www.semanticscholar.
org/paper/Analysis-of-Mobile-Phone-Impact-on-Student-Academic-Jumoke-
Blessing/00ad8a3fdad62581f83d0a8cf46. Accessed 22 July 2019
19. Uğur, N., & Koç, T.: Mobile phones as distracting tools in the classroom: college
students perspective. Alphanumeric J. 3(2) (2015). https://doi.org/10.17093/aj.
2015.3.2.5000145549
20. Osei Frimpong, K., Asare, S., Otoo-Arthur, D.: The effects of mobile phone usage
on the academic performance of Ghanaian Students, a case of Presbyterian Uni-
versity College Asante-Akyem Campus (2016)
Contribution of Application of Deep
Learning Approaches on Biomedical Data
in the Diagnosis of Neurological
Disorders: A Review on Recent Findings
1 Introduction
The brain is the most vital part of the human body that receives input from the
sensory organs of the body and sends output signals to the muscles thereby act-
ing as the command center for the human body. Different regions of the brain are
spatially distributed but continuously share information with each other form-
ing a functionally connected network [36]. At present, people with neurological
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 87–97, 2020.
https://doi.org/10.1007/978-981-15-3666-3_8
88 W. Ahmad Mir et al.
2 Background
Symbolic approaches of Artificial Intelligence in medical health diagnosis can
be traced back to the 1980s. Although Artificial systems that can learn from
Contribution of Application of Deep Learning Approaches 89
the trial and error method had surfaced in the early days of the 1960s [31].
There has been a tremendous development in health data analysis using machine
learning. Accuracy rates of modern classifiers, noise filters, and feature selection
approaches have advanced beyond limits. The availability of systematic knowl-
edge sources, well defined, reasonably sized subdomains, efficient and powerful
processors are the reason behind this success [30]. The concept of inferring a
result after making a system to learn from data samples marks a branch of Arti-
ficial Intelligence known as Machine Learning [7,28]. This learning process may
be divided into two parts (a) from a known data set we estimate unknown depen-
dencies in the system and (b) based on the estimated dependencies; we predict
the new results from unknown data [22]. It is imperative to stress upon that
Machine Learning (ML) offers a highly sophisticated, computer-aided system of
analyzing high dimensional and multimodal biomedical data with the help of
developed and objectively oriented algorithms [32]. ML has shown interesting
results in the area of biomedical research. There are two commonly known ML
techniques (a) Supervised Learning and (b) Unsupervised learning. In case of
supervised learning, a function is derived that takes a specific input and pre-
dicts output values in close approximation to the desired output. A training
data set is taken as input, and an algorithm is designed to derive the required
function which produces output on unknown data to a close degree of fidelity to
the actual outcome. For the Unsupervised learning, the data set consists of only
input values, and an algorithm is designed to identify patterns, trends, and sim-
ilarities in the data thereby clustering the similar inputs [8,21]. A combination
of the supervised and unsupervised learning approaches widely applied is known
as semi-supervised learning. When unlabeled data sets outnumber the labeled
data sets, a more accurate learning model can be constructed with the help of
semi-supervised learning [22]. In the last few years, Deep Learning has helped
in improving the quality of diagnosis [38]. Physiological conditions concerning a
particular individual and the hidden relationships between a particular neural
disorder and the associated brain electrical activity can be more accurately pre-
dicted by the deep learning approaches. Deep learning has the capacity of faster,
reliable and accurate diagnosis and drifts from currently used decision support
systems to deep learning approach is on the cards, despite deep learning being in
its infancy. Deep Learning effectively combines the feature extraction and classi-
fication processes [34]. A robust convolutional neural network (CNN) model can
be developed after extracting features from the input data which improves the
diagnostic performance of the model during the testing phase [29].
3 Deep Learning
Deep learning methods are learning methods with multiple levels of representa-
tion that have become popular for effectively analyzing unstructured data [15].
These multiple levels are obtained by composing different nonlinear modules
that transform the representation from simpler to some abstract levels. These
transformations can be used in systems to learn very complex functions [23].
90 W. Ahmad Mir et al.
One of the important features of deep learning is that the features of the layers
are not designed by human engineers but are learned after following a general-
purpose learning process from the data itself. Deep Learning methods are com-
posed of multiple layers of neurons stacked together to produce discriminating
classes after proper feature selection procedure. Deep learning, a new trend in
machine learning has evolved as a de-facto approach for applications spread over
a vast spectrum of research areas like medical data analysis, social media analy-
sis, natural language processing, showing tremendous accuracy, robustness and
generalizability over the conventional approaches [11].
comparison to the other works published recently, the proposed model does not
provide the best of the performance. This model has managed to produce only
88.67% of accuracy, 95% of sensitivity and 90% specificity. The authors have
proposed a bagging algorithm and enhancement in sample space for improving
the performance metrics of the model.
Ullah, et al. [35] have proposed an automatic system for prediction of epilepsy
using a standard deep CNN which performs on EEG signals using an ensemble
of P-1D-CNN model. The whole system consists of three modules (a) an Input
module which takes EEG signals and breaks the EEG signal into parts and then
passes these fragments to P-1D CNN model (b) Signal segments are classified
by base P-1D CNN model using an ensemble approach (c) Decision making as
per majority vote by fusing local decisions. Although deep CNN models require
a huge amount of data for training purposes, the authors have introduced the
scheme of data augmentation where deep ensemble classifiers are built after
training the P-1D CNN model using the copies as base models, examining dif-
ferent parts of the input signal. An ensemble takes the input and performs the
process of classification. Local decisions are fused together and a final decision
is taken using the majority vote. The model was tested on University of Bonn
Datasets, which is a benchmark dataset for epilepsy-related cases. The model
has outperformed various previous machine learning models and has produced
an accuracy of 99.1%. Further, the model requires less memory and space and
produces results in less than 0.048 s.
Yuan et al. [39] introduced a novel channel-aware attention framework (Chan-
nelAtt) which is an end to end multi-view deep learning model which uses a
multi-channel approach in high-level space. The authors have proposed tech-
niques for learning attentional representations based on EEG signals in the
time-frequency domain. The authors have demonstrated that the system model
outperforms ordinary machine learning techniques in many ways. The perfor-
mance metrics of the proposed system showed an accuracy of 99.61%, sensitivity
of 96.51% and specificity of 97.85%. The comparison of various deep learning
techniques discussed above has been summarized in Table 1.
This interference keeps on developing and gets worse with time. In the US alone
5.7 million people are suffering from Alzheimer’s dementia [4]. Alzheimer’s is
a progressive and degenerative mental disorder caused due to the death of the
neuron cells thus causing shrinkage of the hippocampus [19]. This number is
expected to increase in the next 30 years. In the US the number of deaths due to
Alzheimer’s disease have shown a steep increase of 123 in the last decade caus-
ing emotional distress to caregivers and puts negative impacts on their mental
health. Presently, there is no cure to Alzheimer’s disease but early detection may
help prevent the progress of the disease [19]. Machine learning approaches that
can help in the diagnosis of AD have been proposed for early detection of the
disease. As the disease has many manifestations and the mild cognitive impair-
ments associated with it may progress to AD or may render the patient stable,
therefore multiple classification techniques are required to access the severity of
the disease properly. In the presence of these constraints, Deep learning algo-
rithms have shown promising results. Some of the recent contributions in this
regard are summarized as follows:
Liu, et al. [26] have introduced a novel method which contains stacked
autoencoders and a softmax output layer based on deep learning to overcome
the bottleneck of conventional machine learning methods and aids in diagnosis
Alzheimer’s disease. Unlike binary classification methods based on support vec-
tor machines (SVM), the authors have introduced multiclass classification which
uses small prior knowledge, has enhanced dimensionality reduction capability
and can reverse the synergy between data modalities simultaneously at the time
of data fusion. The authors were able to produce an accuracy of 87.76 %, the sen-
sitivity of 88.57% and specificity of 83.75%. The cost of misclassifying between
different groups usually varies; therefore, the greater sensitivity figures are very
beneficial for diagnosis.
In a very significant contribution, Bi and Wang [6] have introduced a novel
method for early detection of Alzheimer’s disease. The model is based upon a dis-
criminative deep probabilistic approach which has the capability of multi-task
learning and can classify EEG spectrum image into three classes. It employs
a deep convolutional generative model for multi-task learning. The interesting
feature of this model is that it has been able to bridge the connection between
feature extraction and classification. The authors have employed the system as a
demonstration of Discriminative contractive Slab and Spike Convolutional Deep
Boltzmann Machine (DCssCDBM). The authors are of the view that DCss-
CDBM may be used as a standalone nonlinear classifier with other classifiers.
The proposed method has been able to produce the accuracy of 95.04% and
sensitivity of 90.15% which is better than some known discriminative models
like CDBN. However, the discriminant version of the DCssCDBM is yet to be
introduced to the diagnosis of other neurological disorders like Epilepsy which
might show some positive results.
Spasov, et al. [33] have developed a deep learning architecture which has been
able to classify Mild Cognitive Impairment (MCI) patients and patients who have
a high likelihood of developing Alzheimer’s disease as it has been found that
Contribution of Application of Deep Learning Approaches 93
clinical precursors of AD remain stable over time and do not turn to AD. It is
therefore imperative to classify patients based on the severity of the clinical con-
ditions. The authors have introduced deep learning methods which in addition to
structural magnetic resonance imaging (sMRI) have combined neuropsychologi-
cal and demographic data as input to the system. The deep learning algorithm
proposed by the authors has been able to differentiate between the MCI patients
developing AD and the patients who will remain stable and will not develop AD
over three years. The proposed system was able to produce an accuracy of 86%,
the sensitivity of 87.5% and specificity of 85%. Lee, et al. [24] have applied a
deep learning approach using Recurrent Neural Network (RNN), an architec-
ture specially designed for sequential data. The authors believe that RNN can
help in finding hidden temporal patterns from variable-length sequences. The
authors have applied multiple GRUs for longitudinal multidomain data. The
proposed model was able to predict correctly the conversion of MCI to AD in
only 75% of cases. The model achieved its best when it incorporated longitudinal
multi-domain data with an accuracy of 86%. In another significant contribution,
Hosseini-Asl et al. [18] have adopted a 3D deep convolutional network (3D-CNN)
which can learn generic features of AD. The authors have validated their model
on the ADNI data set. This model outperforms some of the state-of-the-art con-
ventional machine learning models despite using single image modality (MRI)
without any previous skull stripping. The model has been able to produce an
accuracy of about 90.82%. The comparison of various deep learning techniques
discussed above has been summarized in Table 2.
regard. Some of the recent contributions in this area of research are reviewed as
as follows.
Grover, et al. [15] have devised a method using deep learning to diagnose the
severity of Parkinson’s disease based on the voice of the patients. The authors
have deployed the “TensorFlow” deep learning library of python for the diag-
nostic purpose. The experimentation has been done based on Motor UPDRS
score and on the input data set of 16 voice features, the classification accuracy
obtained was 81.6%. The authors have found that motor UPDRS score is a
better classification basis than total UPDRS score and hence it is regarded as
a better metric for predictive analysis. However, the experimentation has been
done on a small data set that too on voice features only. The results would have
been better if gait and handwriting features were also taken into consideration.
In a significant contribution Choi, et al. [9] have developed a deep learn-
ing based FP-CIT SPECT interpretation system which works on image data
and can classify normal and PD images and can accurately interpret variability
in the images which cannot be done by human evaluation. The authors have
developed a deep Convolutional Neural Network and validated it for discrimi-
nating Parkinsonian and non-Parkinsonian tremor on independent SPECT data.
The data used for validation of the model has been independently taken from
the PPMI database (www.ppmi-info.org/data) and SNUH cohort. The accuracy,
sensitivity, and specificity obtained is 96%, 94.2%, 100% and 98%, 98.6%, 100%
respectively for the PPMI and SNUH dataset The authors suggested that the
proposed model can be applied to the reclassification of SWEDD group and can
further be extended to imaging analysis of other diseases and can further help
in making imaging biomarkers.
In another study conducted by Hammerla, et al. [16], the authors have devel-
oped an assessment model based on deep learning approach which can clas-
sify disease state in naturalistic settings, taking into consideration the practical
usability constraints. The model has been validated on a large data set that con-
tains 5,500 hours of movement data that has been collected from 34 participants
in a natural setting. On a HOME data set after training the system, it was able
to obtain an f1 score of 58.1% and on the LAB data set it was able to produce an
accuracy of 59.4%, specificity of 88% and sensitivity of 57%. The results cannot
be compared to prior results as no other systems have followed similar training
and methodology.
Hosseini-Asl, et al. [18] have devised a model to predict the PD with the help
of deep 3D convolutional neural network (3D-CNN) capable of learning generic
features and adaptable to different domain datasets. The proposed model is built
on 3D convolutional autoencoder capable of capturing variations in sMRI scans.
The model has been experimented on CAD Dementia MRI dataset without skull
stripping and has been able to outperform various classifiers viz accuracy. For
enhancing the generality of the features, the authors have used pre-training and
layer freezing strategies. Classification performance has been done after ten-fold
cross-validation on 210 subjects of the ADNI data set. The proposed model was
able to produce an overall accuracy of 92.56%. The comparison of various deep
learning techniques discussed above has been summarized in Table 3.
Contribution of Application of Deep Learning Approaches 95
4 Conclusion
Deep learning as a powerful learning method has much to offer to the data
scientists in general and neurologists in particular. This paper has reviewed
some major contributions of deep learning methods that have been applied to
medical data in recent times. The current trends in deep learning are finding
its relevance particularly in unsupervised learning and training strategies in this
area of learning has regained traction [25]. Deep learning as an experienced
modeling approach is a valuable addition to healthcare decision making. This
procedure can do a large body of work and can reveal unexpressed knowledge
during the diagnosis process. In future, deep learning is expected to play an
important role in improving the quality of diagnosis and understanding numerous
clinical events. It can do accurate and reliable diagnosis, thus making physicians
and healthcare providers smarter in the future.
References
1. Acharya, U.R., Oh, S.L., Hagiwara, Y., Tan, J.H., Adeli, H.: Deep convolutional
neural network for the automated detection and diagnosis of seizure using EEG
signals. Comput. Biol. Med. 100, 270–278 (2018)
2. Acharya, U.R., Sree, S.V., Suri, J.S.: Automatic detection of epileptic EEG signals
using higher order cumulant features. International J. Neural Syst. 21(05), 403–414
(2011)
3. Jabardi, M.H., Al-Fatlawi, A.H., Ling, S.H.: Efficient diagnosis system for parkin-
son’s disease using deep belief network. IEEE Congress on Evolutionary Compu-
tation (CEC), Vancouver, Canada (2016)
4. Association, Alzheimer’s, et al. 2018 alzheimer’s disease facts and figures.
Alzheimer’s Dement. 14(3), 367–429 (2018)
5. Avcu, M.T., Zhang, Z., Chan, D.W.S.: Seizure detection using least eeg channels
by deep convolutional neural network. In: ICASSP 2019–2019 IEEE International
Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1120–1124.
IEEE (2019)
6. Bi, X., Wang, H.: Early alzheimer’s disease diagnosis based on eeg spectral images
using deep learning. Neural Networks 114, 119–135 (2019)
7. Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, New York
(2006)
8. Butler, K.T., Davies, D.W., Cartwright, H., Isayev, O., Walsh, A.: Machine learning
for molecular and materials science. Nature 559(7715), 547 (2018)
96 W. Ahmad Mir et al.
9. Choi, H., Ha, S., Im, H.J., Paek, S.H., Lee, D.S.: Refining diagnosis of parkinson’s
disease with deep learning-based interpretation of dopamine transporter imaging.
NeuroImage Clin. 16, 586–594 (2017)
10. de Lau, L., Breteler, M.M.B.: Epidemiology of parkinson’s disease. Lancet Neurol.
5(6), 525–535 (2006)
11. Deng, L., Yu, D., et al.: Deep learning: methods and applications. Found. Trends R
Sign. Process. 7(3–4), 197–387 (2014)
12. Devasahayam, S.R.: Signals and Systems in Biomedical Engineering: Signal Pro-
cessing and Physiological Systems Modeling. Springer, Boston (2012). https://doi.
org/10.1007/978-1-4614-5332-1
13. Faust, O., Bairy, M.G.: Nonlinear analysis of physiological signals: a review. J.
Mech. Med. Biol. 12(04), 1240015 (2012)
14. Foster, K.R., Koprowski, R., Skufca, J.D.: Machine learning, medical diagnosis,
and biomedical engineering research-commentary. Biomed. Eng. Online 13(1), 94
(2014)
15. Grover, S., Bhartia, S., Yadav, A., Seeja, K.R., et al.: Predicting severity of parkin-
son’s disease using deep learning. Procedia Comput. Sci. 132, 1788–1794 (2018)
16. Hammerla, N.Y., Fisher, J., Andras, P., Rochester, L., Walker, R., Plötz, T.:
PD disease state assessment in naturalistic environments using deep learning. In:
Twenty-Ninth AAAI Conference on Artificial Intelligence (2015)
17. Hornik, K.: Approximation capabilities of multilayer feedforward networks. Neural
Networks 4(2), 251–257 (1991)
18. Hosseini-Asl, E., Keynton, R., El-Baz, A.: Alzheimer’s disease diagnostics by adap-
tation of 3D convolutional network. In: 2016 IEEE International Conference on
Image Processing (ICIP), pp. 126–130. IEEE (2016)
19. Ahmad, M.O., Swamy, M.S., Jabason, E.: Deep structural and clinical feature
learning for semi-supervised multiclass prediction of alzheimer’s disease. In: 2018
IEEE 61st International Midwest Symposium on Circuits and Systems (MWS-
CAS), pp. 791–794. IEEE (2018)
20. Joshi, V., Pachori, R.B., Vijesh, A.: Classification of ictal and seizure-free EEG
signals using fractional linear prediction. Biomed. Sign. Process. Control 9, 1–5
(2014)
21. Kireeva II, N., Baskin, H.A.G., Horvath, D., Marcou, G., Varnek, A.: Genera-
tive topographic mapping (GTM): universal tool for data visualization, structure-
activity modeling and dataset comparison. Molecular Inform. 31(3–4), 301–312
(2012)
22. Kourou, K., Exarchos, T.P., Exarchos, K.P., Karamouzis, M.V., Fotiadis, D.I.:
Machine learning applications in cancer prognosis and prediction. Comput. Struct.
Biotechnol. J. 13, 8–17 (2015)
23. LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436–444
(2015)
24. Lee, G., Nho, K., Kang, B., Sohn, K.-A., Kim, D.: Predicting alzheimer’s disease
progression using multi-modal deep learning approach. Sci. Rep. 9(1), 1952 (2019)
25. Litjens, G., et al.: A survey on deep learning in medical image analysis. Med. Image
Anal. 42, 60–88 (2017)
26. Liu, S., Liu, S., Cai, W., Pujol, S., Kikinis, R., Feng, D.: Early diagnosis of
alzheimer’s disease with deep learning. In: 2014 IEEE 11th international sym-
posium on biomedical imaging (ISBI), pp. 1015–1018. IEEE (2014)
27. McAndrew, P.D., Potash, D.L., Higgins, B., Wayand, J., Held, J.: Expert system
for providing interactive assistance in solving problems such as health care man-
agement, 14 May 1996. US Patent 5,517,405
Contribution of Application of Deep Learning Approaches 97
28. Mitchell, T.M.: The discipline of machine learning, vol. 9. Carnegie Mellon Uni-
versity, School of Computer Science, Machine Learning (2006)
29. Oh, S.L., et al.: A deep learning approach for parkinson’s disease diagnosis from
EEG signals. Neural Comput. Appl. 1–7 (2018)
30. Peek, N., Combi, C., Marin, R., Bellazzi, R.: Thirty years of artificial intelligence
in medicine (aime) conferences: a review of research themes. Artif. Intell. Med.
65(1), 61–73 (2015)
31. Quaglini, S., Barahona, P., Andreassen, S. (eds.): AIME 2001. LNCS (LNAI), vol.
2101. Springer, Heidelberg (2001). https://doi.org/10.1007/3-540-48229-6
32. Sajda, P.: Machine learning for detection and diagnosis of disease. Annu. Rev.
Biomed. Eng. 8, 537–565 (2006)
33. Spasov, S., Passamonti, L., Duggento, A., Liò, P., Toschi, N., Initiative, A.D.N.,
et al.: A parameter-efficient deep learning approach to predict conversion from mild
cognitive impairment to alzheimer’s disease. Neuroimage 189, 276–287 (2019)
34. Tan, J.H., Acharya, U.R., Bhandary, S.V., Chua, K.C., Sivaprasad, S.: Segmenta-
tion of optic disc, fovea and retinal vasculature using a single convolutional neural
network. J. Comput. Sci. 20, 70–79 (2017)
35. Ullah, I., Hussain, M., Aboalsamh, H., et al.: An automated system for epilepsy
detection using EEG brain signals based on deep learning approach. Expert Syst.
Appl. 107, 61–71 (2018)
36. Van Den Heuvel, M.P., Pol, H.E.H.: Exploring the brain network: a review
on resting-state fmri functional connectivity. European Neuropsychopharmacol.
20(8), 519–534 (2010)
37. WHO. Neurological disorders public health disorders. Neurological disorders
38. Yang, H.-C., Islam, M.M., Li, Y.-C.J.: Potentiality of deep learning application in
healthcare. Comput. Methods Programs Biomed. 161, 1 (2018)
39. Yuan, Y., et al.: A novel channel-aware attention framework for multi-channel eeg
seizure detection via multi-view deep learning. In: 2018 IEEE EMBS International
Conference on Biomedical & Health Informatics (BHI), pp. 206–209. IEEE (2018)
Integration of BYOD Technology
in Traditional Classroom: A Statistical
Approach
1 Introduction
A survey in 2005 [1] showed that there was an exponential rise in the use of
tablets by the students compared to the previous years. Today, there are more
mobile devices as compared to the number of humans in the world. We can
utilize this increment in mobile devices for the good of education and learning.
BYOD concept gives a cost effective solution for the problems rising in traditional
classroom. In a classroom, teacher plays an important role encouraging and
motivating student along with course work. Introduction of mobile devices inside
the classroom can give teachers freedom to interact and attain things which
were difficult before. It not only enhances classroom interaction, but also gives
a platform to students as well as teacher to interact and collaborate whenever
they want. With the advancement of mobile technologies, MOOCs, m-learning,
Supported by National Institute of Technology Agartala.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 98–110, 2020.
https://doi.org/10.1007/978-981-15-3666-3_9
BYOD Technology in Traditional Classroom 99
e-learning has grown rapidly giving a base for improving quality of education.
In this study we have utilized statistical way to prove the usability of mobile
devices in the classroom. These mobile devices hold an immense possibility in
changing the way we look at the world. Though, it is still in its initial phase but
with the rise of mobile devices it will surely find its way into the classroom.
Today in the era of being smart, education and technology go hand in hand.
Various researchers have explored the domains on the use of mobile technology
in the traditional classroom. Few works that matched with our idea of incorpo-
rating mobile technology in traditional classroom are as follows:
1. In a study [4] by Wang, Xiao, Kong, and Huang in 2016, the influence of
digital devices on classroom was being analysed. The authors observed 54
lectures in Hong Kong, Beijing, and Shenzhen. It was found that the class-
room when equipped with digital mobile devices gave better understanding of
learning and interaction between teacher and student. The study emphasised
on the use of mobile devices inside the classroom for better interaction and
understanding.
2. A research on the use of mobile devices for teacher education [5] by Evrim
explores the domain of mobile technology for student as well as teacher. The
author had six findings in her study including variation that existed between
practical and theoretical use of mobile device, existing trends of mobile tech-
nology, etc. In the end of the study, it was suggested to use these technologies
for the betterment of education system.
3. A case study [6] that suggested BYOD as the cost effective solution for learn-
ing in Australian Regional University explores over the use of mobile tech-
nology inside the classroom. They proposed some practical, low cost solution
for classroom learning and interaction using BYOD concept.
4. A study [7] in England in year 2012 compared the progress and quality of
learning of two different school. One school was allowed to use mobile tech-
nology in their classroom, while the other school was not allowed to use
mobile devices. In conclusion author recommended to use mobile technology
inside the classroom. The school which was equipped with mobile technology
showed better performance result along with increased attention and interac-
tion among students and teachers.
5. A study on the role of iPad in primary school [8], it was found that with
the incorporation of mobile technology i.e. iPad inside the classroom, learn-
ing can be made interactive. Author integrated iPad in primary classroom
for around eight months and found these devices as the facilitator inside the
classroom. They also analysed the behaviour of students while using mobile
technology in classroom. Later they suggested to use mobile devices in class-
room for enhancement of education in terms of classroom interaction and
better understanding.
BYOD Technology in Traditional Classroom 101
3 Statistical Analysis
Smart phones when joined with education system can give exponential growth
to education system.
Figures 7 and 8 represents a Q-Q plot which stands for quantile-quantile plot.
It is formed by plotting two different set of quantities against each other [10]. In
both of the figure we see the group of points forming approximately straight line.
This method calculates and compares the two different probability distribution.
The Q-Q plot for treatment group has almost all its points on the straight line
forming a better representation, where as in control group points are not that
much concise as in the case of treatment group i.e. points are a bit far from the
straight line. Hence, from the Q-Q plot we can easily infer that the treatment
group has much more effective result than the control group.
While analyzing for the sample data collected from the classroom we also
used linear regression to find if mobile phones are feasible inside the classroom
or not. Linear regression is used to model the relationship between some depen-
dent variable and independent variable [11]. Here, we consider that the data of
treatment group is dependent on the use of mobile phones inside the classroom.
We follow certain assumptions like, the response parameter is continuous, all
the predictors are linearly related to response variable, and the error variance
is constant across each value. Figure 9 represents the summary of the regression
model. R is the multiple correlation coefficients; R square is the proportion of the
regression model and RMSE is the root mean square error. Also, ANOVA stands
for analysis of variance [12], it is used to find the difference between the group
mean of a particular data value used in Fig. 10, representing the sum of square
value for regression model and then residual model, and next it calculates the
total sum of square for both the models. Degree of freedom, represented by df is
calculated for both the models separately and then the total df of both regres-
sion and residual model. F value is used as the criteria for adding or removing
104 P. Kumari et al.
p value is 0.535 which says that the alternate hypothesis yields better result. As
per alternate hypothesis, mobile technologies should be implemented inside the
classroom. Mobile technologies along with laptop, tablet and smart phone can
take teaching-learning process to a new level.
To support our analysis strongly we also used Bayesian paired sample t-test
to interpret the probability of adoption of mobile devices inside the traditional
classroom. It considers the two hypotheses H0 and H1 to calculate and analyze
the procurement of data. T-test is a type of statistical model that takes two
groups and find distribution and difference between the mean. It takes into
account the data of both the groups and finds if the null hypothesis is accepted
or rejected. In our work we have used Bayes factor as the function of statistics
to compare mean and other relevant data. BF10 collects the truthfulness of
alternate hypothesis over null hypothesis, whereas BF01 measures the evidence
for null hypothesis over alternate. In Fig. 11, BF10 can be seen as 31.140, this
acts as the evidence for treatment group i.e. alternate hypothesis, which means
that Bayes factor is 31 times in favor of alternate hypothesis. There is an error of
3.313e-4, which is approximately equal to 5. We can neglect this part till it is less
than 10 Fig. 12 shows various descriptive including N as the sample size, which is
the total number of student considered for this analysis. It also represents mean,
standard deviation (SD), standard error of mean (SE). Credible interval is the
gap where the factors that we have considered have given certain probability.
Figure 13 represents the graphical form of the mean between the two control and
treatment group.
Fig. 14. Prior and Posterior distribution of null and alternate hypothesis
which the points accounts for null hypothesis and above it the points give the
evidence of treatment group i.e. alternate hypothesis. Since we have more data
points accounting for treatment group, it proves with a strong point that mobile
devices are good for classroom. Incorporating these technologies inside the class-
room with enable enhanced interaction, better learning, quality education, and
collaborative teaching learning approach.
4 Limitation
Smart mobile technology involving BYOD in the classroom do come up with cer-
tain limitations and challenges. These mobile devices unknowingly invite various
distractions and threat inside the classroom. Some limitations of mobile devices
inside the classroom are as follows:
– Mobile phones can act as a source of distraction [13] for learning process going
on inside the class. Any vibration due to calls or notification can distract
student.
– Mobile devices with network connection are a gateway for many dreadful
things. Students will have access to all the data over the internet.
– There can be health issues due to large use of mobile phones as these devices
use radio frequency for communication, and these radiations can affect our
body turning out to be harmful [14].
– Since traditional classroom is implemented using mobile devices, battery issue
of these devices is also a limitation.
On one hand mobile technology in classroom can provide us with ample of advan-
tages, but on the other there are so many disadvantages attached to it. We can
limit the use of mobile phones just for the educational purpose. Various rules and
norms can be made on the use of this technology inside the classroom. If handled
properly this implementation can mark the beginning of new era in classroom
teaching learning methodology.
5 Conclusion
Our world is moving at a very fast scale; today we have more mobile devices as
compared to total number of humans on the earth. Mobile technology not only
allows us to access learning data from any geographical location, but also has
enhanced teacher-student interaction. These pervasive devices have turned out to
be a paramount in traditional classroom. This research work analyses the imple-
mentation of BYOD technology along with mobile devices inside the traditional
classroom. This work is presented from the statistical point of view, analyz-
ing class data using regression, Bayesian t-test and other statistical method. In
the course of our study we found that our treatment group got more evidence
compared to our control group proving alternate hypothesis H1 to be true. On
analyzing the result of Bayesian paired t-test, we found that maximum points
that we studied support mobile devices in the classroom with positive results.
This study recommends the use of mobile technology inside the classroom
on the basis of statistical result obtained during analysis. With the growth of
mobile technology, enhancement of traditional classroom should also take place.
Incorporating this mobile technology in the classroom will result better interac-
tion among the students along with quality education, with the power of learning
new and more every day.
110 P. Kumari et al.
References
1. Traxler, J.: Defining mobile learning. In: IADIS International Conference Mobile
Learning, pp. 261–266, June 2005
2. Mbabazi, B.P., Ali, G., Geoffrey, A., Lawrence, N.: Mobile devices for learning in
universities: challenges and effects of usage (2018)
3. The Rise of MOOCs: How Massive Open Online Education Started and Where it
is Now (2019). Examined Existence
4. Wang, X.C., Kong, S.C., Huang, R.H.: Influence of digital equipment on interaction
quality in technology-rich classroom. In: 2016 IEEE 16th International Conference
on Advanced Learning Technologies (ICALT), pp. 455–459. IEEE, July 2016
5. Baran, E.: A review of research on mobile learning in teacher education. J. Educ.
Technol. Soc. 17(4), 17–32 (2014)
6. Sher, V., Hatala, M., Gašević, D.: On multi-device use: using technological modal-
ity profiles to explain differences in students’ learning. In: Proceedings of the 9th
International Conference on Learning Analytics & Knowledge, pp. 1–10. ACM,
March 2019
7. Walker, R.: “I don’t think I would be where I am right now”. Pupil perspectives
on using mobile devices for learning. Res. Learn. Technol. 21 (2013)
8. Mann, A.M., Hinrichs, U., Read, J.C., Quigley, A.: Facilitator, functionary, friend
or foe?: Studying the role of iPads within learning activities across a school year.
In: Proceedings of the 2016 CHI Conference on Human Factors in Computing
Systems, pp. 1833–1845. ACM, May 2016
9. Crisp, G., Guàrdia, L., Hillier, M.: Using e-Assessment to enhance student learning
and evidence learning outcomes (2016)
10. Turner, S.D.: qqman: an R package for visualizing GWAS results using QQ and
manhattan plots. Biorxiv, 005165 (2014)
11. Linear regression. En.wikipedia.org (2019). https://en.wikipedia.org/wiki/Linear
regression. Accessed 5 Sept 2019
12. St, L., Wold, S.: Analysis of variance (ANOVA). Chemometr. Intell. Lab. Syst.
6(4), 259–272 (1989)
13. Elgan, M.: Smartphones make people distracted and unproductive. (2019).
Computerworld. https://www.computerworld.com/article/3215276/smartphones-
make-people-distracted-and-unproductive.html. Accessed 18 July 2019
14. “Mobile Phones And Your Health” (2019). www.betterhealth.vic.gov.au. Accessed
22 Aug 2019
Statistical Analysis of Road Accident
Data of UK Using Operational
Intelligence Tool - Splunk
1 Introduction
There are colossal number of impacts happening around the globe and the demise
and fatal injury rates are essentially developing as demonstrated by assessments
and statistics distributed by various organizations over the world. Impacts can
happen due to many reasons like liquor, drugs, weather conditions, careless-
ness, and so forth. Traffic engineers are accused of influencing transportation
frameworks to keep running as effectively as could be expected under the cir-
cumstances, however the assignment appears to be unmanageable. In any case,
with mechanical advancements and complex data analytics, governments are
better ready to fight gridlock now than any time before. In the long run big data
analytics will make driving considerably more secure. Thus, ample opportunity
is there to adequately deal with the road accident analysis. Prescient analytics
gives an all new way to extract valuable urban mobility data from networked
infrastructure, associated vehicles and cell phones, for real-time evaluation of
traffic patterns, and deploying necessary management strategies. Data sources
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 111–119, 2020.
https://doi.org/10.1007/978-981-15-3666-3_10
112 T. Deb et al.
within our cities means analytics offer potential to open up a whole new era of
smart transport.
We have utilized an operational intelligence tool called Splunk [1,2] to dissect
our information in view of the accompanying reasons:
Features of Splunk
There is a need to manage traffic in a smart way using data analytics and
machine learning as the management of traffic with the conventional way such
as the signaling system is not having a major effect in the road accident pre-
vention and also to ensure the smooth flow of traffic at peak times and during
emergency situations, such as following an accident, for instance. The analysis
information and the predictions can provide researchers with insights as what
Statistical Analysis of Road Accident Data of UK 113
factors contribute most to causing accidents and how to tackle accidents in the
urban areas. This will ensure safety to the people out there. Hence, a smart
living experience.
3 Literature Survey
Urban areas have a few difficulties for us to handle; one of them is the issue of
urban portability. The United Nations anticipated that half of the total populace
would live in urban zones toward the finish of 2008. The developing populace
and absence of accessible physical space have made traffic administration pro-
gressively difficult in Singapore. So with the increasing traffic, the probability of
road accident is getting increased. By 2020, travel demand is supposed to ascend
from 8.9 million trips for each day to around 14.3 million, implying the stamped
increment in the city-state’s populace in Singapore. Simultaneously, Singapore
confronts real imperatives in space, with 12% of land officially possessed by the
3,300 km-street arrange and another 15% dedicated to lodging [4,5]. These days,
forecasts say that by 2050, 86% of the developed world and 64% of the devel-
oping world will be urbanized. The Fig. 1 demonstrates the accident scenario in
the world for the year 2010.
We make use of different types to predict the causes of highway traffic acci-
dents by analyzing the data. Beshah’s and his team [7] analyzed 14,254 accidents
that happened in Ehiopia between May 2005 and September 2008 to figure out
their patterns and predict the role of commuters on possible injury risks. Yan
et al. [8] analysis revealed the causes of highway traffic collisions to be human
error, vehicles, environment of the traffic and traffic management. They set up
114 T. Deb et al.
a dynamic system wherein if one of the factor goes out of order, an accident will
occur. R Geetha Ramani’s et al. researched various data mining techniques to
help in the prediction of road accidents due to the commuters [9]. They ana-
lyzed a dataset of about 159,417 casualties in 157,463 accidents of Great Britain
to find out road accident patterns related to pedestrian characteristics. Sharaf
Alkheder et al. used artificial neural network [7] to predict the injury severity of
traffic accidents based on 5973 traffic accident records occurred in Abu Dhabi
over a 6-year period (from 2008 to 2013). They used a tool called WEKA to
build ANN classifier. These papers endeavored to make sense of traffic accident
patterns based on human behaviour. However, it is hard to discover relationship
between static information about people and impacts since human conduct is
difficult to foresee. What we center around are the conditions that surround with
the traffic accidents. This includes road surface conditions, climate conditions,
number of casualties and other factors.
Data taken from kaggle.com contained 3 data sets, viz, 2005–2007, 2009–
2011 and 2012–2014. We combined all the 3 CSV files together and named it
all accidents.csv. It was observed that the total number of accidents reported
from 2005 till 2014 has decreased over the time and this can be visualized in
Fig. 2.
Further more, in Fig. 3, all of the accidents are pictured that happened in
between 2005 to 2014 in a cluster map which basically shows aggregated values
in a geographic region.
Data Visualization and Statistics. Data visualization is projecting the rele-
vant information in pictorial or graphical form. For instance, iPython queries can
be visualized using matplotlib plotting library and seaborn visualization library.
Apart from these, there are other powerful tools available for data visualization
like Cognos, Clickview, Tableau and R for raw data statistics. Splunk on its own
has the capability for data visualization as well as raw data statistics.
Statistical Analysis of Road Accident Data of UK 115
Something else to observe was that the most number of crashes were
accounted in between 2:00 PM to 7:00 PM where top time of impacts is between
4:00PM to 6:00PM. The slightest is amid 8:00 PM to 8:00 AM. This is appeared
in Fig. 5.
116 T. Deb et al.
More than 70% of the impacts happened in the lighting condition - “Day-
light: Street light present” and least in - “Darkness: Street lights present but
unlit”. Figures 7 and 8 shows number of casualties from 1 to 8 in various lighting
conditions.
s o u r c e = ‘ ‘ a l l a c c i d e n t s . csv ’ ’ h o s t = ‘ ‘ l o c a l h o s t . l o c a l d o m a i n ’ ’
i n d e x = ‘ ‘ web ’ ’ s o u r c e t y p e = ‘ ‘ csv ’ ’ | c h a r t count o v e r
L i g h t C o n d i t i o n s by N u m b e r o f C a s u a l t i e s
As per the analysis, Figs. 9 and 10 shows most severe accidents occurred in
dry road conditions followed by damp road surface conditions.
5 Future Work
The statistical analysis on the traffic data can be put on a machine learning
model so that we can get an idea of which factors could have any sort of impact
or influence on the road traffic accident. Machine learning is a cutting edge field
in computer science that seeks to get computers to carry out tasks without being
explicitly programmed to carry out a given task [10]. Machine learning uses many
techniques to create algorithms to learn and make predictions from data sets. It
is used in data mining which is a technique to discover patterns and models in
data sets where relationships are previously unknown.
To create good machine learning systems, we need to focus on 4 things -
Data preparation capabilities, Algorithms basic and advanced. Automation and
iterative processes and Scalability.
Predictive analytics for traffic accidents information will provide capability
of integrating different technologies with the existing communication infrastruc-
tures that can help diminish casualties; increase safety across the capacity and
adequacy of street networks. We can achieve a unique approach of analyzing
accident data using Operational Intelligence and Machine Learning. We can do
horizontal scaling as well to increase our computational power using big data
tools like Apache Hadoop and Spark framework that can help in distributed
processing, sorting the outputs.
Statistical Analysis of Road Accident Data of UK 119
6 Conclusion
In this paper, we have adopted the approach to leverage the Splunk platform
for road accident data analysis. We have analyzed the data in a single platform
compared to using multiple platforms for data storage for statistics and analysis.
In our thorough analysis, we projected and visualized information about impacts
that happened in UK in between 2005–2014. Through our investigation and by
looking at the visualizations we can infer that:
– Road accidents have decreased over the years gradually from 2005 to 2014.
– Most of the accidents happened in between Thursday and Saturday than the
remaining days.
– Peak time for accidents is between 4:00PM to 6:00PM
– Around 80% of the collisions reported were of level 3 i.e. severe.
– Dry road surface conditions accounted for the most number of accidents fol-
lowed by wet road surface.
– More than 70% of the accidents happened in “Daylight:street light present”
lighting condition.
– Also the accidents reported were most severe in dry road surface conditions.
Further machine learning techniques can be applied on this data analysis for
predicting accident hotspots, accident severity and much more.
References
1. Splunk. https://www.splunk.com/. Accessed 4 Sept 2019
2. Splunk features. https://www.quora.com/in/What-are-the-benefits-of-using-Spl
unk. Accessed 1 Aug 2019
3. https://www.kaggle.com/daveianhickey/2000-16-traffic-flow-england-scotland-
wales
4. Intelligent Transportation Systems - Next Wave of the Connected Car. https:
//www.intellimec.com/insights/intelligent-transportation-systemsnext-wave-conn
ected-car/
5. Urbanization of World. http://www.civilservicesexpert.com/geography/urbanisat
ion-of-world/
6. Road Traffic Accident status. http://www.armstrongforensic.com/blog/index.
php/2010/04/12/south-asia-leads-world-in-road-traffic-accidents/
7. Beshah, T., Ejigu, D., Abraham, A., Snasel, V., Kromer, P.: Pattern recognition
and knowledge discovery from road traffic accident data in Ethiopia: implications
for improving road safety. In: 2011 World Congress on Information and Commu-
nication Technologies (WICT), pp. 1241–1246 (2011)
8. Yan, Z. Wang, X., Du, L.: Design method of highway traffic safety analysis model.
In: 2011 IEEE International Conference on Transportation, Mechanical and Elec-
trical Engineering (TMEE) (2012). https://doi.org/10.1109/TMEE.2011.6199168
9. Ramani, R.G., Shanthi, S.: Classifier prediction evaluation in modeling road traffic
accident data. In: 2012 IEEE International Conference on Computational Intelli-
gence and Computing Research (ICCIC), pp. 1–4 (2012)
10. Machine Learning - Brilliant Math Science Wiki. https://brilliant.org/wiki/
machine-learning/
Security
Detection of Routing Infrastructure
Attack in TCP Connection
Abstract. From the last few years, routing infrastructure attacks like
Distributed Denial of Service Attack (DDoS) and Denial of Service (DoS)
Attack have been the most trending topic in the domain of Network secu-
rity. And these days most of the people prefer online payment, online
shopping, online class etc. So the attackers tries to flood the router dur-
ing this time. And We are focusing on Distributed Denial of Service
Attack. In this paper, it implement a method to detect (DDoS) Dis-
tributed Denial of Service Attack in router by using Ns3 simulator. For
this simulation we used unbalanced dumbbell topology. And the process
of imposing attack scenario in dumbbell topology is defined in section
VI. This simulated DDoS attacks are presented in graph by using the
proposed algorithm to detect the routing infrastructure attack in the
network.
1 Introduction
Routing is one of the most important parts of infrastructure to keep the network
running. Its main function is to forward the packets from one network to another
network. And this is usually done by the device called router [1]. Router does not
just route the traffic to another networks, they provide a database that stored
all the information of the network and choose the shortest route to forward the
packet to their destination. It consists of three components. They are IP address,
metrics and topology. IP address comprises of the IP address of the sender’s and
the receiver’s. Metrics provides the details of cost values used by the routers to
route a packet. By using metric values, the router determines the best path to
forward the packets in a network. And topology is the arrangement of the sender
nodes and the receiver nodes through the lines of connection. In [2], routing table
stores all the route information of the nodes (senders, receivers). From that route
information the sender will choose the best route i.e. shortest route to send the
packets to another host. During this transmission of packets from one host to
Supported by organization x.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 123–131, 2020.
https://doi.org/10.1007/978-981-15-3666-3_11
124 A. D. Chanu and B. Sharma
another host, an attacker tries to disrupt one or more primary function of the
router. Routing infrastructure attack is a kind of attack that occurred in router
to disrupt in the peering session [11]. During this process, several attacks try
to forge the network. As mentioned in, for exchanging the route information
among the neighbors, a router needs to create a peering. During this peering
session the attacker attempts to break the established session by resetting the
TCP connection. Some of the attacks in router are Denial of service attack,
Packet mistreating attack, Routing poisoning attack, Hit and run attack and
Persistent attack. In denial of service (DoS) attack, it attempts to prevent the
users from attempting network information or services by sending more numbers
of ICMP packets from multiple sources. Packet mistreating attack aims to inject
packet with malicious code to confuse and disrupt the network and router. In
Hit and Run attack, the attackers inject the malicious code into the router and
check whether the network is running properly or not. It is very easy to spot this
kind of attack. Whereas in persistent attack, the attackers inject the malicious
packets to the router until they achieved their goal.
2 DDoS Attack
Distributed Denial of service (DDoS) attack is also a kind of attack that happen
in router, where a group of connected devices try to flood the bandwidth or
resource of target systems such as server, website or the network [9,12]. This
flooding of connection messages or connection request to the target system may
shut down or crash, thereby denying services to legitimate users or system.
Figure 1 shows the structure of DDoS attack. It consists of four components
such as Attacker, Master, Handler and the Victim. The attackers will handle the
port number, TTL value of the victim. Whereas the handler will check which
agents are up and running, it can be TCP, UDP or ICMP. And the agent will
generate powerful packet streams to the victim to flood the bandwidth of the
network or the resources of the victim system [10]. To detect the DDoS attack,
first it is mandatory to know about router’s functionality. The DDoS (Router
Infrastructure attack) attack can be described as follows:-
– At first the network is stable and everything works fine.
– Hackers start to control some host for generating traffic from the controlled
host to the router.
– When the attack on the router is initiated, then some network connections
to that router started dropping the data packets.
– During that attack, some connections are affected and finally all the connec-
tions to the routers are shut down completely.
network and render it unresponsive. This can be done by sending faster connec-
tion request than the targeted machine. In SYN flood attack, the attacker starts
sending repeated SYN packets to every port on the targeted server or system to
flood their bandwidth. The server unknowingly started receiving multiple legit-
imate requests to establish the three-way handshake connection. And this will
wait the legitimate users to establish the connection. During this time no new
connection cannot be establish (Fig. 2).
3 Related Works
In [1], authors proposed a scheme to detect and mitigate against the TCP SYN
flooding attacks using three counter measures algorithm to detect the spoofed
126 A. D. Chanu and B. Sharma
IP packets. The three counter measures are C-1, C-2, C-3. C-1 is used to record
the initial SYN packets of each connection. C-2 is used to record the completed
SYN packets. C-3 is used to record the other flooded SYN packets. From this
mechanism they are capable of detecting various SYN flooding attack with high
accuracy in short response time. This mitigation scheme starts working once it
detects the SYN flood. And the received SYN packet is monitored by using three
C’s. The results are
– If the SYN packet is not in any of the three C’s then drop the SYN packet
– If it is in C-2 then pass the packet to C-3.
– If the packet is in C-3. TCP connection is completed; pass this packet from
C-3 to C-2.
As mentioned in [2], authors developed a scalable method for detecting the
routing attacks and locating the malicious nodes called slip. In slip it includes the
synchronization to provide synchronization for the updates of the routing table.
And they undergo a simulation to detect the attack under different network.
As mentioned in [3], the authors proposed an approach called hop count
based packet processing to counter DDOS attack. This DDOS attacks are very
difficult to detect in the source node since the attackers used spoofed IP address.
So, in this approach they marked all the packets which came from the same hop
count and through the same router with the same identification number. At the
receiving side, it checked the hop count value for the incoming packet with that
stored value. Thus, attack packets are identified early and reduced the spoofing
threats also. As mentioned in [4] nowadays, most of the commercial activities like
online shopping, online banking depends on the internet. And denial - of - service
attacks become a main threat for this internet infrastructure. They assumed all
clients are outside the LAN in which the server is present so that the proposed
system can monitor all the incoming packets connecting to the server. In the
detection architecture they divide it into three part i.e. flow classifier, object
module, filter. In flow classifier it consists of IP address of the source, IP address
o f the destination and the IP port number and it maintained the arrival time
of the packets at the edge router. Whereas in object module it will monitor the
flow unti it is considered normal. It changed its parameter to detect different
kinds of attacks. And lastly the filter is used to block malicious flow by the
object module. If the burst length is greater than or equal to RTT (round trip
time) of other connections with the same server and time period is equal to fixed
minimum RTO (Recovery time objective) then they conclude that there is an
attack in edge router.
As mentioned in [5], Distributed denial of service has become a serious prob-
lem to a network. They proposed a defense mechanism to detect the malicious
traffic mainly the SYN flooding attack. SYN flooding attack is a kind of attack in
which the SYN packet is much higher than the RST or FIN packets. In defense
mechanism they use four cases. Case 1: To establish a connection between a
client and a server SYN, SYN ACK, ACK packets are required. During this con-
nection if a server received SYN FIN packet then they conclude that the packet
is malicious. Case 2: The attacker source port or destination port should not be
Detection of Routing Infrastructure Attack in TCP Connection 127
considered as 0. If the ACK flag is set then the acknowledgement number should
not be considered as 0. Case 3: If a hacker uses some IP address to send the SYN
packets to the server. And the server will reply SYN+ACK packet with ICMP
message to the client to get the information like whether the client receive the
SYN+ACK or not. Case 4: By tracing the route from where it is sent and where
to receive.
As mentioned in [6] DDOS attack is an attempt to prevent the legitimate
users from using services provided by the service providers. So, to detect such
kind of attacks the author proposed an approach based on entropy. Entropy is
the measurement of the randomness and uniformity of the IP address. Some
researchers examined that higher volume of traffic, incomplete connections,
flooding of packets are characteristics of the attack. It can be calculated by ana-
lyzing a series of packets, refer as the window. For normal network connection
the value of the entropy may vary up to some extent but in case of DDOS attack
it will have vast difference in result. They calculate normalized entropy to find
out the overall probability distribution in packet window W. If the normalized
entropy < threshold value then, mark window as suspected.
As mentioned in [7], the authors proposed an algorithm to detect DDOS
attack in the network using Entropy Based Anomaly Detection Algorithm.
Entropy value is calculated with respect to time and packet windows. For identi-
fying the attacks, they mixed the attack traffic with legitimate traffic. And again,
the entropy of the mixed traffic is calculated. By using PERL programming lan-
guage, they used the retrieved data as input to the program for calculating
entropy. Two different approaches are used for detecting DDOS attack. In first,
they calculate the entropy with respect to time window. In time window app-
roach, entropy of the traffic is calculated with respect to equal time stamps. In
the second approach, entropy value is calculated with respect to packet window.
For packet window, equal numbers of flowing packets are taken from network
traffic.
As mentioned in [8], A Distributed Denial of Service attack generates mali-
cious packets from different computers to shut down the communication between
the legitimate users and the servers. In order to mitigate the DDoS attack, it
includes several steps like detection, characterization and trace back. So, the
author used entropy approach to detect the flooding attacks. First, they calcu-
late the entropy for all the clients connecting to a router. Secondly entropy-based
trace back method is applied to find the edge routers from where the whole attack
traffic is entering into the ISP domain of the data center. From this they plot a
graph between the entropy based on trace back with the time. If the values of
entropy get fluctuate then they concluded that there is a DDoS attack in that
period of time.
5 Imposing Attack
Create a new socket with TCP interface by sending packets only with SYN flag
by using the following process,
– Using Factory
• CreateSocket (): Ptr<Socket>
– Using Helper
• Install (node: Ptr<Node>)
• Install the new socket factory class instances to the nodes.
– Using SYN flood socket
• GetType ID
• Bind()
• Connect()
• Listen()
• Set allow broadcast with the SYN flag
For the same topology with attack scenario, gradually the time of simula-
tion is increasing. It has been observed that number RST packets are gradually
increasing with high numbers in comparison to the same topology with same sim-
ulation time without attack. Efficiency of detection methodology is also observed
in both cases (Table 1 and Fig. 7).
130 A. D. Chanu and B. Sharma
Fig. 7. Calculation of RST packets and SYN packeGraph of attacks and non-attacks
for RST packets.
8 Conclusion
Routing Infrastructure attack like DDoS attacks are very complex problem, and
numerous approaches have been proposed to detect this DDoS attack. In this
paper, a simple method is proposed to detect the routing infrastructure attack
i.e. DDoS attack in router by using Ns3 simulator. We used dumbbell topology
for the attack scenario and non-attack scenario and it has been observed that
number of RST packet for attack scenario is more than the RST packet of non-
attack scenario with respect to simulation time.
References
1. Gavaskar, S., Surendiran, R., Ramaraj, E.: Three counter defense mechanism for
TCP SYN flooding attacks. Int. J. Comput. Appl. 6(6), 0975–8887 (2010)
2. Prathap, A., Sailaja, R.: Detection and Prevention of Denial of Service Attacks
Using Distributed Denial-of-Service Detection Mechanism
Detection of Routing Infrastructure Attack in TCP Connection 131
3. KrishnaKumar, B., Krishna Kumar, P., Sukanesh, R.: Hop count-based packet
processing approach to counter DDoS attacks. In: 2010 International Conference
on Recent Trends in Information, Telecommunication and Computing (ITC). IEEE
(2010)
4. Shevtekar, A., Anantharam, K., Ansari, N.: Low rate TCP denial-of-service attack
detection at edge routers. IEEE Commun. Lett. 9(4), 363–365 (2005)
5. Kumarasamy, S., Gowrishankar, A.: An active defense mechanism for TCP SYN
flooding attacks. arXiv preprint arXiv:1201.2103 (2012)
6. Sharma, S., Sahu, S.K., Jena, S.K.: On selection of attributes for entropy-based
detection of DDoS. In: 2015 International Conference on Advances in Computing,
Communications and Informatics (ICACCI). IEEE (2015)
7. Singh, J., Sachdeva, M., Kumar, K.: Detection of DDOS attacks using source IP
based entropy. Int. J. Comput. Sci. Eng. Inf. Technol. Res. (IJCSEITR) 3, 2249–
6831 (2013)
8. Bhandari, A., Sangal, A.L., Kumar, K.: Destination address entropy based detec-
tion and traceback approach against distributed denial of service attacks. Int. J.
Comput. Netw. Inf. Secur. 7(8), 9 (2015)
9. Zhang, J., et al.: An advanced entropy-based DDOS detection scheme. In: 2010
International Conference on Information Networking and Automation (ICINA),
vol. 2. IEEE (2010)
10. https://www.cisco.com/c/en/us/td/docs/solutions/Enterprise/Security/
Baseline Security/securebasebook/sec chap3.html
11. Navaz, A.S., Sangeetha, V., Prabhadevi, C.: Entropy based anomaly detection
system to prevent DDoS attacks in cloud. arXiv preprint arXiv:1308.6745 (2013)
12. Jun, J.-H., Oh, H., Kim, S.-H.: DDoS flooding attack detection through a step-by-
step investigation. In: Networked Embedded Systems for Enterprise Applications
(2011)
Optimizing e-KYC Process Using
Distributed Ledger Technology
and Smart Contracts
1 Introduction
Know Your Customer (KYC) is the process of a business or organization verify-
ing the identity of its clients or customers. It is required to assess the suitability
of customers. Especially, banks and other financial organizations are employing
the KYC process to make sure that their customers provide the due diligence
information required to fulfill the anti-money laundering regulations. Therefore,
KYC has become a mandatory and crucial procedure for most of the businesses
[1]. However, there is a lot of efforts and cost involved in the KYC verifica-
tion. In the current system, when a customer opens first account with a bank,
the process cost recurs (please refer Fig. 1 [2] as this process is labor and time
intensive. Some banks don’t have enough staff to attend the number of Anti-
Money Laundering (AML) alerts, and it has more than 85% false positive rate.
Hence, an adaptive and time efficient system is needed which should offer more
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 132–145, 2020.
https://doi.org/10.1007/978-981-15-3666-3_12
Optimizing e-KYC Process Using Distributed Ledger Technology 133
granular data. An inefficient system costs banks as they are failing to fulfill KYC
AML compliance and thus, fined by governing authorities. In 2018, Dutch bank
ING was failed to fulfill Dutch AML compliance, and as a result, it was fined by
900 million dollars [3]. Thus, there is a need of an optimized system which can
achieve an efficient KYC and AML, while creating the least amount of friction
for banks and its customers.
In literature, a very few attempts have been made towards simplifying the KYC
process and increasing its efficiency without compromising security. We have
observed that the e-KYC process needs immutability of the records. The cost
of e-KYC should be further reduced and if multiple banks are involved for a
customer, then the overall cost should be shared among them. In addition, it
should have resistance to compromise capability. We have investigated a role of
the national regulator such as RBI in India for a DLT-based e-KYC solution.
However, the identity of the individuals is revealed during KYC approval pro-
cess and storing the documents. Although blockchains offer data transparency
and tracking, the privacy and auditability of a customer is also important for
many applications. The proposed solution attempts to achieve anonymous KYC
details sharing among organizations with the consent of users and cost distribu-
tion of KYC process among the involved organizations. To offer anonymity of
transactions, zero knowledge proofs are used in most of the proposals. Zerocash
[6] offers the decentralized anonymous payments. It guarantees strong privacy
for transactions. However, it lacks in auditability. As many applications needs
auditability, later Zerocash [6] was enhanced to offer privacy and accountability.
In addition, it focuses more on payments and thus, it is very difficult to use in
applications which require data or asset transfer.
IBM has developed “Proof-of-Concept Blockchain-based Shared KYC” in
January 2018 in collaboration Deutsche Bank and HSBC [7]. Shyft [8], a
blockchain-based network allows users to transfer data with regulatory com-
pliance. It has a creditability feature which offers users with a reputation score
based on compliance and historical transaction activity. Here, customers, data
attesters and data consumers interact with each other via the shyft network, and
thus the anonymity of customer relationship is not maintained. Other solutions
like kyc-chain.com [9], kyc.legal [10], and Fides (norbloc.com) [11] are proposed
use blockchain for KYC verification. However, a successful implementation is not
reported yet for document validation by the trusted participants in the network.
In addition, it is very much important to offer data validation among the banks
with the privacy of the customer-bank relationship. Some solutions such as like
KYC services [12] attempts to offer data standardization and harmonization in
KYC processes among banks. However, it relies on a trusted third party [1].
Optimizing e-KYC Process Using Distributed Ledger Technology 135
Recording Documents. For the first institute recording the KYC documents,
the customer must provide the documents in person and have to be verified
by the bank. It is a tedious process and involves a cost. After the customer has
been verified for those documents and the consent for these documents and bank
has been recorded, the bank then selects a symmetric encryption key such as
AES, encrypts these documents and store them in an off-chain database. This
off-chain database is an IPFS database. After this, the bank records the hash
of these documents along with the link encrypted using the same key on the
Optimizing e-KYC Process Using Distributed Ledger Technology 137
blockchain through record document contract. This link can be later used to
share these documents with other banks and hash helps to prove their integrity.
The contract for recording the document is as follows:
recordDocument (customerID, documentsList, transactionCertificate, docu-
mentsHash, encryptedLink)
– Validate customer signature
– Check for the presence of consent on the ledger by querying
– Records the hash and encrypted link of the encrypted documents on DLT
As shown in Fig. 3, the data stored on blockchain includes the Customer
Profile Table, Customer Consent Table, Encrypted Key Table, Hashed Document
Table. The customer ID is a primary key in each table. The consent record is also
linked by type of document and bank’s transactional certificate. Each customer’s
profile is addressed by a globally unique ID with type, and a list of such IDs
and types are recorded in a profile table. The document hash with the link to
off-chain is stored by encrypting it using symmetric key (AES). Such AES keys
are stored in a separate table by encrytping using the public keys as given in
the transactional certificates. Here, the encrypted documents of a customer are
138 H. Hanbar et al.
stored in this IPFS database. This encrypted data is sent to the permissioned
database with a corresponding link. This link can later be used by other banks
to get the encrypted data.
Consent. A customer gives the consent to a bank for accessing KYC details
by recording “the consent” on the blockchain. A consent is a structured record
including customer ID and the transactional certificates of the bank to which
the customer has given consent along with the list of documents to access. The
smart contract for recording consent is as follows:
recordConsent (consent, transactionCertificate, customerCertificate, signT,
signC)
1. The customer selects the list of documents through the portal and sends a
request to a bank for which he/she wants to give consent.
2. The bank selects a unique transaction certificate for this bank customer rela-
tion and replies to customer with this transaction certificate signed with
enrollment certificate secret key and consent signed with transaction certifi-
cate’s secret key. The bank must maintain a map of customers to transaction
certificates and key used to encrypt documents for later exchange of key.
Optimizing e-KYC Process Using Distributed Ledger Technology 139
Fig. 4. Customer’s consent for bank created through one round trip.
3. The customer then validates these signatures to make sure he/she is interact-
ing with correct bank and then invokes the record consent smart contract on
blockchain to record consent and corresponding transaction certificate. The
smart contract records the consent on the ledger after validating customer
and bank sign.
When other banks with consent from customer wishes to borrow the verified
credential, the exchange of key between these two banks happens without know-
ing each other’s identity through requestKey and fulfillKey contracts as given
below:
requestKey (customerID, documentsList, transactionCertificate, signT)
– Check for consent and sign
– Select another Banks’ Tcert at random
– Ask all banks to fulfil key if Tcert matches
fulfillKey (customerID, documentsList, transactionCertificate A, transac-
tionCertificate B)
– Check the presence of consent for the requestor and provider
– Select another Banks’ Tcert at random
– Record the AES key encrypted using requestor’s TCerts’ public key
Figure 5 shows the process of exchanging the key between banks. It performs
following steps:
1. The second bank first selects a unique transaction certificate for this relation
with the customer and record a request on the blockchain for the key of
those documents after signing this request with its transaction certificate’s
secret key. The requestKey contract selects a random transaction certificate
holder, among all those having consent for this customer, who must satisfy
this request of key. This is done to ensure that not all banks try to fulfil this
request at a time.
140 H. Hanbar et al.
2. The success of this request key notifies all banks, including the bank with key
to customer credentials selected by the contract. This bank realizes that it is
its transaction certificate which has been selected and has key for the creden-
tials of the customer asked in this request. It then looks for the corresponding
symmetric key and encrypt this key with the second banktransaction certifi-
cate’s public key and invokes fulfilKey contract after signing request with its
transaction certificate’s secret key.
3. The smart contract validates the presence of consent for both transaction
certificates and then records this encrypted key on the blockchain.
4. Using this recorded key, second bank is able to verify KYC for a customer
who has given consent.
Here, the cost sharing mechanism works as follows: An average cost per
customer is decided by unanimous agreement of all the parties involved. A fixed
value token system (T ) is used to meet such a requirement. To get the customer
data from a bank, the requesting bank has to pay his share of cost via this token.
This cost is shared among all the banks which currently hold the customer’s
identity. This ensures no bank single handily exploits the systems for its own
benefit and the overall cost of this process is shared among all the banks which
desire the KYC data of that particular customer. Suppose, bank A’s cost is T for
customer C’s KYC verification. When bank B receives confirmation from bank
A for customer C’s KYC verification, it pays cost T/2 to bank A. Thus, cost to
bank A and B becomes T/2. Now, bank C needs to do KYC verification for same
customer and bank A has given confirmation, then bank C has to pay T/3 cost
to bank A. As a result, bank C’s cost is T/3 and bank A’s cost becomes T/6.
Thus, the overall cost of KYC verification of a customer is shared among the
banks. In addition, the bank fulfilling the key request is provided with an extra
fractional reward later to keep them motivated towards this business network.
Optimizing e-KYC Process Using Distributed Ledger Technology 141
For the performance evaluation of the proposed e-KYC solution, we have imple-
mented a prototype using Hyperledger Fabric [15]. As shown in Fig. 6, we set up
a virtual prototype network using Docker on a computer with following spec-
ifications: Operating System: Ubuntu 18.04 (Linux), Processor: Intel core i5
8250U, Memory: 8 Gb. We have considered two organizations, namely: Banks
and Customer portal with one peer for each organization. Peers are owned by
organizations and considered as nodes in the network which host the ledger and
smart contracts. Each organization also have a certification authority (CA) and
membership service provider (MSP) to issue certificates and to provide role to
identities in network.
We have generated the required certificates and keys for the banks and cus-
tomers, with the genesis block which bootstraps the service. We have created a
number of transactions to configure the channel. All the Docker containers are
launched along with the couchDB containers for each peer. CouchDB serves as
the state database to store key value pairs of the ledger and allows rich query
against the chaincode data. Docker starts a virtual network, e-KYC network
for the services running in container to communicate. The peers are then made
to join the channel. A client container is launched to execute commands in the
peer containers, which is used to install the proposed smart contracts (KYC-
chaincode) in both the peers. Here, chaincodes are written in node.js, through
which blockchain is accessed. We have deployed different chaincodes viz; record-
Consent, recordDocument, requestKey and fulfillKey in node.js (as discussed in
Sect. 3.2). APIs of Banks and customer portals are implemented in node.js. These
RESTful APIs are implemented with express.js to record consent and facilitate
142 H. Hanbar et al.
exchange of keys. They wait for consent record request from a customer and
reply with the transaction certificates. We use MongoDB database for banks to
store the mapping between transaction certificates, customerID, and the key for
the encrypted documents. This key is encrypted using a master key of bank.
Bank’s API also waits for any key request made for their transaction certificates
and invokes the fulfil key smart contract. New users register by Fabric Node
SDK through node.js. An admin is enrolled first and his/her ID is used by the
customer portal to register new customers. This returns a wallet ID which is used
by a customer to interact with the blockchain and perform transaction such as
record consent. Same SDK is used to enroll banks and to provide them with
enrollment certificates. In addition, it is used at bank API side to get new trans-
action certificates. All these certificates are X.509 and SHA-256 with ECDSA
used for digital signature generation.
The chaincode mentioned above in the proposed e-KYC solution are tested
using Hyperledger Caliper, a blockchain benchmark tool [17]. Its functions are
subjected to a large number of test transactions at different sent rates. We
have derived the results in terms of throughput and latency in executing the
transaction at different sent rates.
Fig. 10. Latency of the bank’s chaincode in the proposed e-KYC solution.
144 H. Hanbar et al.
5 Conclusion
The KYC revolution has addressed the problem of money laundering and other
unidentified funding. The current KYC process in banks is redundant, time con-
suming and costly for every customer. To address these problems, we have pro-
posed an optimized e-KYC solution which leverages the benefits of the blockchain
and smart contracts. With the help of proposed smart contracts, a customer need
to do one time KYC verification process only and results of same is shared among
other banks with the consent from a customer, while maintaining the confiden-
tiality of bank-customer relationships, improving the accountability, cost sharing
among the involved banks, sufficient throughput and affordable latency as com-
pared to current solutions. We have implemented the prototype of the proposed
solution using hyperledger fabric blockchain and evaluated the results in terms of
throughput and latency in performing the transactions. The experimental results
and analysis of the proposed solution are very encouraging to make the proposed
solution standardized and widely utilized in financial institutions. In future, the
throughput of the proposed solution can be further improved by incorporating
lightweight crypto algorithms and better performance databases in peer nodes
for the fast execution of chaincode.
References
1. Rajput, V.U.: Research on Know Your Customer (KYC). Int. J. Sci. Res. Publ. 3,
541–546 (2013)
2. Parra Moyano, J., Ross, O.: KYC optimization using distributed ledger technology.
Bus. Inf. Syst. Eng. 59(6), 411–423 (2017)
Optimizing e-KYC Process Using Distributed Ledger Technology 145
3. Munserman, R., Blackman, A.: ING to Pay $900 Million to End Dutch
Money Laundering Probe. https://www.bloomberg.com/news/articles/2018-09-
04/ing-to-pay-784-million-in-fines-to-settle-dutch-criminal-case. Accessed 13 Aug
2019
4. Martens, D., Van Serooskerken, T., Alexander, V., Steenhagen, M.: Exploring the
potential of blockchain for KYC. J. Digit. Bank. 2(2), 123–131 (2017)
5. Lootsma, Y.: Blockchain as the newest regtech application-the opportunity to
reduce the burden of KYC for financial institutions. Bank. Finan. Serv. Policy
Rep. 36, 16–21 (2017)
6. Ben-Sasson, E., et al.: Zerocash: decentralized anonymous payments from Bitcoin.
In: IEEE Symposium on Security and Privacy, pp. 1–56 (2014)
7. Curry, M.: Blockchain for KYC: game-changing RegTech innovation, IBM
RegTech In-novations. https://www.ibm.com/blogs/insights-on-business/
banking/blockchain-kyc-game-changing-regtech-innovation/. Accessed 13 Aug
2019
8. Shyft. White paper. https://www.shyft.network/pdfs/Shyft-Network-Inc-
Whitepaper V3.1.pdf. Accessed 13 Aug 2019
9. KYC-CHAIN. https://kyc-chain.com/. Accessed 13 Aug 2019
10. KYC Legal. https://kyc.legal/. Accessed 13 Aug 2019
11. Fides. https://www.norbloc.com/. Accessed 13 Aug 2019
12. KYC Services. https://ihsmarkit.com/products/kyc-services.html. Accessed 13
Aug 2019
13. Nakamoto, S.: Bitcoin: A Peer to Peer Electronic Cash System. https://bitcoin.
org/bitcoin.pdf. Accessed 13 Aug 2019
14. Buterin, V.: A Next-Generation Smart Contract and Decentralized Applica-
tion Platform. https://github.com/ethereum/wiki/wiki/White-Paper. Accessed 13
Aug 2019
15. Hyperledger Fabric v1.4: Protocol Specification. https://github.com/hyperledger/
fabric/tree/release-1.4/docs. Accessed 13 Aug 2019
16. Benet, J.: IPFS Content Addressed, Versioned, P2P File System. https://
github.com/ipfs/ipfs/blob/master/papers/ipfs-cap2pfs/ipfs-p2p-file-system.pdf.
Accessed 13 Aug 2019
17. Hyperledger Caliper. https://www.hyperledger.org/projects/caliper. Accessed 13
Aug 2019
Detecting Fake Image: A Review
for Stopping Image Manipulation
Abstract. Data manipulation getting bigger threat day by day with the
dynamic tech touch for the time being. Image is represented by underly-
ing pixelated data consisting by its area elements. By the blessings and
high availability of smart technology and device, images took important
part of humans memorable life events. This is the evidence with most
consideration for significance by human eye view. A true image can be a
big game player both in social and practical situation. Moreover the tech-
nological manipulation of an image named fake image can make violation
in major perspective consideration rather than any thinking flow or data
to obtain the difference of right or wrong. Here the detailed information
obtained from the conducted a literature review on the fake detection
identification techniques is presented. The review paper contains infor-
mation related to different fake image detection techniques instead of
making detection true and false image. Several detection techniques had
been studied like iris recognition, Support Vector Machine (SVM) and
Purkinje image-based. Simultaneously we have considered biometric sys-
tems for security aspects as well as 2D to 3D image transformation prob-
lems. For web based applications demosaicing detection method and a
colour image change splicing technology we have analyzed. Moreover we
found underwater dam methods can be used for crack detection, where
we focused on fake colorized image detection. Most importantly we have
studied on fake smile identification to enrice image forgery technology
stronger.
1 Introduction
downgraded in its reliability aspect over the course of time. The ordinary pho-
tographs were more trustworthy than digital ones. The development of photo
improvement programs such as Adobe Photoshop, Corel Paint Shop, GIMP,
Pixelmator, Photoscape, and Photoplus has played a major role in digital pho-
tography’s credibility demise. With the use of these software’s, it is very easy to
alter any photograph like iris or smile or both alternation [5,10] of an image can
change the output significantly. In addition, if there is a shadow [1] alteration
occurred the view of the considered image will be changed completely. Due to
this, it has become very difficult to decide whether a photograph is original, or
it is altered. This, in turn, has lost its trustworthiness as a piece of evidence in
every important field. The ‘digital image tamper detection’ field is blooming as
a result, to authenticate the digital photographs. It helps to filer the original
images from the altered photographs.
With the great ease of computer graphics and digital imaging, the content
of the image can be altered very easily. None can judge the image if it is real or
a fake one. To make a forgery picture, two resources are usually used, these are
copy-move an image and splicing. In this case, a part of the image is duplicated
and then pasted on different other areas to coat up any unwanted part in the
identical picture. The capabilities of multimedia devices and mobile photography
have increased image fakeness. This activity spoils the name of the deserving
persons. It also influences legal decisions. So, there is a strong need for developing
a healthy framework to analyze the originality of the images. This paper brings
an overview of the current research progress on fake image identification. This
review paper provides information about the work carried out by the different
authors. Here the work is classified based on the image classification technique.
It shows the most effective and most used techniques.
2 Literature Review
2.1 Fake Iris Detection
Precise iris segmentation and localization are the most important objects dur-
ing the implementation of the iris recognition method. The efficiency of the iris
segmentation also matters a lot from an eye reflection that determines the rate
of success for a featured extraction algorithm. With the specific method, we
can easily calculate the accurate outer iris boundary through the tracing feature
of the object all the time irrespective of the shape & size. Inner iris boundary
gives us the result after comparing the two eye images of equal titles at various
intensities so that it can efficiently detect the pupil size. Moreover, the variation
in pupil size is implemented at a time of iris detection. Hence, this approach is
proved more useful than any other based on the fake iris spoofing. This method
has highly encouraging techniques for the iris recognition system. The vital algo-
rithm is implemented through the Phoenix database of 384 images both eyes of
64 subjects. Furthermore, the profit rate in the particular iris localization from
eye image which is more than 99% always. It also removes the possibilities of all
losses of iris texture characteristics in the spatial domain that makes it distinct
148 J. I. Lubna and S. M. A. K. Chowdhury
from other conventional methods. And processing time is also great compared
to these techniques [24].
An analysis on [12] the usage of Support Vector Machine (SVM) for detecting
fake iris in biometric authentication systems. Iris recognition technology can be
certain drawbacks like usage of an artificial eye or printed contact lens. So,
detection of the fake iris is important. This paper proposes the usage of Support
Vector Machine (SVM) as it has high classification ability and is user-friendly.
It is suggested to categorize lower half of Iris patter as Region of Interest (ROI)
to avoid corruption of image. Properties like contrast, correlation, energy and
homogeneity in a Grey Level Co-occurrence Matrix (GLCM) can be useful to
extract textural features. Once the feature values are obtained, a vector is formed
and then classified by Support Vector Machine (SVM). If the output is positive,
it means that it is a live iris. The experiment is carried out using MATLAB.
The total sample size is 2250 images, 1100 are used for testing and the rest for
training. The main issue with the Grey Level Co-occurrence Matrix (GLCM)
approach is the high amount of computing resources required. To solve this,
during the scaling process, grey levels are set at a lower number. Even with this
change, the rate of correct classification is 100% and the average execution time
is 31.3 ms for feature extraction and 14.7 ms for classification [27]. Hence the
scheme can have practical applications.
In order to detect fake iris, a new method, based on Purkinje image has been
proposed. Although previous research has been done on this topic, it has serious
drawbacks [14]. This study proposes to use collimated Infrared Light Emitted
Diode (IR-LED) in order to calculate theoretical positions between the images
and enhance the fake detection algorithm. Based on the focus value, we can
discern if the image is focused enough. The proposed structure for the iris camera
includes dual Infrared Light Emitted Diode (IR-LEDs), two collimated Infrared
Light Emitted Diode (IR-LEDs) and a Universal Serial Bus (USB) camera with
a Charged Coupled Device (CCD) sensor. Designing a shaping model of the
Purkinje image using Gullstrand eye model makes it difficult to use a fake iris
to bypass the system. Based on theoretical values, it is possible to calculate the
exact location where the images are formed. For the experiment, 30 people tried
to recognize 10 times each, resulting in a total of 300 images. In addition, 15
counterfeit samples were used [13].
Several programs have been launched by the government for national iden-
tification like Aadhar in India. It includes iris biometric that presents a unique
identity to millions of people. With the passage of time and development of
modern technologies, iris recognition has become vulnerable to advanced sensor
spoof attacks [30]. In this paper, researchers have investigated another way to
deal with consequently distinguish artificial eye/iris pictures displayed by the
client to the genuine iris imaging sensor for trading off the trustworthiness of
sent iris acknowledgment framework. Recent research has revealed the use of cos-
metic texture lenses to deceive iris recognition devices. Experimental data has
been achieved by performing tests on record obtained from the publicly altered
iris image database. The development of an anti-spoofing technology is based on
Detecting Fake Image: A Review for Stopping Image Manipulation 149
the global and localized area surrounding the iris, the iris created an application
at the same time uses the highlights from the iris area and the whole eye picture
so as to carry out a solid examination of genuine or fake iris images [14]. The
component considered in this technology utilizes powerful descriptors for the
force conveyance, surface irregularity, edge quality, and surface range to assess
the validness of the iris pictures.
The security of biometric systems is a very concerning topic these days [26].
By using various sophisticated techniques like face masking and gummy fingers
spoofing attacks are being conducted on the biometric systems to gain illegal
access to systems and data. It has been an area where a lot of research has
been done since such fraudulent attacks have been on the rise [19]. A robust and
efficient system is needed to meet the challenges that this threat poses. A new
method to detect fraudulent biometric impression uses an assumption that the
image quality and properties from a real sample will be different from a fake
image sample collected during a hack attack on a biometric system and uses a
very basic but robust level of complexity to ensure it is practical for use in real-
time [25]. It uses livens detection via quality assessment as well as many other
traits of biometrics. The system uses Support Vector Machine (SVM) Classifier
to differentiate between a real and fake sample. This system can also in the
future use the sample parameters to test hand geometry, finger or vein prints by
using further detailed quality parameters for testing.
Biometric Detection of persons has become one of the main tools for the orga-
nizations to know the legitimacy of the photo. For the real authentication of the
images, some organizations have also started the implementation of some authen-
tic protection steps which is a valuable process in reconstructed samples. Novel
software-based fake detection method will be useful during the multi-biometric
systems which will work like obstacles for the frauds and fraudulent cases will
be detected efficiently in less time. Hence, the introduction motive proves which
will drastically work like a reliable ally for organizations enhancing their security
all round the clock [4,9]. Besides, by adding the feature of enhancing the secu-
rity of the biometric recognition framework, organizations also add the action
assessment swiftly and most importantly they’ll be user-friendly. Hence, after the
inclusion of the specific method, we can also neglect the complexity by which
method is drastically implemented in the real-time applications with the most
commonly used 25 images with their quality features extracted from any single
image which is needed during the authentication. This specific image need for
these common images will also identify the legitimate or impostor samples at a
time. The analysis of the common image quality through a particular method
will also assist organizations in getting highly valuable information and thus it
also discriminates itself from the fake characteristics of the obtained result [22].
150 J. I. Lubna and S. M. A. K. Chowdhury
Data about image change detection is very crucial in many areas like satellites.
They use image matching algorithm to provide quick, convenient and cost effec-
tive data to detect change. Image change detection is very useful for urban &
rural planning and monitoring environmental changes [18]. One of the biggest
bottlenecks in this area of research is a 2D format of images that make it very dif-
ficult to separate man-made constructions like buildings, highways, and bridges
from natural changes in the landscape. For that, this paper discusses a new
system which obtains 3D images via stereo technology to resolve this problem.
Kullback Leibler Divergence (KLD) is used to compare the similarity of two
images, the results are combined with the Digital Surface Model (DSM) images
using Dempster Shafer Fusion theory (DS Fusion). Radiometric information at
pixel level does not suffice to the cause thus object-based descriptions of the
images are used to get the accurate data. This new method uses the Digital
Surface Model (DSM) technology and algorithms and the data generated from
the stereo data derived from the space can be a very useful and reliable source.
Copy move is one type of forgery to create a fake image. It is a complex issue
and very difficult to detect. In this paper, the researcher proposed a method
to identify copy-move forgery [17]. Despite leaving no trail to detect forgery,
tampering an image to do region duplication damages the image at the pixel
level. Their method uses the robust color moments derived from the image and
makes it possible to detect forgery. The method uses the HSV Histogram based
method and the features retrieved from the image detects duplicated regions in
the image, since any region duplication or copy-move forgery requires duplicated
regions to be used from the same image and thus the field of the copied image
properties are going to be similar to one part or another of the same image due
to duplication. This is a very robust and efficient method to catch duplicate
elements. The proposed system is very efficient in working over two separate
databases and has shown great results in the detection of the copied section. In
a large database of non-tampered and forged images, this system will successfully
detect images which have been subject to region duplication and thus will be
very useful in detecting multiple attacks of copy-move forgery.
Biometric systems in place at the moment are not capable enough to withstand
spoofing attacks which gain illegal access to systems & data by falsifying the
identity. This developed system uses image quality assessment & texture analysis
of an image to determine whether the face on the other end is a person or just a
face print [16]. The system is based on image quality assessment and differences
between the two samples’ artifacts. The difference between a no rigid 3D human
face and a rigid planar photograph of a face is exploited by this technique and
the difference between image pigments, light reflections & shadows are used to
identify fake images [29]. By using Local Binary Patterns (LBP) technique, the
method identifies printing quality defects to make the distinction & it’s a robust,
quick & can be done without end-user approval. This method can also be used
for face recognition & has given successful results on testing is done on a publicly
available database which consisted of several real and many fake images and the
testing results were very accurate. This system can be successfully implemented
in other spoofing attacks that use masks or 3D models on the basis that human
skin has a unique texture which is almost impossible to replicate in that level of
detail from a mask [21].
submerged condition commotion, the ship’s very own clamor and resonation sig-
nal, Doppler move and engendering misfortune during the time spent acoustic
transmission brings the sonar picture issues as low complexity, solid dot commo-
tion, fluffy objective shape, etc. Low Resolution, poor picture quality, less valu-
able data, and different issues, genuinely influencing the submerged discovery
and tasks. The picture division calculation dependent on k-implies bunching is
examined, and the ideal grouping number screening and sonar picture subsurface
division are figured it out. The sonar picture quick division calculation dependent
on the Independent Chip Model (ICM) calculation and the item form location of
sonar picture dependent on level set strategy is acknowledged in MATLAB [27].
The outcomes demonstrate that the proposed calculation can improve the com-
motion decrease impact of the sonar picture under resonation obstruction and
get a superior picture recognition impact. Joined with the underlying division
aftereffects of k-implies grouping, the last division is accomplished by refresh-
ing the markings through the iterative restrictive model (ICM). The guideline
and imperfections of the C-V model division steady four-stage in the level set
are talked about in detail. Considering the improved calculation proposed by
Professor Li Chumming, we can separate the framework of the objective in the
picture precisely, which have tremendous flexibility.
that their SVD scheme is sensitive for any minute area modified in any image
and it is great and effective in detecting the false image.
Efficiency & automation in the process of detecting wheel treat defects is the
need of the hour and traditional methods like matching key points are inadequate
to detect the wheel tread defects and that is why a new method was proposed
that uses wheel tread extraction & registration [15]. This system, first of all,
detects the defects of the tread. By using Helmholtz Principle, lines & ellipses
in images are detected, registrations are then performed between the template
& under detecting tread image. Finally, the differences in pre and post-testing
images identify tread defect candidates which are comparing two images in ques-
tion for the tread defects. The testing shows that results of up to 96% could be
achieved between under detected treads & the template. The defect can be iden-
tified accurately according to the results of this method. This solves the problem
of speed, automation & accuracy of tread defect location identification. The next
stage of testing & development plans to recognize the type of defects as well. By
using SVM classifiers, detailed information of the defect like the depth and the
area can be retrieved along with the type of the defect. Labeled images may also
be used to train a CNN model that predicts the type of defects.
Medical science drastically depends upon the digital images these days which
has also increased the query solving efficiency during the authentication process
of the images [8]. Authenticity verification is one of the most vital processes for
scientific research of the digital image. With the help of authentication, we can
also obtain the purpose like forensic investigations and governments documents
verification greatly. Robust image editing software like Microsoft paint & Pho-
toshop helps us in tampering through the digital image kept for spiteful usage
anytime. Furthermore, fake insurance claim cases in the medical sector have
also been increasing in the past few years so the medial sector has also started
implementing medical imaging. Through the past process, Researchers applied
the algorithm that could address the problem of detecting and localize imaging.
Besides, the algorithm drastically depends upon the firmly based representation
of images implementing the discrete wavelet transform method that will also
assist in detection and localization in tempering. Our study shows the algorithm
is thoroughly based on some robust procedures which will remove the threat of
harmful manipulation and thus these are also sensitive for the tempering process
occurring. The proposed algorithm is great in many aspects as well, where the
less computational resource consumption has also been discussed consequently.
As of late, the railroad business in China demonstrates a quick improvement.
Railway technology in china is not that advanced like European countries and
Japan. This paper centers around the detection of track hurdle based on image
processing, canny edge detection algorithm. They are focused on railroad imped-
iment location strategies dependent on the qualities of the tracks; additionally
advanced the identification window idea to lessen the obstruction of the mind-
boggling foundation to disentangle the rail deterrent discovery. In conclusion,
they affirm through tests that the multi-step recognition program undergoes a
compelling impact in obstacle automatic detection. Through an examination of
156 J. I. Lubna and S. M. A. K. Chowdhury
the three techniques referenced above, we arrive at the resolution that “Strat-
egy dependent on the trustworthiness of the Rails and Sleepers” is most precise,
notwithstanding, somewhat, the other two strategies rely upon the light condi-
tion and deterrent hues. For rapid trains, the braking separation is a lot bigger
than the separation that the camera can identify; essentially utilizing the pic-
ture acknowledgment may go out on a limb. In any case, as an optional discovery
technique, picture recognition can consequently perceive harms of train tracks
rather than the human eye, simultaneously, it additionally provides the funda-
mental pieces of information to encourage investigation of the tracks and train
exemptions (Fig. 1).
Fig. 1. Figure shows some comparison and accuracy of methods year wise.
The edge identification examinations of two genuine pictures are led by meth-
ods for two calculations. The similar trial results demonstrate the new calculation
of programmed edge is powerful [30]. The outcomes are additionally superior to
Detecting Fake Image: A Review for Stopping Image Manipulation 157
the old style Otsu techniques. The improved Sobel administrator and heredi-
tary calculations are used to upgrade division edge of inclination picture and
proposed another programmed ideal limit calculation. The recent calculation
defeated numerous deficiencies of traditional Sobel administrators, for example,
over-segmentation and affect ability. Experiments demonstrate that the estima-
tion speed and hostile to commotion capacity of the new calculations get more
grounded. Be that as it may, we may likewise locate the identified edges are not
fine enough, and a superior refining administrator will be created later on.
3 Conclusions
Research on fake image identification is one of the most preferred topics among
researchers. A number of researchers have come forward to carry on the research
on this topic. This kind of research helps to human mankind to resolve the prob-
lems related to the image forgery. The review of different classification methods
is identified. Mainly the researcher’s used some techniques for detecting fake
image identification. And they are Histogram of Oriented Gradients (HOG) [1],
Fake Iris Detection, Fake Biometric Detection, Image Change Detection, Demo-
saicing Detection, Crack Detection, Fake Smile Detection, Fake Colorized Image
Detection, Region duplication detection, Face Spoofing Detection and etc. [6].
The detailed comparison of the different methods and their relative pros and
cons are given in this paper. The MATLAB 9.0 algorithm for iris detection was
developed. It is tested with 1 GB ram on the 2.4 GHz CPU. And used the CASIA
Iris database was chosen for tests, which is accessible in the public domain. The
database comprises of 30 pictures (320 * 280) photographic and each picture con-
sists of 3 distinct pictures placed. For segmentation based methods Daughman’s
algorithm can give better results. SVM, PCA, and ICA can give better feature
extraction facilities and can give much better fake image identification. In IRIS
based fake image identification the quality evaluation can be done by Laplacian
of Gaussian (LoG). IRIS recognition is the best method in fake image identifi-
cation. Fake smile identification methods can be combined with the above fake
image identification methods to combine emotional effects with image identifi-
cation.
References
1. Arulananth, T., Sujitha, M., Nalini, M., Srividya, B., Raviteja, K.: Fake shadow
detection using local histogram of oriented gradients (HOG) features. In: 2017
International Conference of Electronics, Communication and Aerospace Technol-
ogy (ICECA) (2017). https://doi.org/10.1109/iceca.2017.8212765
2. Bachtiar, M., Gusti, D., Wijaya, I., Hidajat, M.: Web-based application develop-
ment for false images detection for multi images through demosaicing detection.
In: 2018 International Conference on Information Management and Technology
(ICIMTech) (2018). https://doi.org/10.1109/icimtech.2018.8528175
158 J. I. Lubna and S. M. A. K. Chowdhury
3. Ballado, A., et al.: Philippine currency paper bill counterfeit detection through
image processing using Canny Edge Technology. In: 2015 International Conference
on Humanoid, Nanotechnology, Information Technology, Communication and Con-
trol, Environment and Management (HNICEM) (2015). https://doi.org/10.1109/
hnicem.2015.7393184
4. Bhakt, N., Joshi, P., Dhyani, P.: A novel framework for real and fake smile detection
from videos. In: 2018 Second International Conference on Electronics, Communi-
cation and Aerospace Technology (ICECA) (2018). https://doi.org/10.1109/iceca.
2018.8474594
5. Bodade, R., Talbar, S., Batnagar, A.: Dynamic iris localisation: a novel approach
suitable for fake iris detection. In: National Conference on Signal and Image Pro-
cessing Applications (2009). https://doi.org/10.1049/ic.2009.0123
6. Bulla, A., Shreedarshan, K.: Fake shadow detection using local HOG features. In:
2016 IEEE International Conference on Recent Trends in Electronics, Information
& Communication Technology (RTEICT) (2016). https://doi.org/10.1109/rteict.
2016.7808043
7. Chen, C., Wang, J., Zou, L., Fu, J., Ma, C.: A novel crack detection algorithm of
underwater dam image. In: 2012 International Conference on Systems and Infor-
matics (ICSAI2012) (2012). https://doi.org/10.1109/icsai.2012.6223399
8. Gadhiya, T., Roy, A., Mitra, S., Mall, V.: Use of discrete wavelet transform
method for detection and localization of tampering in a digital medical image.
In: 2017 IEEE Region 10 Symposium (TENSYMP) (2017). https://doi.org/10.
1109/tenconspring.2017.8070082
9. Galbally, J., Marcel, S., Fierrez, J.: Image quality assessment for fake biometric
detection: application to iris, fingerprint, and face recognition. IEEE Trans. Image
Process. 23(2), 710–724 (2014). https://doi.org/10.1109/tip.2013.2292332
10. Gunadi, I., Harjoko, A., Wardoyo, R., Ramdhani, N.: Fake smile detection using
linear support vector machine. In: 2015 International Conference on Data and Soft-
ware Engineering (ICoDSE) (2015). https://doi.org/10.1109/icodse.2015.7436980
11. Guo, Y., Cao, X., Zhang, W., Wang, R.: Fake colorized image detection. IEEE
Trans. Inf. Forensics Secur. 13(8), 1932–1944 (2018). https://doi.org/10.1109/tifs.
2018.2806926
12. He, X., An, S., Shi, P.: Statistical texture analysis-based approach for fake iris
detection using support vector machines. In: Lee, S.-W., Li, S.Z. (eds.) ICB 2007.
LNCS, vol. 4642, pp. 540–546. Springer, Heidelberg (2007). https://doi.org/10.
1007/978-3-540-74549-5 57
13. Hou, B., Wei, Q., Zheng, Y., Wang, S.: Unsupervised change detection in SAR
image based on gauss-log ratio image fusion and compressed projection. IEEE J.
Sel. Top. Appl. Earth Obser. Remote Sens. 7(8), 3297–3317 (2014). https://doi.
org/10.1109/jstars.2014.2328344
14. Lee, E.C., Park, K.R., Kim, J.: Fake iris detection by using purkinje image. In:
Zhang, D., Jain, A.K. (eds.) ICB 2006. LNCS, vol. 3832, pp. 397–403. Springer,
Heidelberg (2005). https://doi.org/10.1007/11608288 53
15. Lv, S., Zhou, F., Wei, Z.: Train wheel tread defects detection based on image reg-
istration. In: 2017 IEEE International Conference on Imaging Systems and Tech-
niques (IST) (2017). https://doi.org/10.1109/ist.2017.8261509
16. Maatta, J., Hadid, A., Pietikainen, M.: Face spoofing detection from single images
using micro-texture analysis. In: 2011 International Joint Conference on Biometrics
(IJCB) (2011). https://doi.org/10.1109/ijcb.2011.6117510
Detecting Fake Image: A Review for Stopping Image Manipulation 159
17. Malviya, A., Ladhake, S.: Region duplication detection using color histogram and
moments in digital image. In: 2016 International Conference on Inventive Computa-
tion Technologies (ICICT) (2016). https://doi.org/10.1109/inventive.2016.7823199
18. Patil, R., Pete, D.: Image change detection using stereo imagery and digital surface
mode. In: 2015 International Conference on Information Processing (ICIP) (2015).
https://doi.org/10.1109/infop.2015.7489376
19. Pravallika, P., Prasad, K.: SVM classification for fake biometric detection using
image quality assessment: application to iris, face and palm print. In: 2016 Interna-
tional Conference on Inventive Computation Technologies (ICICT) (2016). https://
doi.org/10.1109/inventive.2016.7823189
20. Pritam, D., Dewan, J.: Detection of fire using image processing techniques with
LUV color space. In: 2017 2nd International Conference for Convergence in Tech-
nology (I2CT) (2017). https://doi.org/10.1109/i2ct.2017.8226309
21. Rebhi, A., Abid, S., Fnaiech, F.: Texture defect detection method based on H-
image and Hotteling model T2 . In: 2014 1st International Conference on Advanced
Technologies for Signal and Image Processing (ATSIP) (2014). https://doi.org/10.
1109/atsip.2014.6834589
22. Reno, A., David, D.: An application of image change detection-urbanization. In:
2015 International Conference on Circuits, Power and Computing Technologies,
ICCPCT 2015 (2015). https://doi.org/10.1109/iccpct.2015.7159368
23. Rosario-Torres, S., Velez-Reyes, M.: Speeding up the MATLABTM hyperspectral
image analysis toolbox using GPUs and the jacket toolbox. In: 2009 First Work-
shop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing
(2009). https://doi.org/10.1109/whispers.2009.5289089
24. Sanchez, C., Niemeijer, M., Suttorp Schulten, M., Abramoff, M., van Ginneken,
B.: Improving hard exudate detection in retinal images through a combination
of local and contextual information. In: 2010 IEEE International Symposium on
Biomedical Imaging: From Nano to Macro (2010). https://doi.org/10.1109/isbi.
2010.5490429
25. Tan, C., Kumar, A.: Integrating ocular and iris descriptors for fake iris image
detection. In: 2nd International Workshop on Biometrics and Forensics (2014).
https://doi.org/10.1109/iwbf.2014.6914251
26. Wang, W., Dong, J., Tan, T.: Effective image splicing detection based on image
chroma. In: 2009 16th IEEE International Conference on Image Processing (ICIP)
(2009). https://doi.org/10.1109/icip.2009.5413549
27. Xinyu, T., Xuewu, Z., Xiaolong, X., Jinbao, S., Yan, X.: Methods for underwater
sonar image processing in objection detection. In: 2017 International Conference
on Computer Systems, Electronics and Control (ICCSEC) (2017). https://doi.org/
10.1109/iccsec.2017.8446701
28. Yao, T., Dai, S., Wang, P., He, Y.: Image based obstacle detection for automatic
train supervision. In: 2012 5th International Congress on Image and Signal Pro-
cessing (2012). https://doi.org/10.1109/cisp.2012.6469703
29. Zhang, L., He, X.: Fake shadow detection based on SIFT features matching. In:
2010 WASE International Conference on Information Engineering (2010). https://
doi.org/10.1109/icie.2010.58
30. Jin-Yu, Z., Yan, C., Xian-Xiang, H.: Edge detection of images based on improved
Sobel operator and genetic algorithms. In: 2009 International Conference on Image
Analysis and Signal Processing (2009). https://doi.org/10.1109/iasp.2009.5054605
A View on the Blockchain as a Solution
to the Healthcare Industry: Challenges
and Opportunities
1 Introduction
Generally, the modern healthcare systems provide more facilities to the doc-
tors to gather, analyze, and monitor the health information of remote patients
due to the advanced computing approaches such as the Internet of Things (IoT)
and clouds. Technology development has eroded the privacy and security protec-
tion of our data. The privacy and integrity of the healthcare systems information
must be protected from an unauthorized user, the internal, and external attacks.
The extensive research work is going on to secure the health care systems such
as Electronic Medical Report (EMR), Electronic Health Records (EHR), Health
Information System (HIS), and Personal Health Records (PHR) [1]. Recently,
researchers have been proposed many cryptographic algorithms to ensure the
privacy and security of the healthcare system. Healthcare system consists of
a large diversity of data’s are generated, scattered, stored and retrieved. For
example, consider the patient undergoes to treatment; the patient needs to take
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 160–169, 2020.
https://doi.org/10.1007/978-981-15-3666-3_14
A View on the Blockchain as a Solution to the Healthcare Industry 161
various tests, scans, CT scan, etc. The data will be available at the test cen-
ter as well as with a physician. The data could be accessed by a physician in
the same hospital as well as a physician in another hospital. The advancement
of technology plays a vital role to provide security, privacy, and integrity of
health care data management system. Blockchain is a tamper-proof distributed
or decentralized online database which is undisputable, non-rewritable, read a
dissimilar data structure, consists of the large number of blocks which are linked
together using non alterable key referencing mechanism. Blockchain is a network
of independent entities which work together and form a peer to peer network.
The blocks of data which is stored in the network are synchronized together
using the consensus mechanism and offers a validity of data over the network.
The blockchain [2] consists of a list of blocks called a data structure. The data
structure of blockchain consists of the following elements: timestamp, the hash
value of the previous block, transaction data. The properties of data structure
are as follows: (1) secured blocks (data records) (2) peer to peer network (3)
consensus mechanism (4) security mechanism which offers unchallengeable of
the data. These four properties play vital role in blockchain mechanism for the
important consequences. The data availability of blockchain is either permission
or permissionless. The block-chain is distributed ledger, which does not depend
on a third party to perform transaction between two entities. The transaction
between the two parties can be carried out using consensus rules to ensure the
trustworthy of the parties. The trustworthy can be ensured by the participants
in the network instead of any centralized parties. The inherent characteristics
of the blockchain technology are the transaction cannot be changed anymore.
The security of blockchain [3] depends on the cryptographic encryption func-
tion and consensus mechanism. The large resources are needed to carry out
the malicious attack on the block of transaction. The complexity of attacks as
increases whenever the new block of the transaction is attached to the chain of
an existing transaction block. The aforementioned features of blockchain offer
the following services such as integrity, privacy, security, traceability, and non-
repudiation of data while storing the information in a decentralized way. The
exhaustive research works have been carried out in the field of the banking
sector, real estate, government bid, and finance using blockchain. Recently, the
researchers started to address the potential of using blockchain technology in the
field of healthcare. The blockchain technology provides the solution to the exist-
ing challenges in healthcare applications. The main contribution of this paper
is to addresses the challenges, possible solution and its of blockchain technology
in the healthcare applications. The organization of this paper as follows: Sect. 2
describes the literature of the healthcare system. Section 3 discusses the litera-
ture of blockchain. Section 4 discusses the uses of blockchain for the healthcare
industry. Section 5 concludes this paper.
162 Sharmila et al.
Blockchain has many features that can be used for healthcare. The features of
blockchain are as follows:
• Decentralized storage
• Authentication
• Disintermediation
• Cost reduction
• Immutability
A View on the Blockchain as a Solution to the Healthcare Industry 165
• Privacy leakage
• Lack of Standardization
• Key Management
• Scalability issues due to IoT
• Software Vulnerabilities
166 Sharmila et al.
4.2 MedRec
MedRec [5] is decentralized data management system used to handle EMRs using
blockchain which was proposed by Azaria, Ekblaw, Vieira, and Lippman. In this
MedRec, the stakeholders of the healthcare industry participated as the miners
and access, aggregate, mining data as an anonymous and securing network using
PoW. The main limitation not addressed by MedRec is a key replacement and
legal issues.
5 Conclusion
In conclusion, blockchain offers a solution for some of the challenges faced by the
healthcare industry. Blockchain facilitates secure data access and sharing, secure
the privacy of the patients, patients can keep track of medical records, etc. In
spite of complex design of healthcare system, the blockchain resolves the key
challenges of the healthcare industry including data dissemination, concurrent
access, a huge amount of data processing, access control, limit the storage of data
and cost of data processing. Blockchain not only offers a wide range of applica-
tion in the healthcare industry but it also offers vast research opportunities on
employing blockchain in the healthcare industry.
References
1. Abouelmehdi, K., Beni-Hessane, A., Khaloufi, H.: Big healthcare data: preserving
security and privacy. J. Big Data 5(1), 1–18 (2018). https://doi.org/10.1186/s40537-
017-0110-7
2. Mettler, M.: Blockchain technology in healthcare: the revolution starts here. In:
IEEE 18th International Conference on e-Health Networking, Applications and Ser-
vices (Healthcom) (2016)
3. Zhang, J., Xue, N., Huang, X.: A secure system for pervasive social network-based
healthcare. IEEE Access 4(1), 9239–9250 (2016)
4. Torre, I.: A framework for personal data protection in the IoT. In: 11th International
Conference for Internet Technology and Secured Transactions, ICITST (2016)
5. Cai, Y., Zhu, D.: Fraud detections for online businesses: a perspective from
blockchain technology. Finan. Innov. 2(1), 20 (2016). https://doi.org/10.1186/
s40854-016-0039-4
6. Kiyomoto, S., Rahman, M.S., Basu, A.: On blockchain-based anonymized dataset
distribution platform. In: 2017 IEEE 15th International Conference on Software
Engineering Research, Management and Applications (SERA) (2016)
7. Sharmila, Kumar, D., Som, K., Kumar, P., Chaudhary, K.: General outlook of wire-
less body area sensor networks. In: Singh, M., Gupta, P., Tyagi, V., Flusser, J.,
Ören, T., Kashyap, R. (eds.) ICACDS 2019. CCIS, vol. 1046, pp. 58–67. Springer,
Singapore (2019). https://doi.org/10.1007/978-981-13-9942-8 6
8. Swan, M.: Blockchain: Blueprint for a New Economy. O’Reilly Media Inc., Newton
(2015)
9. Zheng, S., Xie, H., Dai, X., Chen, H.: An overview of blockchain technology: archi-
tecture, consensus, and future trends. In: 2017 IEEE International Congress on Big
Data (BigData Congress), pp. 557–564. IEEE (2017)
An Approach Towards Encrypting Paired
Digits Using Dynamic Programming
and Diffie-Hellman Key Exchange
1 Introduction
With the advancement in security mechanisms the level of attacks have also
improved significantly. In cryptography the algorithm is as strong as the key asso-
ciated to it. Hence protecting the integrity of the key is very much essential. Public
key Cryptography like RSA uses the discrete logarithm problem to encipher the
plain text. Many equal blocks of plain text are taken separately and applied with
RSA in each block. RSA inherently uses discrete logarithm problem to encipher
each block of plain text. Such discrete logarithm problem takes enough amount
of time to encipher each block of data. Unfortunately, there are many thousands
block of data in a plain text. Then, ultimately we will take huge amount of time
and demand high computation power or require upgraded architecture.
One of the mechanisms of secure key transfer is the Diffie Hellman Key
Exchange. This section briefly discusses about the idea and methodology that has
been used in this cryptography technique. In [5] the author has discussed about the
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 170–181, 2020.
https://doi.org/10.1007/978-981-15-3666-3_15
An Approach Towards Encrypting Paired Digits 171
key exchange to multiple users using Logical Key Hierarchy (LKH) and stated the
reliability of Diffie-Hellman key exchange to multiple peers by broadcasting. In [6]
the paper basically discusses about the improvement of Kerberos protocol based on
Diffie-Hellman key exchange and they have shown the results, demonstrating that
the proposed system gives better results by reducing password guessing attacks. In
[7,8] the author has proposed a better modification of Diffie-Hellman key exchange
for Zero Knowledge Proof (ZKP) and has designed an algorithm in order to achieve
it. They have proved ZKP concept by using various situations of key exchanges and
analyzed them. Similarly, many authors have analyzed different situations of key
exchange to demonstrate that Diffie-Hellman key exchange is reliable and easy to
implement. Diffie Hellman is vulnerable to man-in-the-middle attack in wireless
communication, as it is not able to verify the communication. In [9] the author
proposes a improved version of Diffie Hellman Key Exchange using String com-
parison mechanism. Diffie Hellman key exchange can be used in embedded devices
[10] this is achieved by using socket communication through the network. In [11]
the author presents a secure cryptographic method of Key Generation and Shar-
ing using Diffie Hellman Key Exchange Protocol, after the key generation the mes-
sage is encrypted using AES and sent to the recipient. Elliptic Curve Cryptography
has the capabilities of encryption, Digital Signature and Key exchange, in [12] the
author presents a Elliptic Curve Diffie Hellman key exchange. Gowda in [13] pro-
posed a variation of Diffie Hellman that determines the pixel position of the image
to hide data, the public key exchange is done using Diffie Hellman, where the pub-
lic key is a product of the public key and the private key. In this [14] the author
uses a shared secret key for encryption and decryption, this key is shared using
Diffie Hellman Key Exchange, DNA Hybridization technique is used to generate
cipher text from DNA sequences. Joshi [15] uses the Diffie Hellman key exchange
for node to node authentication in a Wireless Sensor Network(WSN) they used
cryptographic concepts and cluster head mechanism to eliminate the weaknesses
of Diffie Hellman. In this paper [16] proposed three different that securely integrate
Diffie Hellman Key Exchange into Digital Signatures.
The Diffie Hellman key exchange [1] has been used here only to exchange the
common key which in turn will generate a sequence of values in tabular form.
These values in the sequence are referred as codes. These codes are considered
to be a cipher of paired digits. In order to encrypt a 2-digit number, we take
half of the digits corresponding to x-axis and the other half to be in y-axis. The
value corresponding to the intersection of the x-axis and y-axis is the required
code. By this simple idea, we will encipher the entire plain-text into cipher of
larger length. During deciphering we take back the reverse process by searching
the cipher in the tabular data set.
After implementation and analysis, it has been inferred that the cryptography
with dynamic programming enhances the speed of execution of the algorithm.
The table formed dynamically fetches the codes (refer Sect. 3) in a easy and
faster manner. Even though the table formation uses enormous system stack
if implemented through recursion but then it decreases the running time when
implemented by dynamic programming in cost of space.
172 A. K. Das et al.
The rest of the paper is segmented as follows: Sect. 2 discusses about the
preliminary assessments of Diffie-Hellman Key exchange. In Sect. 3 the proposed
methodology, the encryption and decryption mechanism is elaborated. Section 4
describes the results and analysis of the proposed method. The comparison with
RSA has been done in Sect. 5. Lastly, the scope and area of application has been
stated in Sect. 6.
2 Preliminaries
2.1 Diffie-Hellman Key Exchange
We have used this protocol in order to arrive at a common key [2]. This com-
mon key will help us in building the sequence table. The algorithm takes few
requirements into considerations. A random integer XA < q and XB < q along
with the primitive root of a prime number p whose powers modulo p generates
integers from 1 to (p − 1). Here YA and YB are published to one another for key
exchange.
The key is generated as:
Modulo(g,x,p):
1. assign i = g
2. do
(a) if (i > p)
i. i = imodp
(b) else
i. i = ai
ii. i = imodp
3. k = k − 1
4. while(k! = 1)
5. return i
An Approach Towards Encrypting Paired Digits 173
The dividend is squared each time and checked through a condition if it is greater
than divisor. If it is greater, (modp) is taken into consideration but if not the
case, the dividend is squared again. This procedure is repeated until the power
of g get reduced to 1.
and sends 24 through the channel. The 24 is received in the intruder’s buffer and
the intruder calculates 2411 mod31 = 11 and sets 11 as CK for party A. Now,
the intruder calculates 311 mod31 = 13 and sends 13 to both the parties A and
B. In return the party B calculates 317 mod31 = 22 and sends 22 to the party A.
The intruder, using the integer 22 calculates 2211 mod31 = 10 and arrives 10 as
CK for party B. This whole procedure can be done by two private keys of the
intruder.
3 Proposed Methodology
The KS (also called codes) are generated by the initial seed value obtained
from the common key (K). A table in a two-dimensional matrix which is filled
up dynamically by a recursive equation as shown in Table 1. The cell[x,y] is
represented in Table 1 as value corresponding to the intersection of x and y. The
x-axis is the rows where as the y-axis is the columns of the table. The cell[0,0]
and cell[0,1] are filled by half of the digits of the common key. For Ex. CK = 136,
then cell[x] = 13 and cell[y] = 6, CK = 1234, then cell[x] = 12 and cell[y]= 34.
It uses floor function after dividing the CK. So, we can use partition function
for splitting the CK into two half partitions. The lef tP artition function divides
the key for cell[0,0] and rightP artition for cell[0,1]. The value of c and d are
referred as seed values for Fibonacci series. In Table 1 the first row with 0th index
is filled upto 9th Index by the consecutive Fibonacci numbers [4] with the base
numbers as c and d, i.e. 1, 2, 3, 5, 8...89. The first column with 0th index is filled
by (Fibonacci number from the previous row of 9th ) + 1 as shown in Table 1.
The rest is filled by the sum of the adjacent upper, left and diagonal left cell.
For example, cell[1,1] is filled up by the sum of cell[0,1], cell[1,0] and cell[0,0]
respectively. The recursive equation is given as:
An Approach Towards Encrypting Paired Digits 175
⎧
⎨=fib(y,c,d) ; if x = 0
Cell[x, y, c, d] = =cell(x-1,9,c,d)+1 ; if y = 0
⎩
=cell(x,y-1,c,d)+cell(x-1,y-1,c,d)+cell(x-1,y,c,d); else
(4)
Where c is the left partition of CK where as d is the right one. For Ex. CK =
12, then c = 1 and d = 2.
Index
* 0 1 2 3 4 5 6 7 8 9
0 c=1 d=2 3 5 8 13 21 34 55 89
1 (89)+1 (1)+(2)+(89+1) ... ... ... ... 318 462
2 ..
3 ..
4 ..
5 ..
6 ..
7 19745070
8 167084937
9 ..
The Table 1 is generated as shown by using CK=12. Now this can only be
generated if and only if c and d are known to both the sender and the receiver.
If the adversary somehow learns the key ‘c’ and ‘d’ then he can generate the KS
and can search for the codes in the table. The cipher text (codes) are generated
in ascending order of its value in row-wise manner. The proceeding rows get
higher values than the previous rows. Thus, providing non-redundant codes in
column wise fashion. The strength of the algorithm depends on the variability of
partitioning the CK. So, it is quite obvious to say that the large is the CK, larger
the possibilities of partitioning and more the difficulty in breaking the common
key. The recursive equation as shown in Eq. (4) is the most promising part of
table generation. So, the table generator uses the concept of Non Redundant
Cipher (NRC). The NRC gives us unique values in the table so that a plain text
can have at most one cipher. In order to achieve this, the first column of the
table is filled up by next proceeding integer of the previous row corresponding
to 9th column. For Ex. cell[1,0] is filled up by cell[0, 9] + 1.
The encryption model in discussed in the following lines. To understand
the proposed model better we will refer Fig. 2 for our reference. The proposed
model starts with an input of cipher text. The cipher text is in paired form. So,
the model takes 2-digit numbers as a part of plain text as input. The message
176 A. K. Das et al.
length defines the pair of 2-digit numbers. For Example if message length is
4, there will be 4 pair of 2-digit numbers. The input is then divided into two
partitions the left-partition and the right-partition. Out of these two partitions,
the left partition is mapped to x-axis(horizontal axis) and the right partition
is mapped to y-axis(vertical axis) of the table. The values (codes) in the table
(refer Table 1) is the cipher text. It contains all the cipher text corresponding
to plain text 2-digit integer at index [0,0], [0,1], [0,2],...[9,7], [9,8] and [9,9] from
the table. Therefore, tenth place of the 2-digit integer will get mapped to x-axis
where as unit place of 2-digit number will get mapped to y-axis. The intersection
of indexes of x-axis ans y-axis is value that we want for the cipher of respective
plain text 2-digit integer. For example if one pair of the plain text is say 76
then we shall find the cipher at position [7,6]. Similarly, we can place the same
methodology for other plain message pairs as well. After calculating for all pairs,
the cipher text will be of equal or larger length than the plain text.
Fig. 2. Block Diagram for encryption showing the procedure of key splitting and taking
it into account for key generation.
The function ENCIPHER(M, CK) takes two parameters, first is the message M
and second is the Common Key(CK). The message M is a sequence of paired
digits. Here, the ’loc’ refers to the location of digit’s place where the division is
to be made for left partitioning and right partitioning. CipherBuffer is the buffer
which is present in the sender’s local machine. At the end, the CipherBuffer is
filled with cipher text and sent to the destination through the network.
An Approach Towards Encrypting Paired Digits 177
ENCIPHER(M, CK)
DECRYPT(CIPHER, CK)
Complexity in Decryption
Complexity of decryption includes computation in searching the required cipher
from the table. It takes O(100 entries) x O(k) running time in worst case. Since,
there are only 100 entries that have to be searched. So, in worst case running
time complexity is O(k).
After decryption
Output: 34 41 56 64 24 25 20 42
An Approach Towards Encrypting Paired Digits 179
The Fig. 4 shows the outputs of the decryption. It has been found that the run-
ning time during encryption is 0.08709 s where as during decryption it has taken
1.485 s. The time difference lies in procedure of encryption and decryption. The
encryption uses dynamic programming to fill up the table in O(k) time for each
entry. But on the other hand while encryption, it have used extensive searching
algorithm for finding the cipher in the table in row-major order. Then, it returns
the corresponding x-axis and y-axis value index.
6 Conclusion
This paper proposes an innovative method of Shared Key Cryptography using
the Diffie Hellman Key Exchange mechanism. Our algorithm uses dynamic pro-
gramming and recursion that speeds up the key exchange process, this has per-
formed well with higher range of key values. The performance has been improved
and the complexity has been reduced. This approach has a very wide application
in the field of financial sectors or the commercial sector. For example credit card
or debit card information or the pin for your card can be transmitted to and from
the bank servers using this mechanism as it provides security in a very efficient
manner. Information like OTP or phone numbers can be exchanged in private.
These critical sectors can use our algorithm for a better security infrastructure.
References
1. Diffie, W., Hellman, M.: New directions in cryptography. IEEE Trans. Inf. Theory
22(6), 644–654 (1976)
2. Stallings, W.: Cryptography and Network Security, 4th edn. Pearson Education,
India (2006)
3. Cormen, T.H., Leiserson, C.E., Rivest, R.L., Stein, C.: Introduction to Algorithms.
MIT Press, Cambridge (2009)
4. Horadam, A.F.: A generalized Fibonacci sequence. Am. Math. Mon. 68(5), 455–
459 (1961)
5. Bodur, H., Kara, R.: Implementing Diffie-Hellman key exchange method on logical
key hierarchy for secure broadcast transmission. In: 2017 9th International Con-
ference on Computational Intelligence and Communication Networks (CICN), pp.
144–147. IEEE (2017)
6. Hu, Z., Zhu, Y., Ma, L.: An improved Kerberos protocol based on Diffie-Hellman-
DSA key exchange. In: 2012 18th IEEE International Conference on Networks
(ICON), pp. 400–404. IEEE (2012)
7. Ibrahem, M.K.: Modification of Diffie-Hellman key exchange algorithm for zero
knowledge proof. In: 2012 International Conference on Future Communication Net-
works, pp. 147–152. IEEE, April 2012
8. Velten, M.: Zero-Knowledge, the Magic of Cryptography. Saarland University,
August 2006
An Approach Towards Encrypting Paired Digits 181
9. Taparia, A., Panigrahy, S.K., Jena, S.K.: Secure key exchange using enhanced
Diffie-Hellman protocol based on string comparison. In: 2017 International Confer-
ence on Wireless Communications, Signal Processing and Networking (WiSPNET),
pp. 722–726. IEEE (2017)
10. Deshpande, P., Santhanalakshmi, S., Lakshmi, P., Vishwa, A.: Experimental study
of Diffie-Hellman key exchange algorithm on embedded devices. In: 2017 Interna-
tional Conference on Energy, Communication, Data Analytics and Soft Computing
(ICECDS), pp. 2042–2047. IEEE (2017)
11. Yusfrizal, Y., Meizar, A., Kurniawan, H., Agustin, F.: Key management using
combination of Diffie-Hellman key exchange with AES encryption. In: 2018 6th
International Conference on Cyber and IT Service Management (CITSM), pp.
1–6. IEEE (2018)
12. Mehibel, N., Hamadouche, M.H.: A new approach of elliptic curve Diffie-Hellman
key exchange. In: 2017 5th International Conference on Electrical Engineering-
Boumerdes (ICEE-B), pp. 1–6. IEEE (2017)
13. Gowda, S.N.: An advanced Diffie-Hellman approach to image steganography. In:
2016 IEEE International Conference on Advanced Networks and Telecommunica-
tions Systems (ANTS), pp. 1–4. IEEE (2016)
14. Aieh, A., Sen, A., Dash, S.R., Dehuri, S.: Deoxyribonucleic acid (DNA) for a shared
secret key cryptosystem with Diffie Hellman key sharing technique. In: Proceedings
of the 2015 Third International Conference on Computer, Communication, Control
and Information Technology (C3IT), pp. 1–6. IEEE (2015)
15. Joshi, P., Verma, M., Verma, P.R.: Secure authentication approach using Diffie-
Hellman key exchange algorithm for WSN. In: 2015 International Conference
on Control, Instrumentation, Communication and Computational Technologies
(ICCICCT), pp. 527–532. IEEE (2015)
16. Harn, L., Mehta, M., Hsin, W.-J.: Integrating Diffie-Hellman key exchange into the
digital signature algorithm (DSA). IEEE Commun. Lett. 8(3), 198–200 (2004)
17. Mahajan, S., Singh, M.: Analysis of RSA algorithm using GPU programming.
arXiv preprint arXiv:1407.1465 (2014)
Detecting Malicious Twitter Bots
Using Machine Learning
1 Introduction
Twitter is a very famous social networking service where users from all over the
world post and interact with messages which are known as “tweets”. Founded
in 2006 by Jack Dorsey, Noah Glass, Biz Stone, and Evan Williams [1] twitter
has seen some remarkable growth over the years. According to statistics, twitter
had around 30 million users which shoot up to 335 million active users in the
2nd quarter of 2018 (Fig. 1) [2].
Twitter has become one of the most used social media platforms because it is
both personal and rapid. It gives people from different spheres of lives to express
themselves and build relationship with their followers.
Different news channels have twitter accounts to share latest news and devel-
opments. Celebrities use Twitter to build a personal connection with their fans.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 182–194, 2020.
https://doi.org/10.1007/978-981-15-3666-3_16
Detecting Malicious Twitter Bots Using Machine Learning 183
Fig. 1. Comparison between active users from 2010 to 2018 on Twitter [2]
Twitter is being used by many brands to market their products and get feed-
back from people about their brand. It helps them to improve their product.
People with same taste and interest form communities on twitter and share their
ideas and hold discussions online. Also people share interesting photographs, new
researches. Authors use twitters to promote their books and the list is endless.
Another interesting feature of twitter is bots. A bot software which controls a
twitter account with the help of Twitter Application program interface is called
Twitter bots [3]. It is like the automation of a twitter account. Automations rules
are set by twitter only. These bots can tweet, re-tweet, like, follow, unfollow,
or direct message any accounts without any human intervention. According to
study conducted by University of Southern California and Indiana University
around 15% of the active twitter accounts are bots [4]. These bots behave just
as humans and therefore it is difficult to identify them. These bots chats with
random users or posts poetry, news or photographs at regular intervals. Some of
the famous bots are:
1. @HundredZeros: Twitterbot that posts URLs to the eBooks that are freely
available on Amazon.
2. @DearAssistant: Gives detailed response to any question asked quickly.
3. @netflix bot: Tweets about new shows on Netflix.
4. @nicetipsbot: Twitter bot which posts fun little life tweets multiple times a
day and many more [5].
Owing to the popularity of twitter, its huge active audience and such inter-
esting features, it also grabs the attention of malicious and unethical users like
hackers, cyber criminals, cyber bullies. There are many malicious and political
bots on twitter which post sensitive contents and malicious URLs. Cyber crim-
inals use these bots to circulate malicious and Phishing URLs through automa-
tion and regular intervals in order to gain sensitive information of various people
leading to phishing scams and frauds. Twitter has seen an increase in the number
of malicious bots in recent years that have distributed fake news and distorted
images, malicious URLs.
184 T. Satija and N. Kar
These malicious twitter bots regularly post malicious URLs, fake news and
try to reach as many people as possible. These malicious URLs can be of some
malicious website or it can be a part of phishing scam asking for some person
sensitive information. These bots serve as a very helpful tool for hackers to
carry out data breaches. Since 2013 till today, total data breaches recorded is
14,717,618,286 [6]. Twitter has been trying to identify such accounts but due to
their human like behaviour, it is difficult to identify them and filter them out
(Fig. 2).
Fig. 2. Number of accounts challenged to be bots by twitter from Sept 2017 to May
2018 [7]
2 Background
Machine learning is considered to be a fast evolving branch which has increase the
pace of automation. Machine learning algorithms like logistic regression, SVM
helps in handling and understanding large amounts of data. These algorithms
help use to develop models which can identify patterns and take decisions with
Detecting Malicious Twitter Bots Using Machine Learning 185
less human involvement [9]. There are several machine learning algorithms which
are used to train models according to our requirements like linear regression is
used when we have one dependent and one independent variable [10]. Logistic
regression is used when more than one independent variable is there [11].
In order to implement machine learning algorithms and train our models,
we need large amount of dataset. In this paper, we are taking datasets to train
our model from Kaggle [12]. It is an online community which provides many
different kinds of datasets in different formats which can be used by machine
learners and data scientist. It allows its members to use published dataset to
build different machine learning models. Users can even publish their own dataset
on this platform which can be used by others.
Since machine learning is powerful and useful tool, researchers have been
trying to incorporate it with security. Lee, Sangho, and Jong Kim have pro-
posed a suspicious URL detection system for Twitter, WARNINGBIRD. They
have considered correlated redirect chains of URLs in a number of tweets. As
hackers have fewer resources and therefore they will have to use them again , a
portion of their redirect chains will be shared. It’s a real time system. WARN-
INGBIRD consists of 4 modules: data collection, feature extraction, training, and
classification. Data was collected from the collection of tweets with URLs and
crawling for URL redirections. From collected data, useful and unique features
were extracted which were used for training the classifier to identify malicious
URLs. In WARNINGBIRD 12 features has been identified for classifying suspi-
cious URLs on Twitter. WARNINGBIRD uses a static crawler implemented in
Python [13]. In a survey paper published in IJRASET, authors have used the
WARNINGBIRD mechanism and implemented it and have given the results [14].
Short URLs (Uniform Resource Locators) are now a days very popular in
Social media networks but at the same time detecting whether they are mali-
cious or not becomes even difficult. Raj Nepali, et.al. proposed a mechanism
using machine learning to develop a classifier to detect malicious short URLs
with visible content features, tweet context, and social features from Twitter.
The system will be implemented as Firefox’s web browser extension and will be
programmed with XML User Interface Language and Javascript to automate
the processes. The system will fetch tweets on its own and extract the features
from the data and for classification submit it to the classifier. The result will be
displayed to the user [15].
Another paper published in 2017, provides basic understanding of how
machine learning can be used to detect malicious URLs and presents information
about the related work which has been done in this domain [16]. Nikan chavoshi
et.al. proposed a warped correlation finder to identify correlated user accounts
in social media websites such as Twitter. The authors suggested that humans
are not highly synchronous for a long duration, therefore highly synchronous
accounts are most likely to be bots. The proposed model works on activity
correlation and does not require a labeled data set as compared to twitter sus-
pension technique and per-user technique [17]. Authors Novotny and Jan pro-
posed a machine learning approach to distinguish between sophisticate and less
186 T. Satija and N. Kar
There are millions of bots operating on twitter and not all of them are malicious.
The goal of this paper is to develop a system which is able to detect those
twitter bots which are posting malicious URLs. The proposed system, called
Find Malicious Bot can be divided into 3 modules. Figure 3 gives a brief about
the 3 modules.
Under this module a Twitter API is created and trending tweets are extracted
according to the required location using WOEID. A WOEID (Where On Earth
IDentifier) is a unique 32-bit reference identifier, assigned by Yahoo!, that iden-
tifies the place on Earth [21,22]. In this paper,tweets related to India and New
York were collected from Twitter using WOEID of India : 23424848 and that of
New York is 2459115.
For over a week, top 1000 treading tweets were collected daily using both
WOEID, resulting in a database of 7000 tweets for each. From all those
tweets,tweets containing any kind of URL were filtered out and all the details
of those tweets were saved in a separate database. And the URLs present in the
tweets were extracted and saved in a separate database. Details extracted from
the tweets are given in Table 1.
Proposed Algorithm
Step 1: Create Twitter API
Step 2: WOEID = 23424975 for India
Step 3: WOEID1 = 2459115 for New York
Step 4: Extract tweets using trends place(woeid) function for India and New
York
Step 5: Now from Extracted tweets for India
Search for tweets which contain any URLs
If URL present in tweet
Extract all the details mentioned in Table 1 from tweet
Step 6: Save it in a csv file.
Step 7: Also save the URL in another file with twitter id of that
Step 8: Function to identify URL in tweet text
def Find(string):
url = re.f indall( http[s]? : //(? : [a − zA − Z]|[0 − 9]|[$−
@.&+]|[!∗, ]|(? : %[0 − 9a − f A − F ][0 − 9a − f A − F ]))+ , string)
return url
The trending tweets were collected from Twitter because bots post tweets very
frequently (sometimes 700 times in a day) as compared to humans and therefore
those tweets become a trend. As per the tweets extracted, tweets from India
have more URLs as compared to that of New York.
3.2.1 Activity
Since bots are automated software, so the frequency of their tweets is much
more as compared to humans. There are some bots which are posting tweets
every minute. This is not human pattern of behaviour. Therefore by tracking
the number of tweets posted by account, bot like activity can be detected. For
example: Netflix Bot(@netflix bot). It will post a new tweet every minute about
some new show, gives updates about going on shows and posts URLs of the same
whereas a legitimate human account will post 2–3 tweets in a day.
3.2.2 Anonymity
Another factor is the degree of anonymity maintained by account. Usually the
less personal information the account gives, the more likely it is to be a bot. If it
will be a legitimate human account, then it will have personal information like
reading their education, jobs and profile pictures whereas bots won’t have such
information. For example: Museum Bot(@MuseumBot). The profile of this bot
doesn’t contain any profile photo.
3.2.3 Amplification
Bots tend to boost the signal from other users by retweeting, liking or quoting
them. Therefore, if timeline of an account consists of a procession of retweets
and word-for-word quotes of news headlines and very few or no original posts,
then there is a very strong possibility that it is a bot. For example: Dear Assis-
tant(@DearAssistant). The timeline of this bot will be filled with retweets and
replies to the questions asked. It contains very less original posts. Another exam-
ple: Museum Bot(@MuseumBot). Its timeline contains URLs and images from
museum always. It will not post anything personal.
Keeping above three points in mind and using the details extracted from
twitter a model is trained to identify whether the twitter account is a bot or not.
The machine learning algorithm used is Bag of Words. This algorithm counts
how many times a word appears in a document. Then these counts are used in
comparison of documents and find their similarities. It can be used in application
like searching, document classification and topic modelling [24]. This algorithm is
used to extract features from text and then those features are used for modelling
purpose. In this algorithm, occurrence of each word is used as a feature for
training a classifier.
Proposed algorithm
Step 1: Dataset about twitter bots is taken from Kaggle and split in 75:25
Step 2: Using 75% of data model is trained on given parameters:
1. Whether account is verified: if verified chances of being bot being mali-
cious is less
2. Tweets name, screen name, description and status is check for some spe-
cific words like
bot—bot—cannabis—tweet me—mishear—follow me—updates—every—
gorilla—yes ofc—forget etc. – these words are usually used in name,
description etc. of bots
190 T. Satija and N. Kar
3. Bag of words algorithm is used to check how many times these words
appear
If frequency is more than 50 ; possibly a bot
4. Then listed count is check. If greater than 16000; less chances of a bot
5. Check for number of retweets; if more than 10,000 and followers less than
200 ; chance of being a bot
Like this several other features are tested and data is saved.
Step 3: Then this trained model is tested on rest 20% data and comparison
is made with predicted data and original data to calculate accuracy.
Step 4: True positive rate (TPR) and false positive rate (FPR) calculated
Step 5: A ROC curve is plotted between true positive rate (TPR) and false
positive rate (FPR) with threshold value of 45◦ .
The data set for malicious bots and legitimate human accounts is taken from
Kaggle and divided in ratio of 75:25 ratios to get optimal result for training
and testing purposes. The accuracy of this model is: 95.44% as per the data set
used. A Receiver Operating Characteristic (ROC) curve is plotted between the
true positive rate (TPR) and false positive rate (FPR) with threshold value of
45◦ . ROC is a curve which is dependent on probability and tells how well model
can distinguish between two classes. Figure 5 shows the ROC plot for the model
trained. Area under curve is near to 1 which means model has a good measure
of separability Equations to calculate TPR and FPR:
is only one dependent variable that is, whether the twitter account is bot or not,
which is dependent on many other parameters like twitter account name, when it
was created, account’s status, number of followers etc. which are all independent
of each other.
Proposed Algorithm
Step 1: Dataset from Kaggle is taken
Step 2: From sklearn.linear model package import LogisticRegression
Step 3: Data from CSV file is split into 80:20 ratio for training and testing
Step 4: Model is trained using features from csv file.
Step 5: Then the trained model is tested on the 20% of data.
Step 6: Then score is calculated based on test result which tells the accuracy
of model.
The data is divided in ratios of 80:20 in order to get optimal result. If the
dataset is divided in 60:40 ratio, then accuracy was coming to be 80.05%. There-
fore, Eighty percent of the data is used for training the model and twenty percent
of data for testing the accuracy of the model. As per the dataset, the model is able
to achieve 93% (approx.) of accuracy. It is a static model just like WARNING-
BIRD [13]. It works on the predefined dataset only. Figure 6 shows the complete
flow of the proposed system and how all modules shall work together.
Figures 7 and 8 shows the details about Module 2 and Module 3.
The proposed system can help in identifying twitter bots which are posting
malicious URLs and report them to twitter. Also it can help in reducing APT
attacks, phishing attacks etc. as the entry point for all these attacks is usually
the malicious URL circulated either through social media or emails. Once if the
source of these malicious URLs are detected then these attacks can be prevented.
192 T. Satija and N. Kar
4 Conclusion
In present day scenario, twitter bots can be very dangerous if they are broad-
casting malicious URLs. They can very easily reach out to many people and can
lead to many data breaches and phishing scams. So in this paper a model is
proposed, which by using machine learning algorithms can successfully identify
twitter bots which are posting malicious URLs. The proposed model has been
tested and so far it is achieving good accuracy. It is able to distinguish between
twitter bots and legitimate human accounts. Also, the model is able to detect
Detecting Malicious Twitter Bots Using Machine Learning 193
malicious URLs. The proposed system can be used to prevent attacks where the
entry point is malicious URL. There is only one limitation. Proposed model is
trained to work with long URLs for now. But now a days these social networking
sites are using URL shortening facility which can reduce the URL to minimum
size possible. So in future, another module will be added to this system which
will be able to work on short URLS and detect whether they are malicious or
not.
References
1. https://en.wikipedia.org/wiki/Twitter
2. https://medium.com/the-startup-growth/how-many-people-are-on-twitter-
662d797d5958
3. Mønsted, B., Sapieżyński, P., Ferrara, E., Lehmann, S.: Evidence of complex con-
tagion of information in social media: an experiment using Twitter bots. PLoS
ONE 12(9), e0184148 (2017)
4. https://beebom.com/best-twitter-bots/
5. https://www.welivesecurity.com/2018/06/28/twitter-bots-disassemble
6. https://breachlevelindex.com/
7. Data from Twitter blog. https://blog.twitter.com/
8. Chang, S., Cohen, T., Ostdiek, B.: What is the machine learning? Phys. Rev. D
97(5), 056009 (2018)
9. Marx, V.: Machine learning, practically speaking. Nat. Meth. 16(6), 463 (2019)
10. Kan, H.J., Kharrazi, H., Chang, H.-Y., Bodycombe, D., Lemke, K., Weiner, J.P.:
Exploring the use of machine learning for risk adjustment: a comparison of standard
and penalized linear regression models in predicting health care costs in older
adults. PLoS ONE 14(3), e0213258 (2019)
11. Kuha, J., Mills, C.: On group comparisons with logistic regression models. Sociol.
Meth. Res. 47(1), 0049124117747306 (2018)
12. Kainkaryam, S., Ong, C., Sen, S., Sharma, A.: Crowdsourcing salt model building:
Kaggle-TGS salt identification challenge. In: 81st EAGE Conference and Exhibi-
tion 2019 (2019)
13. Lee, S., Kim, J.: WarningBird: detecting suspicious URLs in Twitter stream. NDSS
12, 1–13 (2012)
14. Chaudhary, M., Hingoliwala, H.A.: Warning Tweet: a detection system for suspi-
cious URLs in Twitter stream. Int. J. Res. Appl. Sci. Eng. Technol. (IJRASET) 2,
297–305 (2014). ISSN: 2321–9653
15. Alshboul, Y., Nepali, R., Wang, Y.: Detecting malicious short URLs on Twitter
(2015)
16. Sahoo, D., Liu, C., Hoi, S.C.H.: Malicious URL detection using machine learning:
a survey. arXiv preprint arXiv:1701.07179 (2017)
17. Chavoshi, N., Hamooni, H., Mueen, A.: DeBot: Twitter bot detection via warped
correlation. In: ICDM, pp. 817–822 (2016)
18. Novotny, J.: Twitter bot detection & categorization-a comparative study of
machine learning methods (2019)
19. Davis, C.A., Varol, O., Ferrara, E., Flammini, A., Menczer, F.: BotOrNot: a sys-
tem to evaluate social bots. In: Proceedings of the 25th International Conference
Companion on World Wide Web, pp. 273–274. International World Wide Web
Conferences Steering Committee (2016)
194 T. Satija and N. Kar
20. Chen, Z., Tanash, R.S., Stoll, R., Subramanian, D.: Hunting malicious bots on
Twitter: an unsupervised approach. In: Ciampaglia, G.L., Mashhadi, A., Yasseri,
T. (eds.) SocInfo 2017. LNCS, vol. 10540, pp. 501–510. Springer, Cham (2017).
https://doi.org/10.1007/978-3-319-67256-4 40
21. Devadoss, A.K.V., Thirulokachander, V.R., Devadoss, A.K.V.: Efficient daily news
platform generation using natural language processing. Int. J. Inf. Technol. 11(2),
295–311 (2019)
22. Chen, B.-C., Davis, L.S.: Deep representation learning for metadata verification.
In: 2019 IEEE Winter Applications of Computer Vision Workshops (WACVW),
pp. 73–82. IEEE (2019)
23. https://medium.com/dfrlab/botspot-twelve-ways-to-spot-a-bot-aedc7d9c110c
24. Zeng, J., Liu, M., Xiang, F., Ruiyu, G., Leng, L.: Curvature bag of words model
for shape recognition. IEEE Access 7, 57163–57171 (2019)
25. https://en.wikipedia.org/wiki/Receiver operating characteristic
Content Based Routing Algorithm
to Improve QoS in IoMT Networks
1 Introduction
1.1 Internet of Things (IoT)
Smart objects need to communicate themselves forming IoT networks. But as
the devices are low powered and low energy in these networks making the process
of communication challenging. The routing issues are very important to be con-
sidered [1]. In this paper we have proposed an algorithm to choose the reliable
path for the specific type of content such as text, audio, video, image and multi-
media. The reliable path is chosen by taking ACO principle as the basis. ACO is
based on the swarm intelligence principle. It is mostly used to find the solutions
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 195–206, 2020.
https://doi.org/10.1007/978-981-15-3666-3_17
196 S. Agnihotri and K. R. Ramkumar
for the optimization problems. The value of pheromone tells the goodness of the
routing path. The shortest path is chosen on the basis of pheromone using the
Forward Ants and the Backward Ant. Based on this concept, the artificial ants
are used in the computer solvable problems.
2 Literature Survey
Palaniappan and Chellan [4] tell about the energy efficiency maintenance in
MANETs. In these types of networks, there is always dynamic topology and
are self-constructive. The ad-hoc networks have to face new challenges such
as real time communication, resource constraints, bandwidth management and
packet broadcast overhead. Due to these issues, the researchers have always
found difficult to design efficient and robust routing algorithms. The paper also
mentions that the link strength is very important and should be considered.
The authors have used the cross-layer approach and considered the SINR for
the routing layer. The best metrics for reliability of a wireless link are SNR
and BER (Bit Error Ratio), but these cannot be used for every packet received.
Therefore, another metric, LPER (Link Packet Error Rate) is used for the QoS
in MANETS.
Cevik et al. [5] mention the growth of WSN to the WMSN, which are able to
transfer the multimedia content instead of only scalar data. During the transmis-
sion, only scalar data is not sufficient to be processed and transmitted such as
humidity, temperature etc. The paper also mentions the different WMSN archi-
tectures such as Single tier, Single-tier clustered and Multi-tier. The QoS needs
Content Based Routing Algorithm to Improve QoS in IoMT Networks 197
protocols needs to be done to improve the QoS and the network performance.
The metrics such as data rate, end-to-end delay play a vital role for reliable data
transmission in WMSNs. In multimedia data transmission, not 100% reliability
is required, still we would need maximum reliability we can achieve. This paper
proposes a QoS based routing algorithm which is based on ACO to enhance the
reliability in case of the WMSN networks. The route is found out on the basis
of SNR and the predictive analysis of the route is done before transmission of
data.
Where Packet Delivery Ratio is defined as the ratio of the number of packets
received by the receiver to the total number of packets sent by the sender.
The network performance is good if there is high PDR and vice versa. End-to-
End Delay is calculated as the time taken by the packet to reach from the source
Content Based Routing Algorithm to Improve QoS in IoMT Networks 199
to the destination. The delay during the communication can be caused due to
several reasons, may be through interference, queuing, packet conversions etc.
It is actually the summation of transmission delay, queuing delay, propagation
delay and processing delay. It is the time difference between the sending time
and the receiving time of the respective packet.
The network performance is good if the end-to end delay is low and vice versa.
Types of content during Transmission: The payload to be transferred from the
source to the destination can be of any form such as audio, video, image, text
or it may be multimedia. And, also the audio or video data may not be in the
same packet.
• Audio
• Text
• Video
• Images
• Multimedia
The paper [12] mentions the delay limits for the different types of data to be
communicated on the network. The delay which can be considered for different
data is:
Table 1. Acceptable delay limit for different type of content on wireless channel
The signal can be treated as the information which is required and of interest
value whereas the noise is treated as the error which effects the accuracy and loss
of precision can be there. From the source end when the message is transmitted
to the receiver, the message should be optimized and interpreted clearly as much
as possible. SNR not only affects the message to be communicated, it also affects
the throughput in the communication.
BER can be calculated from the given SNR. The formula to convert SNR to
BER is as follows:
BER = 1/2(SN R) (3)
The various factors are considered and the paper proposes algorithm CARA-
IoT, which is content aware. This algorithm takes ACO as the basic principle
and proposes the algorithms for Forward Ant-FANT, Backward Ant-BANT and
the packet delivery algorithm to deliver the payload on the channel.
Step 5: Visit the next neighbor stored in the RT of the current node
Step 6: If the next node is the source and update the pheromone value (consider
distance, hop count and conversion time and noise on the channel) at source RT
and GOTO step 5
Else, repeat step 2 and apply 70%-30% ACO rule for 70% of the current
pheromone value and the 30% of the previously calculated pheromone value
Step 7: Update the RT values
BANT ID Source Destination Pheromone value
Step 8: Stop pheromone calculation process
5 Simulation
The simulator has been developed in C# - IoTRSim. The CARA-IoT algorithm
(Content Awareness Routing Algorithm for IoT Networks) is the enhanced ver-
sion of ACO algorithm which finds the best reliable path not only based on
distance but the conversions made from one communication technology to other
communication technology and most importantly on the basis of delay limits for
the particular type of payload and the speed offered by the channel. The settings
for the IoTRSim are mentioned in Table 2.
Metric Setting
Simulation area 500 m × 500 m
No. of nodes 10–50
No. of ants 10–50
No. of iterations 10–20
Zigbee speed 75 mbps (full range is assumed to be 75 m) with
percentage drop of 25% for every 15 m
Bluetooth speed 20 mbps (full range is assumed to be 10 m) with
percentage drop 10% for every 5 m
Wi-fi speed 50 mbps (full range is assumed to be 50 m) with
percentage drop of 15% for every 10 m
Delay limit As mentioned in Table 1
Payload size Content Based-Text/Audio/Video/Image/Multimedia
Packet transmission size 512 KB
is found out considering the delay on every node and the speed. It may take more
time and as the congested nodes are not taken, the route is most reliable one.
The following graphs show the simulation results and compare the CARA-IoT,
ACO and SPEED algorithms.
The Fig. 1 shows the distance covered. The proposed algorithm CARA-IoT,
is covering longer distance in the routing path as compared to the ACO based
path and the SPEED. As the number of nodes in the simulation is getting higher,
so as the distance.
The Fig. 2 shows the comparison on the basis of time taken. The CARA-
IoT is taking less time as compared to ACO based approach and the SPEED
algorithm. The proposed algorithm proves to be more efficient as even if the
distance covered was high, still CARA-IoT has chosen the path which is taking
less time.
The packet delivery ratio (Fig. 3) has always been maximum in case of pro-
posed algorithm, CARA-IoT. Whereas in case of ACO approach, the PDR has
been varying. The CARA-IoT is having retransmissions to successfully deliver
the payload which makes the PDR as maximum as compared to ACO and
SPEED.
CARA-IoT considers the speed as well as the BER1 on the channel for cal-
culating the best possible path to deliver the different type of payload (that may
not be scalar data). Here, in the Fig. 4 shows that the average speed offered by
the CARA-IoT is higher as compared to ACO approach and SPEED algorithm.
204 S. Agnihotri and K. R. Ramkumar
6 Conclusion
One major problem of Internet is the lack of QoS, where there is always packet
loss, retransmissions and delays. So, reaching to the destinations becomes dif-
cult. The data packets may reach out of time, late, out of sequence or the packets
may not reach to the destination. To reassemble these packets again at the
destination is a major problem. In this paper, the routing path is calculated
based on the algorithm, which in turn finds the reliable, robust, efficient and
intelligent path which considers the type of content for the respective type of
Content Based Routing Algorithm to Improve QoS in IoMT Networks 205
data to be transferred. The results clearly show that there is less delay and
high packet delivery ratio, when the proposed algorithm is applied. There is
certainly improvement in the QoS. The multimedia networks such as WMSNs
and IoMT has several routing challenges. There is need for the development of
such intelligent and efficient routing protocols to match the speed of the growth
of IoT devices and IoT networks. The standard algorithms need to be modified.
The protocol stack and the architecture should be modified to accommodate the
IoT and IoMT networks. Further work can be done by the researchers to resolve
the routing issues mentioned in the paper.
References
1. Abazeed, M., Faisal, N., Zubair, S., Ali, A.: Routing protocols for wireless multi-
media sensor network: a survey. J. Sens. 2013, 1–11 (2013)
2. Kalaivani, P., Sathya, G., Senthilnathan, N.: QoS parameters estimation in
MANET using position based opportunistic routing protocol. Am. J. Comput.
Sci. Eng. Surv. 3(1), 21–27 (2015)
3. Zwyssig, E., Renals, S., Lincoln, M.: On the effect of SNR and superdirective beam-
forming in speaker diarisation in meetings. In: 2012 IEEE International Conference
on Acoustics, Speech and Signal Processing (ICASSP), pp. 4177–4180 (2012)
4. Palaniappan, S., Chellan, K.: Energy-efficient stable routing using QoS moni-
toringagents in MANET. EURASIP J. Wireless Commun. Netw. 1(3), 13 (2015)
5. Cevik, T., Gunagwera, A., Cevik, N.: A survey of multimedia streaming in wireless
sensor networks: progress, issues and design challenges. Int. J. Comput. Netw.
Commun. 7, 95–114 (2015)
6. Singh, R., Verma, A.K.: Efficient image transfer over WSN using cross layer archi-
tecture. Optik 130, 494 (2017)
7. Kandris, D., Tsagkaropoulos, M., Politis, I., Tzes, A., Kotsopoulos, S.: Energy
efficient and perceived QoS aware video routing over wireless multimedia sensor
networks. Ad Hoc Netw. 9(4), 591–607 (2011)
8. Hasan, M.Z., Al-Rizzo, H., Al-Turjman, F.: A survey on multipath routing pro-
tocols for QoS assurances in real-time wireless multimedia sensor networks. IEEE
Commun. Surv. Tutorials 19(3), 1424–1456 (2017)
9. Han, Y., Chen, Y., Wang, B., Liu, K.R.: Enabling heterogeneous connectivity in
Internet of Things: a time-reversal approach. IEEE Internet of Things J. 3(6),
1036–1047 (2016)
10. Almalkawi, I.T., Guerrero Zapata, M., Al-Karaki, J.N., Morillo-Pozo, J.: Wireless
multimedia sensor networks: current trends and future directions. Sensors 10(7),
6662–6717 (2010)
11. Usman, M., Yang, N., Jan, M.A., He, X., Xu, M., Lam, K.-M.: A joint framework
for QoS and QoE for video transmission over wireless multimedia sensor networks.
IEEE Trans. Mob. Comput. 8, 746–759 (2017)
12. Chen, Y., Farley, T.R., Ye, N.: QoS requirements of network applications on the
internet. Inf. Knowl. Syst. Manage. 4, 55–76 (2004)
13. Agnihotri, S., Ramkumar, K.R.: A survey and comparative analysis of the various
routing protocols of Internet of Things. Int. J. Pervasive Comput. Commun. 13(3),
264–281 (2017)
206 S. Agnihotri and K. R. Ramkumar
14. Agnihotri, S., Ramkumar, K.R.: An enhanced routing algorithm for Internet of
Things that resolves heterogeneity issues. In: Proceedings of the 6th International
Conference, BVICAM, INDIACom 2019, New Delhi, pp. 1274–1279. IEEE (2019)
15. He, T., John, A., Chenyang, L., Abdelzaher, T.: SPEED: A Real-Time Routing
Protocol for Sensor Networks. Virginia University Charlottesville Department of
Computer Science, Charlottesville (2002)
An Improved Chaos Based Medical Image
Encryption Using DNA
Encoding Techniques
1 Introduction
In the digital era exchange of data in the form of images, videos and text are
very prevalent. Medical Image is one of the most prized possession. Leak of
medical data can lead to misuse of insurance benefits, wrong treatments, and
criminal intent to high profile personalities. In the recent years, medical image
security has been the prime concern of the health care sectors primarily due to
partnership with insurance agencies who provide medical benefits to patients.
Many Encryption and Decryption techniques have been proposed for securing
medical images from unauthorized access by changing the medical images into
some other form. Medical Image security should be of primary concern, but it is
Fig. 1. Logistic map function to calculate the value of Xn for various values of r with
a sequence of length equal to image size [18].
Fig. 2. 8-bit Linear feedback right shift register with Odd parity checker
An Improved Chaos Based Medical Image Encryption Using DNA 211
K2,0 , K2,1 is produced by XORing (D6 ,D4 ,D2 and D0 ) subsequently and the
alternating pattern follows to increase randomness. K2,j = (Dj+6 ,Dj+4 ,Dj+2 and
Dj ) is general form of the evaluation for the Boolean function. After every right
shift the 0th position is affected.
Rule
Base 1 2 3 4 5 6 7 8
A 00 00 01 01 10 10 11 11
T 11 11 10 10 01 01 00 00
C 01 10 00 11 00 11 01 10
G 10 01 11 00 11 00 10 01
Image Conversion
The algorithm segregates any medical image according to the type it belongs
to, some images had their pixel values loaded as tuples and some had unsigned
integer values for each pixel. Each image can be considered of the size M × N
this image is converted into binary sequences of 8-bit per block image matrix
Imat , where M and N are 1,2,3....n. The values of the image matrix range from
0–255.
212 A. Das et al.
Generate Keys
The generated key sequences K1,i and K2,i are converted into binary sequences
and then XORed to create the Key Ki for encrypting the MRI image.
DNA Sequencing
Here a different approach is taken. The binary sequences that were generated in
Step “Generating Cipher Matrix” are converted into a sequence of DNA strands
using DNA Encoding mechanism. The rule and mechanism have been discussed
in Sect. 2.3, this sequence has been saved into a text file. Figure 4 shows the DNA
sequence that is generated using the DNA encoding rules.
The Decryption mechanism for our proposed algorithm is discussed. The receiver
just receives a text file. The received key has been sent through a secure channel.
In this algorithm, one can send pixel by pixel in an ordered sequence to securely
transmit the key at the receiver end. Figure 5 shows us the Decryption Model in
our proposed algorithm.
Generating Image Matrix. The Cipher Image Matrix and the Key which is
received through a secure channel is then XORed to produce the Image Matrix.
This Image Matrix is in binary form.
Image Iorg . The Idec and Iorg does not have any visual difference as well as they
are of the same size. Hence it can be said that the transmission is lossless.
The simulation code for the model is written in python. The anaconda navigator
is used and the code is implemented in Spyder IDE. The algorithm is compatible
with any image of the size of M × N there is no restriction on the type or
size of the image. The algorithm automatically recognizes the image structure
and implements the algorithm accordingly. The proposed scheme compares the
results of other published papers with the proposed algorithm. The Histogram
Analysis of the Medical Images, Root Square Mean Value, The PSNR Value,
Correlation Coefficient value Analysis, etc. are computed to prove the efficiency
of the algorithm. The proposed approach uses the initial seed value of X0 as a
random value within the range of 0.1–0.5. The bifurcation value r is set in the
range of 3.5 to 3.99 and the LSRSR initial seed value is set as [1,0,1,0,1,0,1,0].
An Improved Chaos Based Medical Image Encryption Using DNA 215
This section depicts what happens if an intruder uses some random key to deci-
pher the encrypted image. If the cipher is intercepted and the intruder say Trudy,
tries to decrypt the image using a similar algorithm as proposed with X0 as 0.2
and r = 3.99 and an L = [1,0,1,1,0,1,1,0] as initial seed and tries to generate the
key. The resultant will be as shown in Fig. 6(c). On the other hand when the
intended receiver who has received the key through a secure channel can decrypt
the cipher image to produce the image as shown in Fig. 6(d). The images Fig. 6
is of size 512 × 512.
Now a chest x-ray image of size 430 × 430 is taken Fig. 7(a) and analysis is
done and similar results are found as in case of Brain MRI Image.
It is observed that there is apparently no visual difference between the Orig-
inal Image and the Decrypted Image.
Fig. 6. (a) Original Image (b) Encrypted Cipher Image (c) Image Generated with
wrong Key by intruder (d) Decrypted Image generated by receiver using correct key Ki
Fig. 7. (a) Original Image (b) Encrypted Cipher Image (c) Image Generated with
wrong Key by intruder (d) Decrypted Image generated by receiver using correct key Ki
Decrypted Image has been plotted. The histogram for Fig. 7(a), (b) and (d) has
been plotted. It can be observed that the histogram plot for the Encrypted Image
is equalized than in case of original images. The Fig. 8(a) shows the chest Xray
image where the variation in the frequency of the pixels is very high, hence it
produces an irregular histogram formation. Whereas in Fig. 8(b) it is seen that
the cipher image is fairly equalized and the frequency of the pixels is uniform,
hence it can be said that the cipher has an even pixel density. In Fig. 8(c) it is
found that the decrypted image also has similar pixel frequency denoting that
the image is lossless. It is found that Fig. 6(a) produces similar results with a
high frequency of pixel 0 and with an even equalization of cipher image.
Fig. 8. (a) Histogram of original image (b) Histogram of encrypted cipher image (c)
Histogram of decrypted image
The Root Mean Square Error (RMSE) is the measure of how the pixel are spread
out in the image. This analysis considers the square root of difference between the
original image and the encrypted cipher image. This can be defined as follows:
1
MxN −1
RM SE = [F (i, j) − G(i, j)]2 (5)
M N i=0
Where F(i,j) is the original image and G(i,j) is the encrypted image, M × N is
the size of the image.
The higher the value of RMSE between the original image and the encrypted
image the higher is the randomness and the cipher image is more immune to
attacks. Now it is seen in Table 2 that the value for RMSE is much higher
for the proposed algorithm, compared to logistic map function only and other
standard algorithms.
It is observed that the RMSE values for both the Lena encrypted image
and the RMSE value for the Medical Image are higher than other proposed
algorithms. Please note that the medical image chosen in our case is different
than the one in [1]. Hence comparison values may differ in case of Medical Images.
An Improved Chaos Based Medical Image Encryption Using DNA 217
Here L is the total number of pixels of comparison, L = 512 × 512 and E(x) is
the mean value of x which is calculated as given in Eq. (11).
1 L
σx = (xi − E(x))2 (10)
L i=0
Equation (10) gives the step deviation at point x. Mean value of xi is given as,
L
1
E(x) = xi (11)
L i=0
Similarly, Step deviation at y and mean of y can also be evaluated. Table 5 shows
us the comparison between the Original image, Logistic Map Encrypted Image,
[1] and Proposed Algorithm.
4 Conclusion
The proposed algorithm produced better results compared to many other stan-
dard algorithms used for medical image encryption and decryption. With high
RMSE value and promising results from histogram analysis the proposed method
can be implemented in health care industries. DNA based cipher text reduces
the cipher size without any loss in data and the efficiency of the algorithm is
very high. The proposed algorithm will provide a secure key exchange of med-
ical images within the hospitals, from hospitals to patients, from hospitals to
authorized insurance agencies, etc. This will lay the foundation for a better
Medical Image Encipherment process in the health-care industry, especially in
India where the infrastructure significantly needs to improve and implement such
systems. Because in the digital era data is of utmost importance hence privacy
in the medical system is the need of the hour. Further tree-based technique can
be used in the future to hide patient-related details in the Medical Image itself.
Data hiding of the report in the Medical image will increase the confidential-
ity of the patient’s data and also rather than sending two different files over
the network the system can consolidate information in a single document and
communicate securely.
References
1. Rohith, S., Hari Bhat, K.N., Nandini Sharma, A.: Image encryption and decryption
using chaotic key sequence generated by sequence of logistic map and sequence of
states of Linear Feedback Shift Register. In: 2014 International Conference on
Advances in Electronics Computers and Communications, pp. 1–6. IEEE (2014)
2. Ravichandran, D., Praveenkumar, P., Rayappan, J.B.B., Amirtharajan, R.: DNA
chaos blend to secure medical privacy. IEEE Trans. Nanobiosci. 16(8), 850–858
(2017)
3. Belazi, A., Talha, M., Kharbech, S., Xiang, W.: Novel medical image encryption
scheme based on chaos and DNA encoding. IEEE Access 7, 36667–36681 (2019)
4. Shankar, K., Elhoseny, M., Chelvi, E.D., Lakshmanaprabu, S.K., Wu, W.: An
efficient optimal key based chaos function for medical image security. IEEE Access
6, 77145–77154 (2018)
5. Prabhavathi, K., Sathisha, C.P., Ravikumar, K.M.: Region of interest based selec-
tive medical image encryption using multi Chaotic system. In: 2017 International
Conference on Electrical, Electronics, Communication, Computer, and Optimiza-
tion Techniques (ICEECCOT), pp. 1–5. IEEE (2017)
6. Kamble, P.R., Patil, S.: Enhancing security of electronic patient record with tamper
detection. In: 2018 3rd International Conference for Convergence in Technology
(I2CT), pp. 1–5. IEEE (2018)
7. Das, P., Deb, S., Kar, N., Bhattacharya, B.: An improved DNA based dual cover
steganography. Procedia Comput. Sci. 46, 604–611 (2015)
8. Bharghavi, G., Kumar, P.S., Geetha, K., Devi, N.S.: An implementation of SLICE
algorithm to enforce security for medical images using DNA approach. In: 2018
International Conference on Communication and Signal Processing (ICCSP), pp.
0984–0988. IEEE (2018)
220 A. Das et al.
9. Abdel-Nabi, H., Al-Haj, A.: Efficient joint encryption and data hiding algorithm
for medical images security. In: 2017 8th International Conference on Information
and Communication Systems (ICICS), pp. 147–152. IEEE (2017)
10. Parameshachari, B.D., Panduranga, H.T., Naveenkumar, S.K.: Partial encryption
of medical images by dual DNA addition using DNA encoding. In: 2017 Inter-
national Conference on Recent Innovations in Signal Processing and Embedded
Systems (RISE), pp. 310–314. IEEE (2017)
11. Yan-Bin, Z., Qun, D.: A new digital chaotic sequence generator based on logis-
tic map. In: 2011 Second International Conference on Innovations in Bio-inspired
Computing and Applications, pp. 175–178. IEEE (2011)
12. Deb, S., Bhuyan, B., Gupta, NCh.: Design and analysis of LFSR-based stream
cipher. In: Mandal, J.K., Saha, G., Kandar, D., Maji, A.K. (eds.) Proceedings of
the International Conference on Computing and Communication Systems. LNNS,
vol. 24, pp. 631–639. Springer, Singapore (2018). https://doi.org/10.1007/978-981-
10-6890-4 61
13. Cisse, I.I., Kim, H., Ha, T.: A rule of seven in Watson-Crick base-pairing of mis-
matched sequences. Nat. Struct. Mol. Biol. 19(6), 623 (2012)
14. Yu, H., Wu, G.-S.: The compounded chaotic sequence research in image encryption
algorithm. In: 2009 WRI Global Congress on Intelligent Systems, vol. 3, pp. 252–
256. IEEE (2009)
15. Kar, N., Mandal, K., Bhattacharya, B.: Improved chaos-based video steganography
using DNA alphabets. ICT Express 4(1), 6–13 (2018)
16. Shannon, C.E.: A mathematical theory of communication. Bell Syst. Tech. J. 27(3),
379–423 (1948)
17. Center for Biomedical Image Computing & Analytics (CBICA). http://
braintumorsegmentation.org/
18. Incompressible Dynamics. http://www.kierandkelly.com/what-is-chaos/logistic-
map
Internet of Things (IoT)
Real Time Adaptive Street Lighting
System
Abstract. With the aim to optimize the cost and performance of hard-
ware and software integration, this paper discusses and inculcates the
outcomes of various experiments that were conducted with different hard-
ware and software technologies available commercially. This Smart Street
light is designed with an aim of both performance and cost optimization,
and includes a built-in weather station, forest fire/general fire notification
and has the ability to continuously send data to the server for Analytics
and Data Acquisition, as well as to the Mobile Application server. The
data through the Mobile Application server can be fetched directly onto
an app with good performance and the data being sent onto the Ana-
lytics and Data Acquisition server can be used for analysis and further
research. Also, this project is sensitive to day and night and optimizes
energy when there is no motion or when the ambient light is higher than
a particular threshold determined experimentally. This paper opens the
need for a better IoT based Street Light project with additional features
and improved performance.
1 Introduction
The growth of Internet of Things (IoT) has brought about a revolution in our day
to day lives. Cheaper and improved sensor technology has provided an impetus
to this revolution of IoT. Each and everything can be connected to the internet
and synchronized to interact, collaborate and contribute. IoT has been responsi-
ble for machine-human inter-dependency. The major advantages that IoT offers
include improved efficiency, reduced interaction time and human effort. It also
provides support to Artificial Intelligence and Data Mining due to the involve-
ment of larger data-sets and the need for extracting useful information in order
to improve efficiency to the application and add ‘Smartness’ to it. IoT has a
great potential for Automation and Data research. The thing that we are con-
sidering in this paper is a Street Light. The main motivation of this research
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 223–239, 2020.
https://doi.org/10.1007/978-981-15-3666-3_19
224 S. Garg et al.
Fig. 1. Basic features of a Smart Street Lighting system illustrating various modules
involved in the system
work is derived from the fact that lighting accounts for more than 15% of elec-
tricity consumption of the world. This also contributes about 1.6 billion tons of
carbon emissions (per year) in the atmosphere. With the aim to cut-short the
energy wastage through the conventional street lighting systems and realizing
the growing demand of Automation, it is important to work in this area. For
making the Street Light smart, we need to define smartness. A ‘Thing’ is said to
be ‘Smart’ if it is able to sense its immediate surroundings and perform actions
that work towards a desired goal. If we fit the Smart Street Light to this def-
inition, we can describe the system as composed to things called Street Lights
equipped with sensors, illustrated in Fig. 1, to sense ambient light, motion, tem-
perature, humidity, faults and moving objects, acquire data through the sensors,
apply some pre-processing algorithms to compress data, and send it to the Cloud
via an IoT gateway. From the Cloud which consists of a better Infrastructure
and storage and Analytical capabilities, users can retrieve beneficial informa-
tion including visibility in foggy days, ambient light, temperature and humidity
in sunny days, traffic density, etc. and researchers can fetch useful data and
analyze those data for further research that includes temperature and humidity
characteristics of a particular location, season, etc.
2 Literature Survey
Many researchers have proposed a number of methodologies and ideas to reduce
the energy wastage of the street lights and also highlighted the ways to make
Real Time Adaptive Street Lighting System 225
them smarter. Mullner et al. [9] has done an innovative literature review in the
field of saving energy with the help of electro-technology as well as Information
and Communication Technology (ICT) in Street Lights. Mahoor et al. [7] dis-
cussed about the state of smart street light which consists of monitoring and
executive programming. Lavric et al. [6] found that the information exchange
between lamp nodes and wireless sensor networks (WSN) can be a feasible since
it is composed of three parts: Control center, Remote concentrator and Street
lighting control terminals. Abhishek et al. [1] presented the idea of sending the
information of center station by ZigBee wireless communication. Gupta et al.
[4] proposed the solution for controlling the intensity of light with the help of
dimmer using PWM and Firing angles. With the use of dimmer we can conserve
15% to 20% of the energy. Escolar et al. [3] reviewed the street light approach
by using a simulator that combines wireless sensor network (WSN) and belief
desire intention (BDI) agents to provide information. It tried to exploit our app-
roach on more dynamic scenarios. Mohamed et al. [8] reviewed the state of smart
street light system based on the innovation technology named as Vehicular Ad-
Hoc Network (VANET). Smart Street Lighting system can be integrated with
VANET to reduce the cost and use the rich service and communication feature
of VANET. Pizzuti et al. [10] put forth an approach for adaptive street lighting
control based on energy demand idea and showed that the conventional street
lighting regulation system he tested showed that adaptive control provides on
an average almost double energy saving. Badgaiyan et al. [2] reviewed on using
ZigBee and wireless sensor network (WSN) – PIR (Proximity Infrared) sensor
and found that this was using low power and was highly efficient from con-
ventional street lighting system. A Smart Street Lighting (SSL) system [9] is
proposed for switching off the street lights based on the desired level of safety
and location of pedestrians. The proposed system consisted of Zigbee enabled
motes, a Base Station, a centralized SSL Server, server applications for system
configurations and Web-Application for monitoring and controlling the lighting
zones and lampposts and smartphones of pedestrians with GPS functionality.
A street lighting system based on ZigBee has also been proposed [7]. The
monitoring of the connected sensors for fault detection, power adjustment and
on/off control is done. The proposed system gathers the parameters of street
light via GPRS and Zigbee communication. It includes three main components:
a terminal for street monitoring, concentrator and monitoring center. The con-
centrator acts as a repeater and works as a bridge between monitoring center
and control terminals. Its main task is to forward command and data. The task
of the control terminal is to fetch the data and make the required adjustments
in the power.
An autonomous-distributed-controlled light system [11] was proposed in
which lights get turned on before a pedestrian actual arrives at that loca-
tion and then gets off after a pedestrian leaves a certain location. There are
access points which consist of the controller and the communication device. An
Automatic Intensity Varying Smart Street Lighting System [1] is presented in
which the initial step was to replace the existing lampposts by LED lights in
order to reduce the energy dissipation as LED lights are more energy efficient.
226 S. Garg et al.
Secondly, Light Dependent Resistor (LDR) has been deployed in order to control
the lights automatically without any human intervention. In order to control the
luminosity of the lights, dimmers are installed. A novel street lighting system [4]
is proposed which uses the renewable energy source i.e. solar energy. The lumi-
nosity of the street lights is controlled based on the intensity of sunlight which
is sensed by LDR. In case of low intensity of solar energy, the resistance value
of LDR is high and based on this value it is decided that whether the street
lights should be on/off. In case of moderate value of resistance, the dimming
function of lights is invoked. PIR sensor is also utilized to detect the motion
of the pedestrian. A novel fuzzy logic based controller [5] is presented whose
main objective is to minimize the wastage of power in street light systems. Solar
energy is utilized in order to save electric power. A fuzzy rule base is designed
which takes natural intensity and occurrence of vehicles as inputs based on the
rule set defined by fuzzy logic.
A hybrid street lighting system based on Vehicular Ad-Hoc Networks
(VANET) [8] and traditional street lighting system is proposed. In VANETs
Road-Side Units (RSUs) are deployed usually 300–400 m apart. The main task
of these RSUs is to report the data to the light controllers to switch on and
off a light. A reliable on-demand street lighting model [10] is proposed in which
the concept of traffic flow rate forecasting model is utilized. There are number
of methods based on 1-h prediction whose results are compared with the model
based on ANN which reports the best results. As per the above discussed litera-
ture, it is clearly concluded that most of the ideas are based on motion detection
of a person or a vehicle on a road.
Fig. 2. Architecture of the proposed solution depicting various units in the Smart
Street Light system
Real Time Adaptive Street Lighting System 227
3 Proposed Methodology
Based on the previous works and current problems we defined the objectives of
our work. The broad objectives of this Smart Street Light Project are as follows:
– Design and development of low cost and energy efficient Smart Lighting Sys-
tem.
– Deployment of multiple sensors for motion detection, ambient light detection,
optical fog sensing, fault detection, fire detection etc.
– To control the intensity of light of deployed LED lights and their power
consumption based on weather conditions, diurnal timings, traffic levels and
motion detection.
– To record essential data and apply analytics to extract useful information
which further sends a feedback to the main network improving its performance
and adding “Smartness”.
Fig. 3. Comparison between open and closed system with GD as the transfer function
and C and r as output and input respectively. H is the feedback element and E represent
the error signal
228 S. Garg et al.
lighting system works using the concept of feedback in order to minimize error.
Illustrated in Fig. 3, Let GD denotes the transfer function of the measurement
system. Let C and r denote the output and input signal respectively. Also, let H
denote the feedback element in a closed loop system and e be the error or the
interference introduces in the system. Consider an open loop system as shown
in Fig. 3. We know that
C = GD .r (1)
Let the transfer function change from GD to GD + ΔGD which changes C to
C + ΔC therefore, in an open loop:
ΔC = ΔGD .r (3)
Now, consider the closed loop system with H as the feedback element and error
introduced is e,
C = GD .e (4)
e = r − CH (5)
C = GD .(r − CH) (6)
GD .r
C= (7)
(1 + GD H)
Let the transfer function change from GD to GD + ΔGD which changes C to
C + ΔC therefore, in a closed loop:
ΔGD .r
ΔC = (9)
1 + (GD + ΔGD ).H
If GD >> ΔGD so,
ΔGD .r
ΔC = (10)
1 + GD .H
It is clear that ΔC in closed loop is very less as compared to that in an open loop.
This method of high gain feedback therefore reduces error in a measurement sys-
tem with the help of feedback. The sensors are positioned in such a way so as to
reduce error from internal temperature and heat generation, hindrance in path of
light from ambience to the LDR, etc. All the field sensors are capable of working
at a voltage of 3.3 V DC or 5 V DC which is supplied from the voltage regulators
connected to the main DC power supply unit which is charged through Solar Panel
of rated 24 V output voltage as well as through the mains with 220 V 50 Hz stan-
dard power supply. The AC Dimmer is a device that is designed to control the
AC voltage, which can transfer current up to 220 V (5 A 10 A) (TRIAC BTA16
Real Time Adaptive Street Lighting System 229
for 600 V/16 A). Discharge lamps do not support dimming. Power part of dimmer
is isolated from the control part, to exclude the possibility of high current disrup-
tion to a microcontroller. The logical level is tolerant to 5 V and 3.3 V, therefore
it can be connected to the microcontroller with 5 V and 3.3 V level logic. Dim-
ming can be achieved by Pulse Skip Modulation: Technique 1: One or more cycles
(sine wave signal) are transferred to the load while following one or several cycles
are blocked. Technique 2: Partial transference of each sine wave to the load. Tech-
nique 3: Generation of modulated full sine signal of different frequency up to few
hundred Hertz. This method requires specialized powerful AC generators with dif-
ferent modulation. Methods 1 and 2 are the easiest to execute with the help of a
Dimmer and program code: in both cases, there is a need of circuit that detects
the zero crossing and can control a TRIAC. Figure 4 shows the circuit diagram of
a basic dimmer circuit where the symbols have their usual meaning. The specifi-
cations of the dimmer used in the project are presented in Table 1. Alarm gener-
ation happens when the flame sensor that is based on YG1006 sensor which is a
high speed and high sensitivity NPN silicon photo-transistor detects a fire source
or other light source of wavelength in the range of 760 nm–1100 nm. Due to the
black epoxy coated on the diode, it is sensitive to Infrared Radiation. When the
sensor detects flame, the signal LED will light up and the digital pin goes low as
designed by the manufacturer using an internal Data Acquisition circuit. Table 2
describes the specifications of the flame sensor used in the street lights. Generic
procedure for Fault Detection:
Table 1. Table depicting the specifications of the dimmer used in the project
Parameter Details
Power Up to 220 V (5 A 10 A)
AC frequency 50/60 Hz
TRIAC BTA16—600B
Isolation Optocoupler
Logic level 3.3 V/5 V
Zero point Logic level
Modulation (DIM/PSM) Logic level ON/OFF TRIAC
Signal current Greater than 10 mA
Environment For indoor and outdoor use Operating temperatures:
20 ◦ C to 80 ◦ C Operating humidity: Dry environment
only
ROHS3 Compliant
230 S. Garg et al.
Fig. 4. Circuit Diagram of a general purpose dimmer used to vary the intensity of AC
Street Light
Table 2. Table depicting the features of the Flame sensor used in the Smart Street
lights
Parameter Details
Working voltage 3.3 V–5 V
Detect range 60◦
Output type Digital/analog
On-board IC LM393 chip
Dimension 3.2 cm × 1.4 cm
– Calculation of Mean of the temperature and Humidity values from the street
lights in the small local area network and then send it to the thingspeak server
and Blynk server.
– Motion/fire/fault flags are interrupt driven and it is generated by each indi-
vidual street light with its Unique ID (UID). The notification, if generated is
sent to the Blynk server with the UID and a mail is generated for fire/fault
which is sent to the Admin with the UID of street light where the interrupt
was generated.
The next important thing is the IoT gateway which links the local microcon-
troller network with the Cloud. The protocols used in the device to device
communication is IEEE802.15.4 (Zigbee) for slave street lights and IEEE802.11
(Wi-Fi) for master street lights as shown in Table 3. REST APIs are used to
read/write data onto the Cloud server with the help of a unique authorization
key which is valid for a particular channel. The data is sent to the server every 5 s
from the gateway using a client and fetched information from the Cloud server
if required if some logic permits. This timing is set to match the flow between
sender and receiver and to reduce constant polling which affects the performance
of the communication and accounts to the lag in actuation. The microcontroller
used is ATMEGA328p of the Atmel series of microcontrollers and it has been
interfaced with ESP8266 WiFi module to add WiFi connectivity to data trans-
fer. The ATMEGA328p is a 8-bit high performance microcontroller which works
optimally with cost and energy effectiveness as shown in Table 4. Zigbee is a
232 S. Garg et al.
good alternative for Street Light to Street Light communication as it has a low
cost, low power utilization, and appreciable performance.
Master Street Light is the main street light with some important tasks. Tasks
of Master Street Light include:
Parameter Value
Microcontroller ATmega328p
Operating voltage 5V
Digital I/O Pins 14 (of which 6 provide PWM output)
Analog Input Pins 6
DC current per I/O Pins 40 mA
DC current for 3.3 V Pin 50 mA
Flash memory 32 KB
SRAM 2 KB
EEPROM 1 KB
Clock speed 16 MHz
ICSP header In System Programming, used when
you want to bootload, you’ll need an
AVR-ISP
The data received from the sensors via an IoT gateway is sent to the cloud
after pre-processing and compression to reduce un-necessary overheads. We have
used two different clouds in order to reduce burden. One of the clouds is respon-
sible for gathering data from the sensors and recording their data so that they
can be further analyzed using Machine Learning and Analytics. We can fetch
the spreadsheet of the data recorded with its timestamp and location parameters
from the Cloud directly. We have installed MATLAB onto the cloud for analysis
which is a great tool for analysis of data through different plotting techniques
Real Time Adaptive Street Lighting System 233
available in MATLAB. Referring Fig. 8, The other cloud server is responsible for
handling requests from the users. The application which is running on a mobile
continuously show real-time status of the sensors and the data that is being sent
by them including ambient light, temperature and humidity, as illustrated in
Fig. 5, and generates an alarm if there is any fault or fire. There are libraries
available which enable the required information to be fetched or sent from mobile
application server to Cloud server with the help of APIs.
Wireless Sensor Networks have limited processing capability and limited battery
power. Due to large collection of input data, it is difficult to manage the data
along with different domains. Hence energy efficient data aggregation technique
is required for efficient data collection. Data aggregation is the method in which
data coming from different sensors is combined and provides useful aggregated
information. Keeping in view the above issue, a novel energy efficient fuzzy logic
based data aggregation technique is proposed. The proposed technique collects,
analyzes, classifies and aggregates the data of different domains automatically
which is reported by various sensors. Further, fuzzy logic technique is applied
as it has capability to deal with dynamic situations and to model the condi-
tions which are inherently imprecisely defined. Proposed data aggregation tech-
nique aggregates the incoming data in an effective manner by reducing energy
Fig. 5. Flowchart depicting the flow of temperature and humidity data from sensor to
the cloud
234 S. Garg et al.
5.1 Results
For Automation of Street Light based on the ambient light intensity, we found
that climatic conditions play a major role. Also, the latitude and the location
Fig. 6. Flowchart depicting the logic involved in fault detection of Smart Street Light
and flow of control
Fig. 8. Glimpse of the user interface of MATLAB’s thingspeak platform showing plots
of data collected
Table 5. Table depicting the variation of voltage with changes in the intensity of light
Luminous flux [lm] Lamp wattage [W] Lamp voltage (AC) [V]
5000 50 220
3600 44 200
2520 37 180
1714 30 160
1114 24 140
Parameter Value
2
Network coverage/m 200 × 200 m2
Number of sensors 50 300
Initial energy/J 0.5
Data packet size/Bit 500
Control packet size/Bit 12
A 2.5
Maximum transmission range/m 30
Fig. 9. Bar plot showing the the variation of voltage with changes in the intensity of
light and lamp wattage
236 S. Garg et al.
decide the angle of incidence of sunlight therefore the installation of Light Depen-
dent Resistors should be done accordingly. Only locations lying along one line of
latitude on the surface of the Earth can receive sunlight at a 90◦ angle on a given
day. All other places receive sunlight at lesser intensities. Regardless of latitude
or time of year, the sun’s angle reaches closest to 90◦ and is therefore at its most
intense at the midpoint of the day: noon. Therefore, we have to create a logic
based on this variation so we decided to fetch the time of operation and location
of a street light in addition to the actual LDR readings sent by the Street Light.
Based on the actual calibrated LDR data, we adjusted the lamp wattage and
lamp voltage with the help of a dimming module and a instruction set executed
by a microcontroller. The fault of the street light can be detected if the LDR at
the street light shows a low reading when it is supposed to be ON. The real time
clock attached to the server lets the microcontroller know when the lamp needs
to be in ON state and to detect whether it is ON or Off can be found out by the
LDR attached to it and a trigger is generated of failure or fault as illustrated
in Fig. 6. With reference to the works of Mahoor et al. [7], we have conducted
experiments in order to obtain data with the aim of building A smart street
lighting control system for optimization of energy consumption and lamp life.
The Table 5 highlights the variation of voltage, wattage and intensity of light
during the day thereby reducing power loss and increasing efficiency. Figure 9 is
a pictorial representation of the data in Table 5 and it was observed that The
electrical cost was reduced by 20% and the Lamp Life was enhanced by 100%
for a time span of 12 h. With the observations made it was recorded that the
flow of vehicles as well as human traffic is at its apex during 6PM to 10PM time
frame so, our operational strategy of street light was such that the intensity and
luminous flux will be at its peak during these hours. Past, 10PM with the usage
Real Time Adaptive Street Lighting System 237
of AC Dimmer module the voltage takes downward leap to 160 V up till 12AM
after that the light operates on 140V till it get completely dimmed out with the
dawn. The path with minimum () is decided by fuzzy member functions and data
is transmitted. Referring Fig. 10, The evaluation of FLDA is done through MAT-
LAB and FuzzyJ Toolkit used for Java to implement the necessary fuzzy logic
functionality. Figure 7 is a block diagram depicting the process whereas Table 6
describes the simulation parameters used in this experiment work. i.e. denotes
the energy dissipated in the transmission of message from source to destination.
Based on the experiments the flowchart for control flow can be summarized in
figures with impetus to motion detection, fire detection as in Fig. 11, detection of
ambient light, detection of fault and temperature and humidity detection. With
these experiments and the logic proposed, we have successfully collaborated all
the essential features of a Smart Street Light and provided a generalized model
that would help in future research endeavours.
Fig. 11. Flowchart expressing the procedure involved in detection of fire and various
steps involved in it
Pi or a NodeMCU board that can connect to the network using Wi-Fi which is
already available in smart cities these days. We will then install a Master Street
Light that can send the data to the server which cuts down the cost of gateways
and also decreases power consumption by these devices. Our novelty lies in the
fact that we make out street lights dim when there is no pedestrian or traffic
around thereby cutting down energy consumption. As depicted in Fig. 9, the
luminous intensity of the lamp is proportional to the energy consumption, we
are enabling the lights of the street light at dusk and disabling them at dawn
by defining a threshold as mentioned earlier and when the are enabled we are
further cutting down energy consuming by not running them at full intensity
by running them in a dim mode. We some traffic is there they run at their full
intensity thereby not hindering with the normal working.
7 Conclusion
From the discussion done in this paper, the characteristics of a smart street light
have been defined in detail. A Smart Street should be energy efficient, able to
sense the environment and act in an appropriate manner based on a certain
set of instructions. It should also have the ability to send data efficiently with
cost effectiveness to a standard server which is available all the time. Motion
detection, ambient light and fire detection adds to its integral features and are
essential to be smart in present scenarios. With all the methodology operated
and results achieved a significant optimized model of smart street lamp light is
presented and it was observed that model provides a vital energy saving model
and corresponding to the results achieved, the model provides sufficient energy
saving. The weather station further will integrate the data of forecast with the
cloud storage and accordingly will give the inputs to smart street light.
References
1. Abhishek, K., Srikanth, K.: Design of smart street lighting system. Int. J. Adv.
Eng. 1(1), 23–27 (2015)
2. Badgaiyan, C., Sehgal, P.: Smart street lighting system. Int. J. Sci. Res. 4(7),
271–274 (2015)
3. Escolar, S., Carretero, J., Marinescu, M.C., Chessa, S.: Estimating energy savings
in smart street lighting by using an adaptive control system. Int. J. Distrib. Sens.
Netw. 10(5), 971587 (2014)
4. Gupta, A., Gupta, S.: Design of automatic intensity varying smart street lighting
system. In: IOP Conference Series: Materials Science and Engineering. vol. 225, p.
012126. IOP Publishing (2017)
5. Jinila, Y.B.: Solar powered intelligent street lighting system based 1 on fuzzy logic
controller. Int. Rev. Electr. Eng. 10(3), 399–403 (2016)
6. Lavric, A., Popa, V., Sfichi, S.: Street lighting control system based on large-
scale WSN: a step towards a smart city. In: 2014 International Conference and
Exposition on Electrical and Power Engineering (EPE), pp. 673–676. IEEE (2014)
Real Time Adaptive Street Lighting System 239
7. Mahoor, M., Najafaabadi, T.A., Salmasi, F.R.: A smart street lighting control
system for optimization of energy consumption and lamp life. In: 2014 22nd Iranian
Conference on Electrical Engineering (ICEE), pp. 1290–1294. IEEE (2014)
8. Mohamed, S.A.E.: Smart street lighting control and monitoring system for elec-
trical power saving by using vanet. Int. J. Commun. Netw. Syst. Sci. 6(08), 351
(2013)
9. Müllner, R., Riener, A.: An energy efficient pedestrian aware smart street lighting
system. Int. J. Pervasive Comput. Commun. 7(2), 147–161 (2011)
10. Pizzuti, S., Annunziato, M., Moretti, F.: Smart street lighting management. Energ.
Effi. 6(3), 607–616 (2013)
11. Popa, M., Cepişcă, C.: Energy consumption saving solutions based on intelligent
street lighting control system. UPB Sci. Bull. Ser. C 73(4), 297–308 (2011)
A Review on Coverage and Connectivity
Through Random Deployment of Nodes
in Wireless Sensor Network
1 Introduction
WSN comprises of considerable size of sensors called motes which form an adhoc
network once they are deployed in the region of interest (ROI). They are used
for detecting environmental changes such as pressure, humidity, temperature,
level of pollution, etc. All nodes have a connection to Base Station (BS) through
wireless links and send the sensed data periodically to the BS. The BS is respon-
sible for aggregating and processing the data and the processed data is send to
the web server through internet [1].
The sensor node can be installed by simply dropping the node from
Unmanned Aerial Vehicle (UAV) in case of accessible region or simply by placing
the node in the field of interest. The network can be extended through addition
of more nodes in a similar way without any extra endeavor. Installation costs
are also lowered since WSN have the capacity to respond to change in topology.
WSN can be heterogeneous also as some nodes with higher energy are deployed
to extend the lifespan of the network (Fig. 1).
Fig. 1. A WSN where sensors detect application specific events and send sensing data
to BS (https://www.researchgate.net/figure/Wireless-Sensor-Network-Architecture-7
fig1 281459697).
The basic unit of WSN [2] is a sensor node which comprises of the following:
(1) sensing transducer which measures various environmental changes within its
vicinity but it has a limited sensing range. (2) Data Processor transforms the
measured or sensed data into an analog signal. (3) Radio Transceiver is used to
transmit the analog signal to the BS through a single hop or multi hop wireless
communication links. (4) Embedded batteries are used as a source of power for
sensing, to process the sensed data and for efficient communication. But the only
disadvantage is that these batteries are non – rechargeable. So extensive research
is going on to facilitate all these tasks using solar powered batteries (Fig. 2).
2 Application of WSN
4 Connectivity Model
The communication model of WSN can be expressed in terms of a graph compris-
ing of vertices and edges, denoted by G = (V, E) sensor node represent vertices
and wireless communication links are denoted by edges. Two sensor nodes will
be neighbor nodes if Euclidean distance is less than or equal to communication
range. Figure 5 illustrates that sensors A and B are neighbors since they are
inside the communication range of each sensor, but node C is isolated.
A WSN remains connected if there exist a single hop or multi hop path
between the sensor nodes. Connectivity in WSN depends upon the location where
the nodes are deployed and their communication range.
5 Coverage Model
Environmental changes are monitored by the sensors deployed within ROI. A
point ‘a’ is covered by the sensor if Euclidean distance between point and the
sensor si is less than or equal to the sensing range rs of the sensor [9].
d(a, si ) ≤ rs (2)
The region of interest is said to be fully covered if there exists at least one
sensor and there is no hole in that ROI. Disk sensing model doesn’t provide
full sensing capability of real time sensors because they are small and cheap
and are unable to provide same sensing capability in each and every direction.
Therefore, a probabilistic model called Elfes sensing model is utilized to model
sensing capability of realistic sensors. Using Elfes sensing model [10], a sensor
can detect an event at a distance within (rs – e) with probability 1, and at a
distance greater than (rs + e) is with probability 0, and within the distance
interval (rs – e, rs + e) with probability e−a(r−rs ) , where rs is average sensing
range and e measure uncertainty in sensing of sensors detection capability.
Fig. 6. Voronoi polygon of a sensor in a square, triangle and hexagonal lattice (https://
link.springer.com/chapter/10.1007/978-81-322-2129-6 2).
246 R. Chowdhuri and M. K. Deb Barma
Voronoi polygon determines the number of nodes required to wrap the full
area. Assume Area is the application field to be covered and Av is the Voronoi
Polygon Area, then number of sensors Ns required to complete the entire region
is given by:
Area
Ns = (3)
Av
The area of the Voronoi Polygon is computed based on the length of the
deployment lattice pattern, i.e., deployment distance lx . This parameter in turn
depends on sensing range and communication range of sensor.
From Fig. 8, it is absolutely clear that all sensors are connected and there is
no isolated sensor. If the degree of a node is 0, then it is called as isolated sensor.
Number of neighbors that a node possesses within its communication range is
defined as degree of a node.
248 R. Chowdhuri and M. K. Deb Barma
Let us consider a sensor node si arbitrarily and the probability of the node
being isolated is almost equivalent to the probability that the node has no neigh-
bors within its communication range and is as follow [1]:
2
Pisolated (si ) = e−pπrc (10)
Since N number of sensor nodes are independently deployed over the appli-
cation area, all sensor node possesses equal probability of being non isolated.
Therefore, probability that there doesn’t prevail any isolated sensor in the net-
work is given as [1]:
N
2
Pnonisolated = Pnonisolated (si ) = (1 − e−pπrc )N (12)
i=1
9 Performance Metrics
9.1 K – Coverage
It refers to minimum K – Coverage. A WSN is presumed to have K – Cover-
age, if every point in the network is covered by minimum k number of sensors.
A Review on Coverage and Connectivity Through Random Deployment 249
Boundary conditions for any type of deployment strategies are not considered
and hence suitable for large scale WSN applications.
10 Conclusion
In this paper, three important aspects of WSN have been investigated in detail:
Node deployment models which are of two types: deterministic and random.
Connectivity model is discussed which assures whether the network is fully con-
nected or not and whether reliable data communication is taking place or not
between sensor nodes and cluster head and also from cluster head to the sink
node or base station for further processing of data. Coverage model is also dis-
cussed to find whether each and every point in the application area is covered
or not without leaving any holes in the region of interest. Coverage and Connec-
tivity are the most essential requirements for maintaining Quality of service in
a network.
References
1. Wang, Y., Zhang, Y., Liu, J., Bhandari, R.: Coverage, connectivity, and deployment
in wireless sensor networks. In: Patnaik, S., Li, X., Yang, Y.-M. (eds.) Recent
Development in Wireless Sensor and Ad-hoc Networks. SCT, pp. 25–44. Springer,
New Delhi (2015). https://doi.org/10.1007/978-81-322-2129-6 2
2. Al-Karaki, J.N., Kamal, A.E.: Routing techniques in wireless sensor networks: a
survey. IEEE Wirel. Commun. 11(6), 6–28 (2004)
250 R. Chowdhuri and M. K. Deb Barma
3. Devi, G., Bal, R.S., Nayak, S.M.: Node deployment and coverage in wireless sensor
network. Int. J. Innov. Res. Adv. Eng. (IJIRAE) 1(2), 139–145 (2015)
4. Wang, Y., Wang, X., Agrawal, D.P., Minai, A.A.: Impact of heterogeneity on cov-
erage and broadcast reachability in wireless sensor networks. In: The Fifteenth
International Conference on Computer Communications and Networks (ICCCN)
(2006)
5. Wang, X., Xing, G., Zhang, Y., Lu, C., Pless, R., Gill, C.: Integrated coverage
and connectivity configuration in wireless sensor networks. In: ACM SenSys 2003
(2003)
6. Zhang, H., Hou, J.C.: Is deterministic deployment worse than random deployment
for wireless sensor networks? In: INFOCOM (2006)
7. Wang, Y., Kutta, A.: Joint and simultaneous K-sensing detection in determinis-
tic and random sensor networks. In: Proceedings of the 26th IEEE International
Parallel and Distributed Processing Symposium Workshop on Dependable Paral-
lel, Distributed and Network-Centric Systems, Shanghai, China, 21–25 May 2012
(2012)
8. Fortune, S.: Voronoi diagrams and Delaunay triangulations. Comput. Euclidean
Geom. 1, 193–233 (1992)
9. Liu, B., Towsley, D.: A study of the coverage of large-scale sensor networks. In:
IEEE International Conference on Mobile Ad-Hoc and Sensor Systems, pp. 475–
483. IEEE (2004)
10. Elfes, A.: Occupancy grids: a stochastic spatial representation for active robot
perception. In: Iyenger, S.S., Elfes, A. (eds.) Autonomous Mobile Robots: Percep-
tion, Mapping and Navigation, vol. 1, pp. 60–70. IEEE Computer Society Press,
Washington, D.C. (1991)
11. Liu, B., Towsley, D.: A study of the coverage of large-scale sensor networks (2004)
12. Wu, H., Zhu, L.: Coverage strategy for heterogeneous nodes in wireless sensor
network based on temporal variability of farmland. https://doi.org/10.17559/TV-
20160517052746. ISSN 1330–3651 (Print), ISSN 1848–6339 (Online)
Role of Open Hardware and IoT in Home
Automation: A Case Study
1 Introduction
Open Source hardware and Internet of Things are the technologies of today’s
world which is penetrating rapidly into our everyday lives with various smart
and innovative solutions which not only simplify our living but also allow us to
save power and make our life style energy efficient and smart. When things get
connected, we can feel more in our control and thus enhancing and empowering
our normal day to day living. IoT is having a great impact on improvising home
automation solutions and designing home automation with having open source
hardware and sensors as the pillars to implement where everything smart solution
will be taken into account by the devices interconnected on IoT. With these all
the personal devices connected to internet will enable us to author our lives.
For example, in medical science these technologies have played a major role in
enabling diagnostic devices and systems to sense upcoming ailments and act
accordingly on it.
Supported by NIT Agartala.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 251–262, 2020.
https://doi.org/10.1007/978-981-15-3666-3_21
252 A. De et al.
The objective of this project is to make life easier through automation and
rooted on few major advances over the technology accessible in automation, such
as advancement in wireless automation technology as well as dropping of price
points as the market gets matured enough to accept home automation require-
ments in substantial amounts. And we have taken a step forward in remotely
observing the temperature and humidity of a room and actuating an actuator
based on the requirement to smartly control various equipment based on the
level of temperature and humidity.
According to the McKinsey report “Disruptive technologies: Advances that
will transform life, business, and the global economy”, the Internet of things
(IoT) is one of the top three technological advancements of the next decade
(together with the mobile internet and the automation of knowledge work). The
report goes on to say that “The Internet of Things is such a sweeping concept
that it is a challenge to even imagine all the possible ways in which it will affect
business, economies, and society” [1,11].
IoT comprises of all the web enabled devices which gather data, send data
and act on the data which they get from the surrounding with the help of various
sensors with embedded software, microprocessors and communication hardware.
These devices are smart and connected devices and can also communicate with
other devices a technique known as machine to machine communication. We can
also interact with the devices to setup them with instructions. Although the
devices do much of their job on their own without much human intervention.
Thus, we can say IoT (Internet of Things) is mostly about connecting various
daily life devices to internet and interact with them to gather information and
also control them when required remotely. One example is our room air condi-
tioner which we can suppose control based on the ambient room temperature
and humidity using sensors using our mobile phone or tablet from anywhere on
the planet.
Open Source Hardware: Open Hardware or Open Source Hardware are the
physical entities or objects whose design features and specifications are licensed
in a manner that the said entity can be studied upon, can be modified, recreated
and also distributed if required by anybody. The term Open Hardware does not
only specify for computers, it can be any type of entity as it is nothing but a
collection of design principles and legal practices which can be defined for any
entity or object to be termed as Open Hardware or Open Source Hardware. It
can thus specify objects such as Computers, Vehicles, Furniture, Buildings, and
so on [12].
This technology is best suited for implementing Home Automation as it gives
the engineer the liberty to choose from a variety of available hardware platforms
to upload their codes and interface sensors and if required the freedom to update,
build and distribute the technology.
Role of Open Hardware and IoT in Home Automation: A Case Study 253
2 Related Work
There has been a lot of work done on Home Automation using Open Source
platform and Internet of Things and is still going on at full pace. As our world
is changing very fast in terms of technological advancement, lifestyle of human
beings and also climate, the need of automation and smart appliances are becom-
ing increasingly high as all these are required to complement our changing envi-
ronment in a smarter and more sustainable way which in turn will bring efficiency
in power consumption, reduction in time required to do a particular task, cost
cutting and above all the power to connect devices and appliances remotely from
anywhere in the world.
Few of the noted applications in this field are discussed here. Sipani et al. [6]
demonstrated temperature and humidity monitoring and control using Arduino
Open source board. Patnaikuni [4] presented a comparative study of various
IoT development boards such as Arduino, Raspberry Pi and ESP8266. Kunal
et al. [2] developed an end to end smart home automation system using Internet
of Things. Malathi et al. [5] developed a home automation system with the help
of ESP8266. Kale and Kulkarni [7] demonstrated a real time temperature and
humidity monitoring system using Arduino and Xbee S2. Saiteja et al. [3] devel-
oped a home automation system using Raspberry Pi board which can control
appliances remotely and also provide security for any mishap when the user is
254 A. De et al.
not present at home. Liu and Wang [8] elaborately presented the future scope
and research prospects of Internet of Things and its applications (Fig. 1).
Figure 2 represents the way the IoT mobile application works to enable Internet
of Things. It is capable of controlling the Open Hardware devices remotely and
display, store and visualize sensor data. The three components of the system
works as follows.
Fig. 2. Communication between IoT mobile application and Open Hardware platform.
256 A. De et al.
IoT Mobile App: It helps us to create user friendly interfaces for various IoT
based projects using variety of widgets.
IoT Cloud Server: It controls all necessary communications between the
mobile device and open hardware.
IoT Libraries: It stores all types of required files compatible with different
Open Hardware platforms (like ESP8266 NodeMCU, Arduino Uno, etc.) and
also controls and process all the to and fro commands from the server.
The mobile application is connected to the Open Hardware platforms via the IoT
cloud server which maintains the communication between them with the help
of various library files required as per the compatibility of the Open Hardware,
thus enabling them to work in coordination. The internet source can be of user
choice like Ethernet, wireless connectivity or 4G connection (SIM Card based)
as available based on the environment.
Figures 3 and 4 shows the architecture of the system and the connection
diagram between the Temperature/Humidity sensor with the Arduino Uno open
source board and the ESP8266 wifi module board respectively [5–7].
Table 1. NodeMCU and Arduino Uno specification and feature based comparison
First, the DHT library is to be added which is available on the Arduino official
website [14]. Once the library is included, the pin number is defined. Now, con-
nect the sensor to the respective pin thus creating a DHT object. At the time of
setup the serial communication is required to be initialized as the serial monitor
will be used to display the result. Read () function is then used to read the data
from the sensor and place the respective values of humidity and temperature
into corresponding humidity (h) and temperature (t) variables. Once the above
tasks are completed the serial monitor will display the humidity and temperature
readings.
Figure 5 shows the ESP8266 being connected with DHT11 Temp/Humid sen-
sor. This reading can also be extended to be viewed remotely via the Blynk App
and also used to trigger an actuator based on a threshold thus helping us to
remotely control appliances based on the threshold level of temperature and
humidity with the help of Open Source hardware and IoT.
Firstly connection to the blynk app is established via the ESP8266 board
which will be interfacing with the DHT11 temperature and humidity sensor.
Once the app is downloaded on our mobile device, we need to create a new
project on the app and select the device as ESP8266. Once project is created,
an authorization token is being issued which is then used as the key to connect
the mobile device to the ESP8266 board via the blynk cloud. For Arduino Uno
board, a separate Wifi shield is required to enable it to connect to the mobile
device via the internet.
Role of Open Hardware and IoT in Home Automation: A Case Study 259
Fig. 6. Real time temperature and humidity reading on mobile device [10]. (Color
figure online)
Fig. 7. Temperature and humidity variation over a period of time. (Color figure online)
Role of Open Hardware and IoT in Home Automation: A Case Study 261
Figure 7 shows the humidity and temperature over a span of time graphi-
cally. The blue and the red lines depicts the temperature and humidity variation
respectively over a span of time. The X-axis represents the span of time and the
Y-axis represents the range of temperature and humidity. Here the temperature
and humidity variation is manually controlled over a very short span of time
(reading taken every 5 s) to visualize the variation scenario. In the real field of
study, the data gathered will be spread over a larger and broader span of time.
Thus, based on the data gathered and tracking done, its possible to take a
decision to smartly control our home appliances, for example.
– If the Temperature and Humidity are above a threshold, Air Conditioning
will turn ON
– If Temperature is below a threshold, Heater will be working, etc.
5 Conclusion
In this paper, the case study experimented can measure temperature and humid-
ity of a room using a sensor with help of an Open Source hardware platform
and also read the real time humidity and temperature values remotely with the
help of a mobile application using IoT based cloud platform, thus helping us to
monitor the readings over a period of time and enable us to analyze the read-
ing and accordingly perform necessary statistical calculations and also trigger
various actuators to control house hold appliances in order to promote smart
home automation and energy efficiency. Thus we can conclude this paper as an
experimental work on a specific case study to visualize home automation from a
particular point of view using Open Source hardware and IoT thus emphasizing
their role in its implementation.
References
1. Lueth, K.L.: IOT basics: getting started with the Internet of Things. IOT Anal.
1–8 (2015)
2. Kunal, D., et al.: Smart home automation using IOT. Int. J. Adv. Res. Comput.
Commun. Eng. (IJARCCE) 6(2), 576–578 (2016)
3. Saiteja, K., et al.: Home automation using IOT. Int. J. Eng. Trends Technol.
(IJETT) (2017). ISSN 2231–5381, 229–233
4. Patnaikuni, D.R.P.: A comparative study of Arduino, Raspberry Pie, ESP8266 as
IOT development board. Int. J. Adv. Res. Comput. Sci. 8(5) (2017). ISSN 0976–
5697
5. Malathi, M., et al.: Home automation on ESP8266. SSRG Int. J. Comput. Sci.
Eng. (2017). ISSN 2348–8387
6. Sipani, J.P., et al.: Temperature and humidity monitoring and control system based
on Arduino and SIM900A GSM shield. Int. J. Electr. Electron. Data Commun.
5(11), 62–68 (2017)
7. Kale, V.S., Kulkarni, R.D.: Real time remote temperature and humidity monitoring
using Arduino and Xbee S2. Int. J. Innov. Res. Electr. Electron. Instrum. Control
Eng. 4(6), 175–179 (2016)
262 A. De et al.
8. Liu, R., Wang, J.: Internet of Things: application and prospect. In: MATEC
Web of Conferences, vol. 100, p. 02034 (2017). https://doi.org/10.1051/matecconf/
201710002034
9. Pyramid Technology. http://pyramidtec.com
10. Blynk IoT Platform. https://www.blynk.cc
11. Disruptive technologies: advances that will transform life, business, and the global
economy (2013). https://www.mckinsey.com/business-functions/digitalmckinsey/
our-insights/disruptive-technologies
12. https://opensource.com/resources/what-open-hardware
13. Ashwini, P., et al.: A comparative study of Arduino Uno, NodeMCU in IoT. Int.
J. Sci. Res. Sci. Eng. Technol. (IJSRSET) 4(7) (2018). ISSN 2394–4099
14. Arduino Official Website. www.arduino.cc
Design and Simulation of Load Balancing
Algorithms in Wireless Mesh Network
Priyanka Sarkar(B)
Abstract. This work deals with balancing the load which is caused due
to the congestion created in Wireless Mesh Network. It is known that
each Mesh Router is associated with a gateway for transmission and
reception of internet traffic. So, if a lot of Mesh Routers are served by
a single gateway from the gateway set, the gateway becomes congested.
This deteriorates the overall performance of the network. Many algo-
rithms for load balancing have been introduced to solve this problem. In
this paper, a hop count based load balancing algorithm and a Last in
first out based load balancing algorithm have been proposed to balance
the load in Wireless Mesh Networks (WMNs) having multiple gateways.
The algorithms in the proposed scheme are executed when the network
is found to be overloaded. The proposed schemes reroutes the flow of
the mesh router from the overloaded domain by taking care that the
flow of the rest of the mesh routers in the overloaded domain should not
be hampered. The proposed schemes are compared with each other and
have been simulated. The simulation results show that the hop count
based load balancing algorithm balances the load more efficiently than
Last in first out based load balancing algorithm.
1 Introduction
Many wireless networks have been introduced into the next generation to have
better services. Now-a-days, wireless LAN technology is gaining lots of popu-
larity as it provides mobile access in remote areas and public spaces. When the
need of mobility is being realized, this technology has been introduced in the
mainstream technology. Recently, due to the rapid development of networks,
people always want to access high-bandwidth network anytime and anywhere
for enjoying ubiquitous network services. For this, new innovative wireless tech-
nologies are introduced. Wireless Mesh Networks (WMNs) are new kind of wire-
less multi-hop network architecture. The performance of WMNs is influenced
by radio techniques, scalability, mesh connectivity, security, compatibility and
Supported by NIT Agartala.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 263–272, 2020.
https://doi.org/10.1007/978-981-15-3666-3_22
264 P. Sarkar
2 Related Work
3 Proposed Scheme
Hop Count Based Load Balance Scheme (HCLB): In this section the hop
count based algorithm for load balancing has been proposed. If the network is
found to be overloaded, load balancing algorithm will be executed. It balances
the load by changing the route of the traffic flow of the sink from the overloaded
domain to the under loaded domain.
At first, the sinks are waiting to be assigned to their primary gateways.
The gateways periodically send beacon signal to advertise their presence.
After receiving the advertisement, the MR registers itself to that gateway as
its primary gateway. When the MR needs to send a data packet to the destina-
tion, it sends the route request to the gateway. After receiving the route request,
the gateway checks whether the Mesh Router is sending route request for the
first time or not. If yes, the entry of the MR under the gateway is done in a
table. Each IGW checks its load every time it receives the request from a new
mesh router. The load of domain d is defined as:
load(d) = total number of sinks under the domain.
If the domain is found to be overloaded, the algorithm is executed to balance
the overload of the network. The overload of a domain d is given by:
0, if load (d) < C
Overload(d) = (1)
load(d) − C, otherwise
Here C is the capacity of the domain d, i.e, the number of MRs, a gateway
can serve without being overloaded.
The sinks under the overloaded domain are sorted in decreasing order of
distance from the overloaded gateway. The MR having the highest hop count
in this order is selected to migrate its flow to another domain. The idea behind
this is that the probability of most distanced node being the boundary node is
very high. The MR then starts selection of new gateway for routing its internet
traffic. The set of neighboring under loaded gateway are arranged in ascending
order of their hop distance from the MR. The gateway, then sends a route error
message to that MR. After receiving the route error message, the MR deletes
its route to the gateway. The gateway having the least hop count is selected as
serving gateway. After the gateway selection, the route to the new gateway is
established. Now the MR continues its flow through the new gateway.
Figure 2(a) shows the algorithm for HCLB. Here C is the capacity of each
gateway and G is the number of gateways in the network.
Design and Simulation of Load Balancing Algorithms in WMN 267
Last in First Out Based Load Balancing (LBLB): This section pro-
poses the hop count based algorithm for load balancing. In this procedure, the
mesh routers are stored in a queue. Each gateway maintains a separate queue.
When a mesh router appears to the network, its existence is known to the gate-
way. Each gateway maintains a Gateway Queue to keep track of the mesh routers
which are associated to that gateway. The gateway, then, makes an entry of the
new mesh router in that queue. The entry will be in Last in first out method. If
the gateway becomes overloaded, it executes the LBLB algorithm. As the name
implies, here the mesh router which has entered last in the gateway queue is first
shifted to another domain for load balancing. This technique is applied because
the mesh router which has arrived lastly in the internet is least associated with
other mesh routers and the transmission and reception of packets.
Figure 2(b) shows the algorithm for LBLB. Here C is the capacity of each
gateway.
4 Performance Analysis
This section presents extensive simulation results that illustrate the impact of
using proposed load balancing schemes. Network Simulator 2 (NS2) has been
used to simulate the proposed schemes. The performance before and after load
balancing has been compared.
Figure 3 shows the scenario used for simulation. The scenario has 11 CBR
flows generated randomly to the 3 gateways. The other simulation parameters
are shown in the Table 1. Finally, the throughput, average delay and calculated
packet delivery ratio has been compared so that the load-balancing ability of
the proposed scheme can be proved. The values of HCLB are compared to the
values before load balancing and LIFO based load balancing (LBLB).
Parameter Values
Scenario size 800 m × 500 m
Transmission range 250 m
Mesh router Generated 11 nodes
Gateway Generated 3 gateways
Traffic type CBR
Flows 11 CBR flows
Packet size 512 bytes
Simulation time 300 s
Design and Simulation of Load Balancing Algorithms in WMN 269
Throughput is the total amount of data transferred from the source to des-
tination or processed in a specified amount of time. Here it is shown that the
throughput of the network is increased after the HCLB algorithm is executed.
The throughput in LBLB is less than HCLB because in LBLB, the MR to be
shifted to the other domain for load balancing may be nearer to the gateway. So,
other MRs may connect to the gateway through the selected MR. If the nearer
MR is shifted to another domain, it deteriorates the performance of other MRs
also. It degrades the throughput. But in HCLB, as the MR having maximum
distance from the gateway is selected to reroute the flow and it increases the
throughput. Figure 4 shows the comparison of throughput for both the schemes.
The average delay of a network is the time taken by a bit of data to travel
across the network from one node or endpoint to another. From the observation,
it is shown that the average delay is decreased after the execution of HCLB.
Before the execution of the load balance scheme, congestion occurs at the gate-
ways and delay is increased. As the proposed scheme distributes the load uni-
formly among the IGWs, it reduces the average delay. In LBLB, the selected
MR to shift another domain may be the intermediate MR from the gateway.
In such situation, if the MR is shifted, other MRs which were connected to the
gateway through this MR, have to configure their routes to gateway again. This
causes delay in the network. This deteriorates the performance. Figure 5 shows
the decrease of average delay after execution of the proposed scheme.
270 P. Sarkar
Packet delivery ratio is the ratio of the number of delivered data packet
from source to destination. The analysis shows that after HCLB is executed, the
packet delivery ratio of the network is increased. HCLB algorithm distributes
the load in such a way so that all the IGWs are utilized properly. In LBLB the
MR which is to reroute its flow may be the intermediate node. In such case,
the other MRs connected to the gateway through this MR has to reconfigure
their routes. For this reason, the packets already arrived to those MRs cannot
be delivered properly. So the delivery of packets gets affected. Figure 6 shows
increase of the packet delivery ratio after the execution of the load balancing
algorithm.
Design and Simulation of Load Balancing Algorithms in WMN 271
References
1. Zhang, Y., Luo, J., Hu, H.: Wireless Mesh Networking: Architectures, Protocols and
Standards. Auerbach Publications, Boca Raton (2006)
2. Akyildiz, I.F., Wang, X., Wang, W.: Wireless mesh networks: a survey. Comput.
Netw. J. 47(4), 445–487 (2005)
3. Akyildiz, I.F., Wang, X.: A survey on wireless mesh networks. Comput. Netw. ISDN
Syst. (2005)
4. Nandiraju, D., Santhanam, L., Nandiraju, N., Agrawal, D.P.: Achieving load bal-
ancing in wireless mesh networks through multiple gateways. In: Mobile Adhoc and
Sensor Systems (MASS), pp. 807–812. IEEE International Conference (2006)
272 P. Sarkar
5. Liao, G.-L., Chen, C.-Y., Hsu, S.-W., Wu, T.-Y., Chao, H.-C.: Adaptive situation-
aware load balance scheme for mobile wireless mesh networks. In: Computer Com-
munications Workshops (INFOCOM WKSHPS), pp. 391–396. IEEE (2011)
6. Galvez, J.J., Ruiz, P.M., Skarmeta, A.F.: Responsive on-line gateway load-balancing
for wireless mesh networks. Ad Hoc Netw. J. 10, 46–61 (2012)
7. Choi, H.-G., Han, S.J.: Load balancing routing for wireless mesh networks: an adap-
tive partitioning approach. In: Consumer Communications and Networking Confer-
ence, CCNC 2008, pp. 1–5 (2008)
8. He, B., Sun, D., Agrawal, D.P.: Diffusion based distributed internet gateway load
balancing in a wireless mesh network. In: Global Telecommunications Conference
2009, pp. 1–6 (2009)
9. Ma, L., Denko, M.K.: A routing metric for load-balancing in wireless mesh net-
works. In: 21st International Conference on Advanced Information Networking and
Applications Workshops (AINAW 2007) (2007)
6LoWSD: A Protocol Platform
for Interoperability Between SDN
and IoT
1 Introduction
The Internet of Things (IoT) is refer to as the Internet of Everything or the
Industrial Internet. This is a new technology paradigm, envisioned as a global
network of machines and devices that are capable of interacting with each other.
The IoT is recognized as one of the most important emerging areas of the future
technology and it succeeds in getting attention from a wide range of indus-
tries. The real utility of IoT for enterprises can be fully realized when connected
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 273–287, 2020.
https://doi.org/10.1007/978-981-15-3666-3_23
274 W. Khongbuh and G. Saha
devices are able to communicate with each other and integrate with respective
vendor-managed inventory systems, customer support systems, business intelli-
gence applications, and business analytic [1].
The Three Layer Architecture of the IoT system is described in the bottom
to top hierarchy as defined below:
1. Data Plane: Data Plane consists of networking elements (switch, router, etc).
The data plane switches maintains a flowtable for the traversal of the data.
This flowtable is configured by the Controller via SouthBound Interface (SBI).
One such SBI is the OpenFlow protocol.
2. Control Plane: Forwarding devices are scheduled by the Controller. The con-
troller is a decision making entity which governs the overall operation of the
network.
3. Application Plane: Network services can be provided to the user via this
plane. The communication between Control Plane and Application is through
NorthBound Interface (NBI).
The SDN based IoT will be able to manage the network more efficiently [6].
Absence of an intermediate protocol for bridging the two technologies came as
a disadvantage. In this light, a new research initiative has been introduced in
this paper, where SDN has been integrated with IoT infrastructure. Further to
encounter the interoperability issue between the two, a new protocol 6LoWSD
has been tried. Some of the notable protocols of IoT are ZigBee, 6LoWPAN,
etc and that of SDN includes OpenFlow, Border Gateway Protocol (BGP), etc.
In this context we have used the IoT protocol - 6LoWPAN with that of SDN
protocol - OpenFlow for experimental purpose. For this purpose, an integrated
communication protocol 6LoWSD has been proposed in this work. This proposed
protocol combines the features of 6LoWPAN and OpenFlow protocol to improve
the interoperability constraint.
The contents of the paper are organized as follows: Sect. 2 explains features
of IoT protocol - 6LoWPAN and SDN protocol - OpenFlow. The Challenging
interoperability issue is discussed in Sect. 3. The literature review is given in
Sect. 4. The proposed 6LoWSD protocol is presented in Sect. 5. Experimentation
and results are discussed in Sect. 6. Finally, Sect. 7 concludes the paper.
276 W. Khongbuh and G. Saha
‘User Datagram Protocol (UDP)’, which can also be compressed using the LoW-
PAN format. The ‘Internet control message protocol version 6 (ICMPv6)’ is used
to control messaging process. Adaptation between IPv6 and LoWPAN format is
done by routers at the edge of the 6LoWPAN, referred to as edge routers. This
transformation is transparent, efficient and stateless in both directions [12].
constrained by energy and resources [7]. With the SDN approach, flexibility can
be achieved as to allow different objects connected to heterogeneous networks to
allow communication with each other.
4 Literature Review
IoT and SDN are the emerging field in the networking domain. Many experts
proposed improved SDN based IoT architecture and with some impasses on pro-
tocol bridging between the two. Some of the most prominent works are described
in this section which can effectively solve the challenging issue between the two.
‘SD-WISE’ (Software-Defined WIreless SEnsor network) [16] focuses their
attention on the two main problems: network flexibility and security. SD-WISE
extends the SDN approach to WSN by introducing a more flexible way to define
flows as well as the possibility to control the duty cycles of the node radios to
increase energy efficiency.
‘SD6WSN’ (software-defined 6LoWPAN wireless sensor network) architec-
ture developed by Miguel et al. [17] aims to control the routing contention of the
data traffic in 6LoWPAN according to the SDN approach. It takes into account
the specific resource constraint WSN issues such as low data transfer rate, high
latency and low processing power. By incorporating the flow-based forwarding
flexibility provided by SDN, a development of specific networking applications
based on a unified view has be achieved.
de Oliveira et al. [18] defined ‘Spotled’ an architecture with distributed and
hierarchical controller. The local controllers use local information to reply to
nodes in its domain and a global controller maintain the whole network state.
This architecture is able to minimize the impact of the control traffic throughput
introduced by the SDN paradigm in WSN, thus reducing the data flow rate
installation on the network.
‘SDN-WISE’ [19] is another SDN solution for WIreless SEnsor networks. It is
a stateless based protocol. SDN-WISE aims to reduce the amount of information
exchanged between the SDN network controller and sensor nodes. SDN-WISE
offers APIs that allow software developers to implement the SDN Controller pro-
grammatically. This increases flexibility and simplicity in network programming.
‘SDIoT’ [20] is another comprehensive SDN based framework model. It aims
to simplifies the IoT management process found in traditional IoT. The archi-
tecture allows to forward, store, and secure the produced data from the IoT
objects. It integrates the SDN, software defined storage, and software defined
security into one software defined based control model. It is less efficient then
SDN-WISE as it merge every entities of the model into one single entity. The
failure of which can degrade the overall performance.
The proposed ‘TinySDN’ architecture [21] discusses about a multiple
controller TinyOS-based architecture and hardware independent systems for
software-defined wireless sensor network nodes. TinySDN provides flexibility in
communication. It introduced memory overhead that does not hinder other fea-
tures related to the WSN application. It suffers from the drawback of delay in
forwarding packets from the control plane level.
280 W. Khongbuh and G. Saha
Han and Ren [22] have used OpenFlow protocol in Wireless Sensor Network
(WSN). The centralized concept of SDN implies by having a master node (con-
troller) programmed by NOX as Network Operating System (NOS) and a centre
nodes (OpenFlow switches). In this work, some solution for WSNs as per with
the OpenFlow paradigm on a traditional network is presented.
The proposed ‘Sensor OpenFlow system’ [23] has two components: an archi-
tecture featuring a clear demarcation between the control and data plane. It
is a core component that standardizes the protocol pertaining to communica-
tion between the two planes. The authors address resource under-utilization and
counter-productivity by promoting sharing of hardware resources between dif-
ferent application and reuse of implemented functionalities.
A good minimal amount of work based on the above mentioned literature deal
with SDN-IoT architecture to address the IoT challenges. The proposed protocol,
6LoWSD intends to provide a probable solution as well for a systematic network
by integrating SDN feature (SBI) into IoT ecosystem. This new collaboration
of 6LoWPAN with OpenFlow architecture helps to resolve the interoperability
constraint of IoT devices. Other issues like power management, scalability and
reliability of the existing IoT devices can be mitigated to some extend using the
proposed 6LoWSD Protocol.
Internet of things (IoT) uses a low-rate wireless personal area network with a
short communication range and within limited resource. Software Defined Net-
work (SDN) has a centralized controller managing the devices of its domain and
maintaining the whole network view. With time the wireless network is expand-
ing exponentially with its devices embedded with various heterogeneous compo-
nents. The IoT system in this respect is facing several challenges with respect to
management, interoperability and power management. In this context our cur-
rent network architecture approach we have integrated the SDN infrastructure
features with the IoT network system. In this regard our main purpose lies in
alleviating the interoperability challenges prevailing on the existing IoT system.
However, the challenge remains here, with the bridging of the different architec-
tures protocols: 6LoWPAN and OpenFlow. Our objective endeavour is therefore
directed as how to manage interoperability between 6LoWPAN and OpenFlow.
The Fig. 8 represents the abstract of the proposed 6LoWSD protocol in block
diagram.
The proposed 6LoWSD protocol operates on the propriety 802.15.4 standard
at the Physical and the Link Layer. The Network Layer interfaces 6LoWPAN
with the adaptation layer. The SDN concepts are used for decoupling the control
and the data plane of the devices. Initially, OpenFlow assigns each nodes as SDN
enabled sensor devices. Thereby some nodes acts as the data plane devices while
one or few others take the role of the controller. The data nodes are simply
nodes based on flowrules and meant for sending data to the controller. The
controller nodes maintains the functionality and the controlling management
6LoWSD: A Protocol Platform for Interoperability Between SDN and IoT 281
else
The Sink controller will -
Create the Entry.
Allocate an entry.
Create the new rules.
Add the new rules.
end if
else
If Packets don’t confirms and have no relevance
Drop the packet
end if
return SDN-IoT Compatible Packets
6 Experimental Setup
6.1 Experimentation
The proposed protocol has been experimented on simulation in Contiki using the
Cooja simulator. While performing experimentation for the proposed protocol
on Contiki platform, two object nodes were considered: an iotnode and a con-
troller. The object nodes comply with the sky mote configuration on Contiki.
The object nodes configuration were compiled according to the flow concept
of the OpenFlow architecture. The initial configuration of the node includes
Node Configuration, Routing tables, Packet transmission and Flow-Tables con-
figuration. The experimental attributes and parameters tends to change with
the mobility of the topology. In the Node Configuration mode, it initializes the
node with an address. It initializes the node with a packet header, a flow table
entry which defines the basic node standard configuration. Here in the Contiki
platform, the node is a sky mote, which is an MSP430-based board with an
IEEE 802.15.4-compatibility. This is equipped with a CC2420 radio chip. The
Routing Table contains the node addresses of the adjoining or the neighbour
nodes. It defines the necessary routing information for the nodes to traverse
along the topology of the network. Packet transmission contains the configura-
tion of the packets: its header contents and the necessary source and destination
configuration. In the Flow tables, rules, actions and statistics are configured.
The working principle of the proposed protocol involves routing configuration
for the low power mesh network topology. As given in the proposed algorithm, it
checks initially whether the beacon signal is broadcasted or unicasted. Then the
nodes are assigned with proper IP addresses. A packet header is created for main-
taining the packet structure of the header as “header; netId; addr; src addr; dest
addr;”. After receiving the beacon signal, the adjacent nodes will also broadcast
their Neighbour Tables, which shows their address and relative distance from
each other. The rules of the flowTable will direct further the transmission or
dropping of the packet as required.
6LoWSD: A Protocol Platform for Interoperability Between SDN and IoT 283
6.3 Results
The results obtained from simulated experiments are presented. The modeling
behavior of the iotnodes and the controller have been executed in simulation.
The simulated network topology used for the experimentation is presented in
Fig. 10. Here, iotnodes are configured on the 802.15.4 standard and they interact
on the network layer with the other networks using the adaptation layer of the
LoWPAN. Accordingly, controller node 1 and 2 acts as the SDN controllers.
These controllers determine the policies and flow rules, which can be applied for
the further operation of the simulated iotnodes. The simulated controller can
be used to create an extended virtual network. As a result, the simulated and
real nodes can be fully integrated and they can interact with each other. This
will be useful for real-time test-bed networking scenario. In the experimentation
process, due to space constraints and memory limitation, the experimentation
was carried out in a limited virtual simulated environment. However, the initial
results produced during the experimentation was quite encouraging.
The results obtained are based on various test scenarios with the parameters
indicated in the Table 2.
284 W. Khongbuh and G. Saha
Parameters Values
IoT nodes 8
Controller 2
Elapsed time 00:10:00
Avg. packets/sec 45
Avg. packets size 40–50 bytes
Burst rate 1.4
Total packets transmitted 55972
Parameters Values
IoT nodes 8
Elapsed time 00:10:00
Avg. packets/sec 55
Avg. packets size 56 bytes
Burst rate 0.05
Total packets transmitted 8863
7 Conclusion
The integration of SDN-IoT model is a very challenging task. The present
IoT network suffers from the resource constraint limitations thus compelling
researchers to develop a resource conscious system. This still remains illusive
despite many efforts by the past researchers. The introduction of the concept
of SDN in IoT system opens a path which helps in leveraging the challenges
of the resource constraint in IoT. But here the main constraint becomes inter-
operability between IoT and SDN platforms. In our proposed work, the SDN’s
OpenFlow concepts has been introduced into the IoT network protocol frame-
work. This has given birth to a new protocol 6LoWSD. The aim of this protocol
is to enhance the interoperability of these two heterogeneous platforms. The
same was experimented on simulation on Contiki using Cooja simulator. The
result obtained was much encouraging in the present limited hardware environ-
ment. The same experiment can be extrapolated in real time SDN based IoT
network also, which remains a future scope of work. The same can also be exper-
imented in dynamic mobile environment which also remain future scope of work.
The results obtained while comparing with the conventional 6LoWPAN under
the given condition, found that the proposed algorithm displayed comparable
performance with respect to interoperability.
References
1. Lee, I., Lee, K.: The Internet of Things (IoT): applications, investments, and chal-
lenges for enterprises. Bus. Horiz. 58(4), 431–440 (2015)
2. Zanella, A., Bui, N., Castellani, A., Vangelista, L., Zorzi, M.: Internet of Things
for smart cities. IEEE Internet Things J. 1(1), 22–32 (2014)
3. Silva, B.N., Khan, M., Han, K.: Internet of Things: a comprehensive review of
enabling technologies, architecture, and challenges. IETE Tech. Rev. 35(2), 205–
220 (2018)
4. Al-Fuqaha, A., Guizani, M., Mohammadi, M., Aledhari, M., Ayyash, M.: Internet
of Things: a survey on enabling technologies, protocols, and applications. IEEE
Commun. Surv. Tutor. 17(4), 2347–2376 (2015)
5. da Cruz, M.A., Rodrigues, J.J.P., Al-Muhtadi, J., Korotaev, V.V., de Albuquerque,
V.H.C.: A reference model for Internet of Things middleware. IEEE Internet Things
J. 5(2), 871–883 (2018)
6. Qin, Z., Denker, G., Giannelli, C., Bellavista, P., Venkatasubramanian, N.: A soft-
ware defined networking architecture for the Internet-of-Things. In: 2014 IEEE
Network Operations and Management Symposium (NOMS), pp. 1–9. IEEE (2014)
7. Jarraya, Y., Madi, T., Debbabi, M.: A survey and a layered taxonomy of software-
defined networking. IEEE Commun. Surv. Tutor. 16(4), 1955–1980 (2014)
8. Hu, F., Hao, Q., Bao, K.: A survey on software-defined network and openflow: from
concept to implementation. IEEE Commun. Surv. Tutor. 16(4), 2181–2206 (2014)
9. Kobo, H.I., Abu-Mahfouz, A.M., Hancke, G.P.: A survey on software-defined wire-
less sensor networks: challenges and design requirements. IEEE Access 5, 1872–
1899 (2017)
6LoWSD: A Protocol Platform for Interoperability Between SDN and IoT 287
10. Thubert, P., Hui, J.W.: Compression format for IPv6 datagrams over IEEE
802.15.4 - based networks. https://tools.ietf.org/html/rfc6282. Accessed 18 May
2019
11. Shelby, Z., Bormann, C.: 6LoWPAN: The Wireless Embedded Internet, vol. 43.
Wiley, Hoboken (2011)
12. Palattella, M.R., et al.: Standardized protocol stack for the Internet of (important)
Things. IEEE Commun. Surv. Tutor. 15(3), 1389–1406 (2012)
13. Baktir, A.C., Ozgovde, A., Ersoy, C.: How can edge computing benefit from
software-defined networking: a survey, use cases, and future directions. IEEE Com-
mun. Surv. Tutor. 19(4), 2359–2391 (2017)
14. Flowgrammable: SDN Protocol Decomposition. http://flowgrammable.org/sdn/
openflow/. Accessed 10 March 2019
15. Kreutz, D., Ramos, F., Verissimo, P.E., Rothenberg, C.E., Azodolmolky, S., Uhlig,
S.: Software-defined networking: a comprehensive survey. Proc. IEEE 103(1), 14–
76 (2015)
16. Anadiotis, A.-C., Galluccio, L., Milardo, S., Morabito, G., Palazzo, S.: SD-WISE:
a software-defined wireless sensor network. Comput. Netw. 159, 84–95 (2019)
17. Miguel, M., Jamhour, E., Pellenz, M., Penna, M.: SDN architecture for 6LoWPAN
wireless sensor networks. Sensors 18(11), 3738 (2018)
18. de Oliveira, B.T., Margi, C.B.: Distributed control plane architecture for software-
defined wireless sensor networks. In: 2016 IEEE International Symposium on Con-
sumer Electronics (ISCE), pp. 85–86. IEEE (2016)
19. Galluccio, L., Milardo, S., Morabito, G., Palazzo, S.: SDN-wise: design, prototyping
and experimentation of a stateful SDN solution for wireless sensor networks. In:
2015 IEEE Conference on Computer Communications (INFOCOM), pp. 513–521.
IEEE (2015)
20. Jararweh, Y., Al-Ayyoub, M., Benkhelifa, E., Vouk, M., Rindos, A., et al.: SDIoT: a
software defined based Internet of Things framework. J. Ambient Intell. Humaniz.
Comput. 6(4), 453–461 (2015)
21. De Oliveira, B.T., Gabriel, L.B., Margi, C.B.: TinySDN: enabling multiple con-
trollers for software-defined wireless sensor networks. IEEE Lat. Am. Trans.
13(11), 3690–3696 (2015)
22. Han, Z., Ren, W.: A novel wireless sensor networks structure based on the SDN.
Int. J. Distrib. Sens. Netw. 10(3), 874047 (2014)
23. Luo, T., Tan, H.P., Quek, T.Q.: Sensor OpenFlow: enabling software-defined wire-
less sensor networks. IEEE Commun. Lett. 16(11), 1896–1899 (2012)
Machine Learning Based Efficient
Multi-copy Routing for OppIoT Networks
1 Introduction
With the development in technogies, more and more devices are getting con-
nected to the Internet, which leads to the new criterion in computing and com-
munication known as the Internet of Things (IoT) [1]. As a result, more devices
are getting prevalently monitored and cannot be deployed statically. These IoT
devices while being transported by humans or different carriers will interact
either by statically or dynamically. For instance, in vehicular based IoT network
statically deployed devices may entail to recognize the mobile agents and employ
them in assembling and delivering their information to the intended target [2].
In recent times, opportunistic mobile networks rose as a novel mechanism for
transmissions in wireless systems. Unlike mobile ad hoc networks (MANETs)
[3] that needs end-to-end connected routes for information transfer, the com-
munication in opportunistic mobile networks takes place on the corporation of
opportunistic connections among mobile nodes without the availability of end-
to-end communication paths. These mobile devices establishes connection when
people comes in contact and these network type are considered as human social
networks. Hence, the opportunistic networks utilize the individual behaviors and
social relations to provide more effective and reliable message distribution sys-
tems. In OppIoT [4], data acquisition, propagation, and distribution of data
which typically necessitate opportunistic connections utilizing mobility, short-
range transmission, and technologies. During such operations, routing protocol
configuration poses challenge due to its inherent complexity in securing connec-
tivity among devices and recognizing a relevant forwarder to forward data packet
to its desired destination. Opportunistic networks [5], which are the derived class
of OppIoT and considered as evolved form of MANETs. In OppNets, nodes are
allowed to establish communication among several different nodes even though
they don’t have routing connection. Nodes are connected momentarily and the
routes keep dynamically changing due to nodes mobility or frequent nodes acti-
vation and deactivation. When the packets are en route between the source
node and the target nodes, any viable node in between can be opportunistically
employed as the next hop whenever it is plausible to deliver the packets to the
intended destination with minimum routing distance. Several protocols related to
OppNets that have been proposed such as epidemic, HBPR, HibOp etc., which
uses parameters like message propagation pattern, context erudition of nodes
(i.e. current position, the possibility of delivery, target node distance, prior and
existing history of the node), last synergy time and various other parameters in
order to predict which node is opted as next hop to deliver the message to the
intended destination. In this work, a machine learning based algorithm which
can be employed to OppNets is proposed. The proposed algorithm that differ
radically from the existing algorithm as this algorithm uses the random forest
to decide the neighboring node to route the data packet based on the present
context and location information. Random forest or otherwise referred to as ran-
dom decision forest are one of the most popular ensemble learning techniques
that are used in both classification and regression. Several decision trees are ran-
domly created with several subsets of data during the training period. The class
290 N. N. Srinidhi et al.
is assigned based on the mode. Random forest algorithm uses the technique of
bootstrap aggregation, which helps in improving the stability and accuracy of
the algorithm along with reducing the variance and overfitting of the proposed
iProphet.
The rest of the paper is organized as follows. Section 2 presents a summary of
various existing works with respect to multi-copy and oppurtunisitic networks.
Problem statement is presented in Sect. 3. Proposed algorithm has been pre-
sented in Sect. 4. The experimental evaluation and results has been explained in
Sect. 5. Conclusion along with future work has been presented in Sect. 6.
2 Related Work
Many protocols that are used in OppNets are related to the nodes context-
information and message distribution. Representative ones are summed as fol-
lows. Epidemic routing [6], is a flood-based routing algorithm whose foremost
objective is to deliver the packet with high probability to the desired destina-
tion with a minimum premise about the topology and the network. The work
is based on epidemic algorithm which manages two buffers one for the message
originating from the node itself and the other buffer is used to store messages
generated from different nodes. Each node contains a summary vector that car-
ries a short description of messages currently cached in the buffer. When nodes
come in contact of one another, they revise their summary vectors and each
node has an extra buffer to prevent verbose connection with other nodes. Nodes
exchange their summary vector with each other after a predefined time and after
exchanging their vectors the nodes rival their vectors to ascertain which message
is missing. Then, the nodes request for the copies of the messages which they
do not contain. But the proposed epidemic protocol uses higher bandwidth and
buffer space which will cause network congestion and it do not examine nodes
parameters to predict the possibility of message delivery. The HBPR (History
Based Prediction for Routing), protocol which was proposed by Dhurandher
et al. [7], uses election of the intermediate node as the message carrier is decided
based on specific parameters. These parameters are the duration for which a
pair of nodes meet, node’s history, time at which a couple of nodes meet, and
the path of the mobility of the node concerning the root and the target node.
The schema consists of three stages, i.e. identifying the home location, wherein
the nodes movement is utilized to recognize the subsequent mode location; the
message production stage, during which the intermediate node is determined,
and also the home location is renewed and lastly, the subsequent hop selection
phase, where metric is determined to help to judge on the assortment of the best
succeeding message forwarder. This protocol has achieved improved results when
compared with Epidemic concerning the message delivery number and overhead
ratio, but only after considering the individual mobility paradigm. Prophet rout-
ing protocol [8], uses non-arbitrary movement and association patterns in prac-
tical application situation to replicate the message to different nodes in order to
enhance the routeing performance. It is established on the basis that if a node
has communicated with a different node or visited a region regularly, then there
Machine Learning Based Efficient Multi-copy Routing for OppIoT Networks 291
3.1 Challenges
There are various challenges that opportunistic network in IoT faces. This is due
to the dynamic nature of nodes over time depending on the application, which
makes it difficult for routing the information and to identify the intermediate
nodes which are used for forwarding messages to destination.
– Data extraction: For training the machine learning based random forest
model, a huge data set has to be obtained. To obtain data that is close to the
real world was a challenge. ONE simulator [10], provides the closest solution
and to extract the data from the ONE simulator, the source code has to be
modified in order to train the model.
292 N. N. Srinidhi et al.
– Model Creation: After extracting the data from ONE simulator, a robust
random forest classifier had to be constructed. In order to do so, features and
algorithm related parameters has to be chosen wisely.
– Model Integration: After the model is constructed, it has to be properly inte-
grated with the ONE simulator in order to classify the nodes.
4 Proposed Algorithm
This section provides in detail explanation of the proposed iProphet algorithm.
4.1 Terminology
– S: Source node
– D: Destination node
– X: Selected forwarder node
– G: Set of all neighboring nodes
– Ni : Neighboring node i in the set S
– Mi : Message
– n: Current node, tasked with making a routing decision
– M : Set of all messages that are generated during data generation phase
– Ms : Set of messages that are delivered
– Md : Set of messages that are dropped or aborted
– E: Set of all the events recorded during data extraction phase of Prophet
– P (Ni ): Delivery probability as calculated by Prophet for the neighbor node
Ni using current node n
– f eature name(x): Value of the attribute feature name in the record x in any
set.
history between the nodes to calculate the delivery probability between them.
Even though this algorithm provides a sufficient performance, the method to
select the next node could be optimized by applying machine learning models.
Prophet does not completely utilize all the context information that are present
in the nodes. With the data driven approach, it is possible to select best pre-
dictors/features and use them to drive the routing decisions. In the proposed
Random Forest Based Optimized Multi-Copy Routing for Opportunistic IoT
Networks, random forest algorithm is used to choose the neighboring node to
route the data packet based on the present context and location information.
Random forest are one of the most popular ensemble learning techniques that are
used for both classification and regression. Several decision trees are randomly
constructed with varied subsets of data during the training period. The class
is assigned based on the mode. Random forest algorithm uses the technique of
bootstrap aggregation, which helps in improving the stability and accuracy of the
algorithm along with reducing the variance and overfitting. Formally, consider a
training set X = x1 , ..., xn along with their corresponding outputs Y = y1 , ..., yn
giving dataset Dn = (X1 , Y1 ), ..., (Xn , Yn ). Random forest is a collection of ran-
dom base regression trees {rn(x, θm , Dn ), m ≥ 1}, where θ1 , θ2 , ... are outputs of
a randomizing variable θ. The outputs of these random trees are aggregated to
get an estimate. Thus, random forest model is given by,
rn(X,
¯ Dn ) = Eθ [rn(X, θ, Dn )] (1)
Random Forest is one of the most popular machine learning algorithms for
both classification and regression. The existing Prophet algorithm does not take
advantage of the available past data, it only uses the current context information
in its static model. Thus, it could be made more robust by integrating machine
learning capabilities to it. Random forest uses the ensemble learning technique.
By constructing several trees and taking an aggregate of them would only make
the model robust. Apart from this, it also provides better classification since it
uses boosting and bagging techniques ands helps in avoiding overfitting com-
pared to their counterparts. Random forest is better suited for this particular
application because it can work on multidimensional input data and the data
294 N. N. Srinidhi et al.
yes no no yes
In the decision tree 1, the first node considers the location co-ordinates of
the current node, the selected neighbor node and makes the decision based on
the quadrant of selected neighbor with respect to current node. Selected neighbor
node is the one that the current node has established the connection with and
the current node needs to make decision about the reliability while transferring
it to the neighbor node. For each quadrant, the next decision node considers the
location co-ordinates of the destination to proceed to the next decision node. If
the Selected neighbor node belong to the first quadrant, then the destinations
lying in the first, second, and fourth quadrant are considered as valid, the Selected
neighbor node could eventually move and therefore allowed to next level of the
decision tree. Similarly, for neighbors in other quadrants, i.e. quadrants of the
destination nodes surrounding the neighbor quadrants are considered as valid.
For e.g, for second quadrant, quadrant 1, 2, 3 are the surrounding quadrants
and quadrant 2, 3, 4 are for third quadrant. If the destination doesn’t lie in the
surrounding quadrant, then the message with that destination is not transferred
to the Selected neighbor node. The next decision node considers the buffer size
of the Selected neighbor node node. The nodes with optimum level of free buffer
size in order to accommodate the incoming message are proceeded down the
tree, where as the nodes with lesser space will be discarded (i.e. the message is
not forwarded). The next decision is based on the criteria of movement speed
of the Selected neighbor node node. The movement speed is compared against
the threshold as calculated by the decision tree and is allowed to proceed along
different paths. If the movement speed is less than the threshold, the nodes are
not immediately treated as unreliable as shown in the decision nodes. Instead,
the decision nodes will make comparison of the feature value across various
other neighbors feature values to make appropriate decision regarding reliability.
Finally, the Selected neighbor node PRoPHET values meeting the threshold set
by their respective groups is considered for final message transfer. This procedure
is repeated for other base learners with different set of features and random
sampling of the data.
For a given node, extracted data about one neighbor at a time will be fed into
trained model. Based on the features, the model classifies the neighbors as good
forwarder or bad forwarder. After going through all the neighbors, data packets
are forwarded to all of the good forwarders as proposed in the Algorithm 1.
Since it is an improvement over the Prophet algorithm, the Prophet probability
is also considered one of the feature for classification. In order to see how the
model performed in real world, the source code of the Prophet algorithm in ONE
simulator was modified to include the trained model.
5 Experimental Evaluation
In this section, environment considered for simulating proposed work along with
discussion about results obtained will be explained.
296 N. N. Srinidhi et al.
Begin;
// Original Prophet
foreach node Ni as i do
dp[i] <– random()
end
for every 5 seconds do
foreach node Ni as i do
dp[i] <– GetP rophetP robabilty(i);
end
end
// End of original Prophet
foreach message m in current node n buffer do
foreach node Ni as n in neighborhood set G do
P(n) <– dp[n];
if dp[n] ≥ lP (Ni ) then
if Ni == X then
update X;
update G;
Label = RandomF orest1(X, G, P (X), P (n));
end
if Label == “deliver” then
Relay the message to node X;
else
if Label == “don’t deliver” then
Don’t relay the message to the node X;
end
end
end
if dp[n] < lP (Ni ) then
if Ni == X then
update X;
update G;
Label = RandomF orest2(X, G, P (X), P (n));
end
if Label == “deliver” then
Relay the message to node X;
else
if Label == “don’t deliver” then
Don’t relay the message to the node X;
end
end
end
end
end
Algorithm 1. Improved Prophet algorithm
Machine Learning Based Efficient Multi-copy Routing for OppIoT Networks 297
In this section, the performance of the proposed iProphet is evaluated using the
Opportunistic Network Environment (ONE) [10] simulator. The simulation is
done on a terrain size of 4,500 × 3,400 m2 with 51 to 198 nodes deployed ran-
domly. The transmission range of the node is set to 20 m. The various parameters
considered during simulation are shown in Table 1. Here the proposed iProphet
performance has been compared with Epidemic [6], Prophet [8] and Spray and
Wait [11] routing protocols, under varying TTL values from 100 to 300 s, number
of nodes from 51 to 198, and message creation interval from 25 to 35 s.
Parameter Value
Simulation area (4,500 × 3,400) m2
Total number of nodes 198
Total nodes group 6
Pedestrains group 3
Nodes number in each pedestrian group 30
Pedestrian walking speed 0.5–1.5 km/h
Pedestrian buffer size 15 Mb
Tram groups 3
Nodes number in each tram group 2
Tram speed 6.5 Km/h
Tram buffer size 59 Mb
Transmission speed of bluetooth 250 K
Transmission range of bluetooth 20 m
High speed interface transmission speed 10 Mb
High speed interface transmission range 1500 m
Time-to-live for message in each group 100 min
Message generation intervals 25–35 s
Size of message 500 Kb–1 Mb
Mobility model Shortage path map based movement model
Simulation time 100000 s
In this section, the performance of the proposed algorithm with respect to deliv-
ery probability, buffer time, hop count, overhead ratio and latency is evaluated
with similar existing algorithm.
Figure 2 compares the delivery probability of various algorithms when the
TTL is varied accordingly. It can be observed that delivery probability of algo-
rithms other than Epidemic and iPRoPHET decreases with the increase in TTL.
298 N. N. Srinidhi et al.
This is due to the fact the lifetime of messages increases with the increase in
TTL and hence the messages remain in the node buffer for longer time than
usual which results in more messages getting dropped. However, in iPRoPHET
algorithm, the delivery probability increases because of the selection of the next
forwarder using random forest classifier which considers PRoPHET probability
and other contextual information of the forwarders. The next forwarders are
selected not only when their PRoPHET probability is greater but also when the
probability is lower which helps to clear the messages from the buffer as quickly
as possible.
Figure 3 compares the Hop Count of various algorithms when the Number
of Nodes is varied accordingly. When the number of nodes is increased, it leads
to corresponding increase in number of forwarders available for routing the mes-
sages. Among these forwarders, the random forest classifier picks the fit nodes
accurately leading to lesser overall hop count of the proposed work.
messages for far longer time than usual and as number of nodes increases, the
outlier count increases keeping the average buffer time slightly larger and outlier
prone.
concluded that: (i) While making each routing decision, random forest classifies
the nodes as reliable or unreliable which led to better selection of intermediate
nodes to carry out efficient routing. (ii) Proposed method outperforms in terms
of delivery probability, hop count, overhead ratio and latency in comparison with
similar existing algorithms. (iii) The classifier trained with large amount of data
extracted using simulation leads to precise classification of the nodes as reliable
or unreliable nodes to provide efficient communication. However the proposed
method will have poor performance in terms of average buffer time in par with
similar multi-copy routing algorithm.
The average buffer time has experienced a performance issue due to the
presence of the outlier nodes, which needs to be addressed as part of future
enhancement. Furthermore, energy consumption of the nodes needs to be opti-
mized for working with compute-intensive machine learning algorithms in the
future work.
Acknowledgment. This research work has been funded by the Science and Engi-
neering Research Board (SERB-DST) Project File No: EEQ/2017/000681. Authors
sincerely thank SERB-DST for intellectual generosity and research support provided.
References
1. Srinidhi, N.N., Lakshmi, J., Dilip Kumar, S.M.: Hybrid energy efficient and QoS
aware algorithm to prolong IoT network lifetime. In: Kumar, N., Venkatesha
Prasad, R. (eds.) UBICNET 2019. LNICST, vol. 276, pp. 80–95. Springer, Cham
(2019). https://doi.org/10.1007/978-3-030-20615-4 6
2. Srinidhi, N.N., Dilip Kumar, S.M., Venugopal, K.R.: Network optimizations in the
internet of things: a review. Eng. Sci. Technol. Int. J. 22(1), 1–21 (2019)
3. Dilip Kumar, S.M., Vijaya Kumar, B.P.: Energy-aware multicast routing in manets:
a genetic algorithm approach. Int. J. (2009)
4. Guo, B., Zhang, D., Wang, Z., Zhou, X., Yu, Z.: Opportunistic IoT: exploring
the harmonious interaction between human and the internet of things. J. Netw.
Comput. Appl. 36, 1531–1539 (2013)
5. Pelusi, L., Passarella, A., Conti, M.: Opportunistic networking: data forwarding
in disconnected mobile ad hoc networks. IEEE Commun. Mag. 44(11), 134–141
(2006)
6. Vahdat, A., Becker, D.: Epidemic routing for partially connected ad hoc networks.
Technical report CS-2000-06 (2000)
7. Dhurandher, S.K., Sharma, D.K., Woungang, I., Bhati, S.: HBPR: history based
prediction for routing in infrastructure-less opportunistic networks. In: IEEE AINA
2013, Barcelona, Spain, 25–28 March 2013 (2013)
8. Lindgren, A., Doria, A., Schelen, O.: Probabilistic routing in intermittently con-
nected networks. ACM SIGMOBILE Mob. Comput. Commun. Rev. 7(3), 19–20
(2003)
9. Dhurandher, S.K., Borah, S.J., Obaidat, M.S., Sharma, D.K., Baruah, B., Gupta,
S.: Probability-based controlled flooding in opportunistic networks. In: Proceed-
ings of International Conference on Wireless Information Networks and System
(WINSYS), Colmar, France, 20–22 July 2015, pp. 3–8 (2015)
302 N. N. Srinidhi et al.
1 Introduction
In the present energy scenario, the demand for electricity is continuously increas-
ing and the global energy consumption will be tripled by the end of 2035 [1].
This work is a part of research project “Development of Intelligent Multipurpose con-
troller for Nano Grid Operation”, funded by Ministry of New & Renewable Energy
Resources, Government of India. Grant No: 24/29/2016-SWES (R&D).
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 303–317, 2020.
https://doi.org/10.1007/978-981-15-3666-3_25
304 G. M. Madhu et al.
2 Literature Review
2.1 Existing Smart Meters for Monitoring and Controlling
the Energy
Researchers have put efforts towards developing smart meters for monitoring
and controlling of individual consumers. Priyanka et al. [11] have presented an
IoT based smart meter for remotely monitoring of the renewable energy sources.
Here, FPGA based controller is used for computing and monitoring. However, the
overall system cost is very high and the controlling features are not implemented
yet. Aboelmaged et al. [12] have proposed an IoT based metering system which
uses Bluetooth technology to send the data over android application which in
turn sends the data to cloud. This approach has many flaws in terms of commu-
nication technologies and system design. Shibu et al. [13] have developed a smart
energy meter for monitoring and controlling of consumer loads. However, it lacks
in the detailed design and technologies used in the development. Lohokare et al.
[14] have presented an IoT ecosystem for wireless home automation systems.
The real-time implementation of this work to test its feasibility is not reported
yet. Barman et al. [15] have presented an IoT based smart meter which uses
open IoT platform. It is capable of uploading the consumption information at
IoT server. However, it has only one directional communication and the control-
ling part are not realized. Ali-Ali et al. [16] have presented an IoT based utility
meter which sends the monitored data to IoT server. Here, Raspberry Pi based
local server is created and the developed meter can monitor electricity, gas and
water consumption. Apart from these, Yaghmaee et al. [17] have developed an
IoT based smart meter which can monitor various electrical parameters of the
consumption, such as power factor, active power, apparent power and reactive
power. Here, an open IoT platform is used. However, it has high latency in the
data uploading. Giridhar et al. [18] have presented MQTT based industrial smart
grid monitoring system which monitors the individual load consumption. This
system lacks in controlling the loads. From the literature, we have observed that
there is a need of a system that can monitor as well as control the electricity
supplied to the consumers with the improved performance.
In literature, various IOT platforms have been investigated to collect sensor val-
ues from hardware and can have less latency. ThingSpeak [19] is an IOT analytic
platform which allows a user to send, aggregate, visualize and analyse the data
from hardware. It offers MATLAB integration support to perform analysis and
processing. It offers both free tire and commercial use where free tire allows up to
8 devices to connect in single channel, with 8500 messages/day limit and has a
delay time of 15 s for every data updating. Thing Board [20] is an open source
IOT platform which supports devices to connect with cloud and to perform
processing, visualization and device management. It supports multiple IOT pro-
tocols such as MQTT, CoAP, HTTP, etc. Other similar platforms are Kaa [21],
306 G. M. Madhu et al.
Fig. 1. Architecture of the proposed smart meter based monitoring and controlling
system (three layered architecture).
Thinger [22] and openIOT [23]. Carriots [24] is an end to end IoT platform which
offers platform as a service (PaaS) to design machine to cloud and machine to
machine projects. It uses application programming interfaces (APIs) for infor-
mation exchange. In contrast to ThingSpeak, it allows only 2 devices to monitor
in free tire. Cloud MQTT [25] has very less latency and can communicate with
HTTP, HTTPS, TCP and UDP protocols. This makes the developer to choose
a protocol accordingly to their design. Here, an MQTT server is implemented in
the local PC as it provides very fast communication and there are no limits on
the storage and number of devices connected.
is performed and values are stored in a database which can be accessed using
subscribe API. The design of the proposed smart meter is as follows:
Smart meters are important in proper operation of the proposed system. A design
of IoT enabled smart meter with tamper detection circuit is shown in Fig. 2. This
smart meter has the capability to directly interact with the IoT server and has
the feature to turn OFF/ON the supply to a consumer based on the commands
from the Server/Authority. The readings measured at every interval are sent to
the server through HTTP requests. A consumer can access these readings along
with billing information using the developed android application. A tampering
detection circuit continuously monitors for meter tampering. On detection of
the meter tampering, the smart meter automatically turns off the supply and
an alert message is sent to the concerned authority.
A work flow of the proposed smart meter is given in Fig. 3. Initially, the proposed
smart meter at individual consumer premises connects to available local network
in turn connecting to the server using SSID (Service Set Identifier) and API keys.
Then, the smart meter reads the power flow parameters and sends them to the
server. At server, these values are processed and stored in the database. Then,
the meter continuously checks the supply flag whose value is controlled by the
power authorities for providing the supply. Based on the supply flag, the smart
meter continuous or interrupts the power supply. In addition, the proposed smart
meter continuously checks for the meter tampering using PIR sensor values. If
there is any tampering detected, then the supply is turned OFF and the smart
meter changes the tampering flag to 1 at the server.
308 G. M. Madhu et al.
The implementation of the proposed system can be considered in two parts viz;
software design and hardware design. In software development, Message Queuing
Telemetry Transport (MQTT) based server is created and an android application
is developed. In hardware development, a low cost IoT enabled smart meter with
tampering detection is developed.
Parameter Value
Voltage range 0–300 V RMS
Current range 0–25 A RMS
Peak current 30 A
Sampling frequency 9615 Hz
Temperature 10–70 deg
Relay capacity 30 A RMS
Supply voltage 5V DC
Communication Wi-Fi 802.11 b/g/n
Fig. 9. The activities developed in Android. (a), (b), (c) - consumer apps. (d), (e), (f)
- authority apps.
Fig. 11. Monitoring of real time energy utilization information (a) from consumer app
(b) event log in smart meter.
consumption information with time stamp and the smart meter sending the data
to server.
Meter Tampering Detection Results: A scenario is created to tamper the
smart meter by making the PIR sensor to give motion detect output. The smart
meter reads the sensor value and sends the information to server detecting the
meter tampering event. Figure 12 shows the server-side data received from the
smart meter.
Fig. 13. Server sending a stop command to smart meter in the event of meter
tampering.
stop the supply and acknowledgement is received from the smart meter. Figure 13
shows the event log happened during this condition.
5 Conclusion
In this paper, we have proposed a open hardware-based IoT enabled real time
smart meter monitoring and controlling system. We have used a new architecture
based on MQTT protocol for developing the monitoring and controlling system.
Android apps are developed for easy monitoring of energy utilization/billing
information and alerting the tampering information for authorities. A prototype
of the proposed system is implemented by considering various test scenarios to
validate the proposed system. The experimental results are very encouraging to
make the proposed system widely utilized in real time environment. In future,
the proposed smart meter can be extended to measure power quality parameters
of individual consumer and to detect the energy theft at each smart meter level.
References
1. World energy outlook 2018. www.webstore.iea.org/world-energy-outlook-2018.
Accessed 15 Oct 2019
2. Ouf, M.M., Issa, M.H.: Energy consumption analysis of school buildings in Mani-
toba, Canada. Int. J. Sustain. Built Environ. 6(3), 359–371 (2017)
3. IEEE Vision for Smart Grid Controls: 2030 and Beyond Reference Model. In IEEE
Vision for Smart Grid Control: 2030 and Beyond Reference Model, pp. 1–10 (2013)
4. Campbell, M., Watson, N.R., Miller, A.: Smart meters to monitor power quality
at consumer premises. In: EEA Conference & Exhibition (2015)
5. Bedi, G., Venayagamoorthy, G.K., Singh, R., Brooks, R.R., Wang, K.: Review of
internet of things (IoT) in electric power and energy systems. IEEE Internet Things
J. 5(2), 847–870 (2018)
An Open-Hardware Approach for IoT 317
6. Abubakar, I., Khalid, S.N., Mustafa, M.W., Shareef, H., Mustapha, M.: Application
of load monitoring in appliances’ energy management-a review. Renew. Sustain.
Energy Rev. 67, 235–245 (2017)
7. Morello, R., De Capua, C., Fulco, G., Mukhopadhyay, S.C.: A smart power meter
to monitor energy flow in smart grids: the role of advanced sensing and IoT in the
electric grid of the future. IEEE Sens. J. 17(23), 7828–7837 (2017)
8. Gan, S., Li, K., Wang, Y., Cameron, C.: IoT based energy consumption monitoring
plat-form for industrial processes. In: 2018 UKACC 12th International Conference
on Control (CONTROL), pp. 236–240 (2018)
9. Gunturi, S.K.S., Reddy, M.S.K.: IoT based domestic energy monitoring device. In:
2018 3rd International Conference for Convergence in Technology (I2CT), pp. 1–4
(2018)
10. Tom, R.J., Sankaranarayanan, S., Rodrigues, J.J.P.C.: Smart energy management
and demand reduction by consumers and utilities in an IoT-fog based power dis-
tribution system. IEEE Internet Things J. 6(5), 7386–7394 (2019)
11. Priyanka, B.D., Joshi, V.M.: IOT based smart grid to remotely monitor and control
renewable energy sources using VLSI. Int. J. Innov. Res. Sci. Eng. Technol. 6(11),
21601–21607 (2017)
12. Aboelmaged, M., Abdelghani, Y., Ghany, M.A.A.E.: Wireless IoT based metering
system for energy efficient smart cites. In: 2017 29th International Conference on
Microelectronics (ICM), pp. 1–4 (2017)
13. Shibu, N.B.S., Hanumanthiah, A., Sai Rohith, S., Yaswanth, C., Hemanth Krishna,
P., Pavan, J.V.S.: Development of IoT enabled smart energy meter with remote
load management. In: 2018 IEEE International Conference on Computational Intel-
ligence and Computing Research (ICCIC), pp. 1–4 (2018)
14. Lohokare, J., Dani, R., Rajurkar, A., Apte, A.: An IoT ecosystem for the imple-
mentation of scalable wireless home automation systems at smart city level. In:
TENCON 2017-2017, pp. 1503–1508 (2017)
15. Barman, B.K., Yadav, S.N., Kumar, S., Gope, S.: IOT based smart energy meter
for efficient energy utilization in smart grid. In: 2018 2nd International Conference
on Power, Energy and Environment: Towards Smart Technology (ICEPE), pp. 1–5
(2018)
16. Al-Ali, A.R., Landolsi, T., Hassan, M.H., Ezzeddine, M., Abdelsalam, M., Baseet,
M.: An IoT-based smart utility meter. In: 2018 2nd International Conference on
Smart Grid and Smart Cities (ICSGSC), pp. 80–83 (2018)
17. Yaghmaee, M.H., Hejazi, H.: Design and implementation of an internet of things
based smart energy metering. In: 2018 IEEE International Conference on Smart
Energy Grid Engineering (SEGE), pp. 191–194 (2018)
18. Giridhar, M., Kayalvizhi, S.: IOT based industrial smart grid monitoring using
MQTT protocol. Int. J. Pure Appl. Math. 118(22), 41–49 (2018)
19. ThingSpeak IoT platform. https://thingspeak.com. Accessed 15 Oct 2019
20. ThingBoard IoT platform. https://thingsboard.io. Accessed 15 Oct 2019
21. Kaa IoT platform. https://www.kaaproject.org. Accessed 15 Oct 2019
22. Thinger IoT platform. https://thinger.io/. Accessed 15 Oct 2019
23. Open IoT. http://www.openiot.eu. Accessed 15 Oct 2019
24. Carriots. https://www.altairsmartworks.com/index.php. Accessed 15 Oct 2019
25. Cloud MQTT. https://www.cloudmqtt.com/docs/index.html. Accessed 15 Oct
2019
Performance of NOMA-MUSA System
for 5G Using m-ZCZ Sequences
1 Introduction
There are growing demands for reliable high data rate in mobile communication
system. The multi user interference occur between various users by simultaneous
sharing of a common channel. MAI provides a limitation on the capacity and
BER of the non-orthogonal multiple access (NOMA) system when more number
of users are active. MAI occurs when multiple users are simultaneously accessing
the same channel. The MAI in NOMA system is because of the larger similar-
ity between the obtained signals of different users in the receiver side under
Rayleigh channel and near far effect. The characteristics and feature of 5G are
high data rate coverage in the edge of the cell, increase in base station capacity,
and user’s quality of experience, reduction in latency, very good energy and spec-
trum efficiency. It is also utilizing spectrum from the available LTE frequency
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 318–328, 2020.
https://doi.org/10.1007/978-981-15-3666-3_26
Performance of NOMA-MUSA System for 5G Using m-ZCZ Sequences 319
band and millimeter frequency bands which supports data rate up to 20 gigabits
per second [5]. Multiple access is an important issue in the 5G communication.
This happens due to the higher cross correlation between the transmitted sig-
nals from different users in 5G and near far effect. NOMA is promising access
technology in 5G to make efficient use of spectrum resources [10]. NOMA’s main
objective is to improve spectrum sharing between different users by exploiting
power domain multiplexing which differs from traditional multiple orthogonal
access OMA [3,6,7,13].
NOMA has capability of handling different users at the equal time, frequency
and code but with varying power levels, which achieves a substantial spectral
efficiency OMA. Performance of NOMA better than OMA and attains better
spectrum utilization [8]. Due to increase of interconnection of devices through
IoT there is a need transferring data from one device to another such as video
conferencing, broadcast services, lifeline communications and E-health applica-
tions. Successive interference cancellation (SIC) with NOMA is achieved these
requirements in 5Gwireless system [4]. Pattern division multiple access (PDMA),
sparse code multiple access (SCMA), and multiuser shared access (MUSA) are
three main schemes of NOMA technologies. Here the performance of MUSA
system is investigated for 5G communication. Wang et al. [11] maximal length
sequence is used due to low cross correlation. Hayashi et al. [2] proved that
zero correlation zone sequences are good candidate for giving best orthogonal-
ity since it has zero auto correlation and cross correlation properties. In this
paper the BER performance of NOMA-MUSA using m-ZCZ are analysed and
compared. This paper is explained as follows: Sect. 2 deals with MUSA, Sect. 3
briefs about other code domain NOMA schemes, Sect. 4 explains about m-ZCZ
sequences, Sect. 5 deals with the proposed NOMA-MUSA system model with m-
ZCZ sequences, result and discussion are summarized in Sect. 6 and conclusion
is narrated in Sect. 7.
2 MUSA
MUSA is the code domain scheme of NOMA [13]. MUSA is utilizing the low
correlation spreading sequences at transmitting end by considering overloading
factor into account [11]. In MUSA the transmitted signal is spreaded by specially
designed sequence. Then these signals are transmitted over the channel simulta-
neously with same frequency. A non-linear type of MUD scheme: SIC is utilized
at the receiver in which the users are successively decoded. In this method it
successively cancels the substantial users by re-encoding the decoded bits. SIC
improves the BER performance of all the users [4].
The same blocks are transmitted by multiple users in SCMA with different code-
books. With sparse codebooks, the collision between users is decreased, there-
fore SCMA able to recover the inter-user interference. Because of Sparsity the
receiver complexity is reduced and able to achieve better performance using
message passing algorithm (MPA).
3.2 PDMA
4 m-ZCZ Sequences
The auto correlation and cross correlation value of m-ZCZ sequence is zero.
Hence it can able to eliminate MAI [7], m-ZCZ sequences are superior to Gold
sequences, Kasami sequences and m-sequences whose correlative properties are
good but not optimal. The auto correlation value must be zero except for time
shift of zero and cross correlation value of all the sequences is zero for every time
shift. Since the m-ZCZ sequences satisfy the above requirements, the number of
users supported by m-ZCZ sequences is larger than the conventional binary ZCZ
codes [5].
Construction of m-ZCZ Sequences
The m-ZCZ sequence is represented as
KL represents the sequence length, J is the sequence size and Mmin is the min-
imum length of oneside ZCZ respectively.
The first sub-sequence of the ith sequence with the length of L is given by,
Where a0 refers m-sequence with its period L, and ai.Mmin refers the sequence
generated by cyclically leftward shifted a0 with i.Mmin chips. This is given as
follows,
ai.M
n
min
= a0 (n + i.Mmin )modL (3)
where n = 0, 1, . . . L − 1 and n is the index of the chip. The second sub-
sequence Zi2 = +1 contains only one “+1 ” chip. The congregated code length is
Performance of NOMA-MUSA System for 5G Using m-ZCZ Sequences 321
L
KL = L + 1. The set size of the sequence is J = where x is the max-
Mmin
imum integer not larger than x.
for KL = 16, Mmin = 4 and the
Ye et al. [12] m-ZCZ sequence have developed
15
set of m-ZCZ sequences containing J = = 3 sequences are defined as,
4
Figures 1 and 2 show the ACF and CCF comparisons between m-sequences
and m-ZCZ sequences, respectively. ACF and CCF of ZCZ sequences are almost
zero when compared to m-sequences [1]. From Figs. 1 and 2, it is noted that
m-ZCZ sequences have improved cross correlation than that of PN-sequences.
The spreaded sequences are transmitted through Rayleigh fading channels. Then
it is processed by the SIC receiver. It is then fed into the decision making device
and they are converted into bit sequence. The sequence is despread and the
output data is obtained.
The transmitted signal in proposed NOMA-MUSA system is expressed as
follows,
Nu
∞
S k (t) =
(k)
ak q(t − jTs − jTf − Cj Tc ) (4)
k=1 j=−∞
where τk ∈ [0, Ts ] is the transmission delay of k th user and n(t) is the noise due to
Rayleigh fading. The proposed NOMA-MUSA system is implemented with SIC
receiver at the receiving end. To eliminate the error propagation, optimal order-
ing technique is used. While SIC is combined with optimal ordering technique,
Performance of NOMA-MUSA System for 5G Using m-ZCZ Sequences 323
knowledge in choosing the signal of s1 first or s2 is high and then subtracts the
respective s1 or s2 from the obtained signal. In this scheme, the strongest signal
is eliminated first followed by the second strongest, and so forth.
⎧
Nu (j+1)Tf (k) (k)
⎪
⎨ j=0 jTf y(t)hi (t − Cj Tc )dt,
(k) (k)
Yi = s + IM AI + n, for i = bi (6)
⎪
⎩ (k)
IM AI + n for i = bi
where s = ak and n is the Rayleigh component and IM AI denotes the MAI
component caused by Nu − 1 interfering users.
Using Matlab R2018a, Monte Carlo simulations have been performed to simulate
the performances of m-sequences and m-ZCZ sequences for different values of
Eb
( ) and varying number of users under AWGN channel.
N0
BER has been obtained by evaluating the transmitted signal with the received
signal and computing the error count for the proposed system. Simulation param-
eters of the proposed system are given in Table 1.
Output Parameter
BER is the output parameter of the proposed system. It is the ratio of the
number of bits in error to the total number of bits transmitted. The BER is
calculated as follows,
N umber of bits in error
BER = (7)
T otal number of bits transmitted
Fig. 4. BER of the proposed NOMA-MUSA system with m-ZCZ sequences as a func-
tion of in AWGN channel (constant power level)
In random power level in which all the users’ power are random values. With
random power level for the users, the MAI of every user is not the same.
Figure 5 depicts the BER performance of NOMA-MUSA system using m-
Eb
ZCZ sequences with respect to in AWGN channel for random power level.
N0
The BER performance of a single user with a random power level is not as
good as the constant power level. The proposed NOMA-MUSA system using m-
Eb
ZCZ sequences shows the BER values of 2 × 10−2 at = 8 dB and 2 × 10−3 at
N0
Eb
= 10 dB for 10 users under random power level (from 0.1 to 1mW). For same
N0
number of users SCMA and PDMA gives the BER of 1.6 × 10−1 and 3 × 10−1
Eb
at = 8 dB, SCMA and PDMA gives the BER of 1 × 10−1 and 2.5 × 10−1 .
N0
From the results, it is proved that the proposed NOMA-MUSA system exhibits
good BER when compared with SCMA-NOMA and PDMA-NOMA system.
326 M. Jayasheela et al.
Eb
From Table 2 a BER of 0.08 is obtained for 12 dB at 50 users. At = 8 dB,
N0
Eb
BER is increased to 0.1370 at the similar number of users. At = 4 dB BER
N0
value of 0.2810 is obtained for the identical number of users. Hence it is inferred
Eb Eb
that the less BER is obtained for = 12 dB. Thus, as value increases,
N0 N0
BER value decreases.
7 Conclusion
A novel NOMA-MUSA system with m-ZCZ is proposed and BER performance
of the system m-ZCZ sequences under AWGN channel is compared with other
NOMA schemes. The proposed NOMA-MUSA system using m-ZCZ sequences
exhibits 3 times BER improvement than that of SCMA-NOMA system and 40
Eb
times PDMA-NOMA at = 8 dB for 10 users under AWGN channel. From
N0
the results, it is also observed that proposed NOMA-MUSA system using m-ZCZ
yields better BER performance than that of other NOMA schemes.
References
1. Bae, J.N., Choi, Y., Kim, J.Y., Kwon, J.W., Kim, D.I.: Efficient interference can-
cellation scheme for wireless body area networks. J. Commun. Netw. 13, 167–174
(2011)
2. Hayashi, T., et al.: A novel class of quadriphase zero-correlation zone sequence sets.
IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 100(4), 953–960 (2017)
3. He, B., et al.: On the design of secure non-orthogonal multiple access systems.
IEEE J. Sel. Areas Commun. 35(10), 2196–2206 (2017)
4. Higuchi, K., Benjebbour, A.: Non-orthogonal multiple access (NOMA) with suc-
cessive interference cancellation for future radio access. IEICE Trans. Commun.
98(3), 403–414 (2015)
5. Li, S., Xu, L.D., Zhao, S.: 5G internet of things: a survey. J. Ind. Inf. Integr. 10,
1–9 (2018)
6. Liu, Y., et al.: Enhancing the physical layer security of non-orthogonal multiple
access in large-scale networks. IEEE Trans. Wirel. Commun. 16(3), 1656–1672
(2017)
7. Liu, Y., Qin, Z., Elkashlan, M., Nallanathan, A., McCann, A.: Non-orthogonal
multiple access in large-scale heterogenous networks. IEEE J. Sel. Areas Commun.
35, 2667–2680 (2017)
8. Lv, L., et al.: Cognitive non-orthogonal multiple access with cooperative relaying:
a new wireless frontier for 5G spectrum sharing. IEEE Commun. Mag. 56(4), 188–
195 (2018)
9. Ping, L.: Code-shift division multiple access for cellular mobile. In: 1997 IEEE 47th
Vehicular Technology Conference. Technology in Motion, vol. 1. IEEE (1997)
10. Qin, Z., et al.: Physical layer security for 5G non-orthogonal multiple access in
large-scale networks. In: 2016 IEEE International Conference on Communications
(ICC). IEEE (2016)
328 M. Jayasheela et al.
11. Wang, B., et al.: Comparison study of non-orthogonal multiple access schemes for
5G. In: 2015 IEEE International Symposium on Broadband Multimedia Systems
and Broadcasting. IEEE (2015)
12. Ye, G., et al.: A novel ZCZ code based on m-sequences and its applications in
CDMA systems. IEEE Commun. Lett. 11(6), 465–467 (2007)
13. Zhang, Z., Sun, H., Hu, R.Q.: Downlink and uplink non-orthogonal multiple access
in a dense wireless network. IEEE J. Sel. Areas Commun. 35(12), 2771–2784 (2017)
IoT Based Multi-sensor Data Acquisition
System for the Application
of Smart Agriculture
1 Introduction
The Internet of Things (IoT) is an enabling technology for the revolution in
health, agriculture, education, entertainment, and industry sector [1]. The tech-
nological developments are allowing every domain to employ new techniques
and methods to enhance their end results. One of them is the modern agri-
culture, employing new technologies and equipments to improve the yield. The
optimization of water usage with help of soil moisture sensor for the irrigation
process has been demonstrated [2]. The IoT and modern techniques transformed
conventional agriculture into smart agriculture. It helps in reducing the human
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 329–342, 2020.
https://doi.org/10.1007/978-981-15-3666-3_27
330 S. Kumar and N. Ch
work by monitoring the crop health and field parameters remotely. IoT utilizes
a remote sensor (mobile phone) that acts as the backbone for collecting infor-
mation for these monitoring and control applications. The system consists of
different sensors to monitor various parameters such as a temperature, humid-
ity, light intensity and soil moisture. The system is capable of communicating
this data with the other devices [3,4] for better control and monitoring. The
major challenges faced in the agricultural sector is data loss during the trans-
mission of data between multi-sensor system and mobile phone [5]. The data loss
will effect the controlling of irrigation system.
The limitation of the existing systems has motivated to propose an IoT based
multi-sensor data acquisition system for multiple activities including control,
monitor and manage different devices together for the application of smart agri-
culture. The system is built with an added feature to monitor the data and con-
trol the irrigation system in real-time without losing data during transmission.
The system is designed to measure different parameters such as a soil moisture,
temperature and relative humidity. Further, it helps to control irrigation process
remotely using a mobile phone. The developed system reduces the involvement
of human being as it makes cost-effective and time-efficient. In addition, the pro-
posed system uses solar energy to provide continuous operation which in turn
increases the autonomy of embedded system and its associated devices. Thus,
it helps farmers to contend with the enormous challenges faced in the agricul-
tural sector including overwatering in the field, unavailability of electric power
for irrigation, etc. The experimental results depicts that the accuracy of sending
control command for irrigation control is 100% and the accuracy of receiving
data from multiple sensor to mobile application is 98%.
This paper is organizes as follows. Section 2 briefly describe the data acquisi-
tion system. Section 3 discusses details about the Architecture of Data Acquisi-
tion System. Experimental results are shown in Sect. 4 and conclusions are given
in Sect. 5.
Transducer sense the physical phenomenon such as a soil moisture and con-
vert them into an electrical signal like voltage or current. However, these signals
may not be in the appropriate form for the measurement. The signals needs to be
modified appropriately as per the requirement in terms of amplitude (low/high)
and frequency. Further, the noise will be removed so that the signal will be
suitable for other circuit blocks.
Signal conditioning is one of the most important components of any data
acquisition system [6]. It is the interface between real-world analog signals and
the rest of the system. Signal conditioning system enhances the quality of signal
coming from a sensor in terms of:
1. Right type of signal, to convert the output signal from a transducer into the
desired form i.e. voltage/current.
2. Right level of the signal, to amplify or attenuate the signals to a right/
acceptable level for the further processing.
3. Noise, to eliminate noise from a signal.
4. Protection, to protect the damage to the next block of a system such as a
micro-controller from the high current or voltage signals.
The circuit diagram for measuring the temperature and humidity (DHT22) are
given in Fig. 3. Initially, micro-controller sends a start signal to the DHT22 to
change from standby status to running status. In return, the DHT22 send a
response signal of 40-bit data that contains temperature and relative humidity
back to the micro-controller unit (MCU). The DHT22 will not response to MCU
IoT Based Multi-sensor DAQ for the Application of Smart Agriculture 333
unless it receives the start signal. However, DHT22 will change from running
status to standby status when transmission has completed or when the start
signal is not received from the MCU for the next operation [5].
A resistor connected between pin1 and pin2 of DHT22 shown in Fig. 3 [5] is
used as a pull-up resistor. This resistor essentially pulls up the voltage at pin2
to 5V in times where the DHT sensor is not actually sending any data (i.e. not
HIGH or LOW). Therefore, we use pull-up resistor to prevent this condition
called floating.
The circuit for solar tracker using LDR is shown in Fig. 5. The minimum
output voltage out of all LDRs are used to rotate the solar panel through servo
IoT Based Multi-sensor DAQ for the Application of Smart Agriculture 335
motor. The servo motor will move the solar panel in the direction where light
intensity is more. Hence, the solar panel effectively charges the battery connected
with it so as to operate the DAQ system. The curve shown in Fig. 6 tells that
the LDR voltage is inversely proportional to the light intensity.
4.5
3.5
3
LDR Voltage (V)
2.5
1.5
0.5
0
0 100 200 300 400 500 600 700 800 900 1000
Illumination (LUX)
Fig. 8. Firebase
sensors are updated in mobile application which further helps to control the
irrigation system.
Whenever control command is given by the user for controlling irrigation
system through mobile application and the same has been gets updated in the
firebase cloud at a particular node in real-time. The firebase cloud communicates
with Wi-Fi NodeMCU and it follows up based on data received from firebase
IoT Based Multi-sensor DAQ for the Application of Smart Agriculture 339
cloud. If NodeMCU receives serial data 1, it sends the signal to the Arduino
micro-controller to turn ON the irrigation system through 5 volt relay, and if
NodeMCU receives serial input 0 then it command the Arduino micro-controller
to turn OFF the irrigation system.
4 Experimental Results
The proposed system has been designed and demonstrated for controlling and
monitoring the irrigation system through a mobile application for the optimum
usage of resources. The Fig. 12 shows an experimental setup for proposed DAQ
system. The optimum usage of water as per the moisture/humidity levels can be
controlled through a mobile application. The system is designed such way that
it can automatically turn on/off the irrigation system through a relay. Further,
the system can be controlled remotely via internet through a mobile application.
Besides that, the power supply to turn on/off the irrigation system is provided
by a battery which is charged by solar tracker.
340 S. Kumar and N. Ch
In this experiment, the soil and environmental information has been collected
from the sensors and it is forward to mobile application so as to control the
irrigation system. These experiments were carried out by sending and receiving
control commands via the internet from the mobile application. A unique control
code has been utilized at firebase cloud to receive the control command from the
mobile application and the same command is forwarded to NodeMCU. Further,
the NodeMCU forwards the command to ATmega328 to control the 5volt relay
which helps in switching the irrigation system.
The experimental results in Table 1 shows that the 49 times out of 50 times
the data transmitted has been received at the end node. It confirms the per-
centage of successful data transmission is 98%. In Table 2, the control command
sent from mobile phone to the end node has accuracy of 100%. It means that
the measuring environmental parameters and controlling the irrigation system
are successfully demonstrated without fail which fulfils the requirements of the
proposed system for the said application.
5 Conclusion
IoT based Multi-Sensor Data Acquisition System was designed and demon-
strated successfully for the application of smart agriculture. As per the require-
ment of the user, it provides the flexibility of controlling irrigation system in
two ways, one is manual control and another is automatic control. The system
automatically updates the data collected from different sensors in real time and
at the same time makes it available to the user through a mobile application.
As a result, it allows the user to maintain soil moisture and other parameters
within an allowable range to yield better crop. This developed system reduces
the involvement of human which makes it cost-effective and time-efficient. Fur-
ther, the proposed system was designed to reduce the data loss close to zero,
and found that the data loss in case of control signal transmission is 0% and in
case of data transmission it was 2%. In addition, the system was integrated with
solar tracker to provide a solar energy for the continuous operation.
Acknowledgment. The design and implementation was carried out at IIIT Senapati
Manipur with the support of TEQIP-III, NPIU and Government of India. Finally,
authors would like to thank Dept. of ECE as well as others who have helped in the
demonstration of DAQ system for the said practical application.
References
1. Al-Fuqaha, A., Guizani, M., Mohammadi, M., Aledhari, M., Ayyash, M.: Inter-
net of things: a survey on enabling technologies, protocols, and applications. IEEE
Commun. Surv. Tutor. 17(4), 2347–2376 (2015)
2. Kim, Y., Evans, R.G., Iversen, W.M.: Remote sensing and control of an irrigation
system using a distributed wireless sensor network. IEEE Trans. Instrum. Meas.
57(7), 1379–1387 (2008)
3. Liu, Y., Zhang, C., Zhu, P.: The temperature humidity monitoring system of soil
based on wireless sensor networks. In: 2011 International Conference on Electric
Information and Control Engineering, pp. 1850–1853. IEEE (2011)
4. Ivanov, S., Bhargava, K., Donnelly, W.: Precision farming: sensor analytics. IEEE
Intell. Syst. 30(4), 76–80 (2015)
5. Kaewmard, N., Saiyod, S.: Sensor data collection and irrigation control on vegetable
crop using smart phone and wireless sensor networks for smart farm. In: 2014 IEEE
Conference on Wireless Sensors (ICWiSE), pp. 106–112. IEEE (2014)
342 S. Kumar and N. Ch
6. Preejith, S., Alex, A., Joseph, J., Sivaprakasam, M.: Design, development and clin-
ical validation of a wrist-based optical heart rate monitor. In: IEEE International
Symposium on Medical Measurements and Applications (MeMeA), pp. 1–6. IEEE
2016 (2016)
7. Hari Ram, V.V., Vishal, H., Dhanalakshmi, S., Vidya, P.M.: Regulation of water in
agriculture field using internet of things. In: IEEE Technological Innovation in ICT
for Agriculture and Rural Development (TIAR), pp. 112–115. IEEE (2015)
8. Zhao, J.-C., Zhang, J.-F., Feng, Y., Guo, J.-X.: The study and application of the
IoT technology in agriculture. In: 2010 3rd International Conference on Computer
Science and Information Technology, vol. 2, pp. 462–465. IEEE (2010)
Context-Oriented User-Centric Search
System for the IoT Based
on Fuzzy Clustering
1 Introduction
The infrastructure of Internet of Things (IoT) is undergoing an expeditious devel-
opment, wherein several physical components (such as sensors, actuators, etc.,)
inclusive of the user are being linked to the digital world. These components
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 343–356, 2020.
https://doi.org/10.1007/978-981-15-3666-3_28
344 S. Pattar et al.
2 Literature Survey
In this section, we present an overview of the recent works that address the
sensor search challenge in the IoT. In the first subsection, we discuss the previous
studies related to the context-based search systems for IoT and in the second
subsection, we illustrate the recent works that apply fuzzy theory to the IoT
search applications. We put forth how our work differs from the existing works
as our contributions in this field.
Context-Oriented Fuzzy Based Search System for the IoT 345
Few recent works on fuzzy-based IoT search systems are discussed in this sub-
section. Based on the data generated by the sensors, Truong et al. [19] put forth
a content-based search approach built through fuzzy theory. Although this work
achieves low communication overhead, it does not provide user-centric results.
Similarly, Truong et al. [20] in one of their other works have proposed an algo-
rithm for computing similarity index of the sensors deployed in Web of Things
(WoT) using fuzzy rules. This approach is highly accurate in computing sim-
ilarity parameters between sensors of the same region but failed to compute
similarity among sensors between two different regions. Jun et al. [21] presented
a fuzzy logic-oriented information retrieval mechanism in IoT integrated Intel-
ligent Transport System (ITS) application. However, this mechanism handles
only simple fuzzy relations that are to be defined beforehand in the ontology.
In Table 1, we compare the different search techniques proposed for the IoT
systems.
From the above works, it can be inferred that the context-based search tech-
niques developed by several authors could not handle user requirements specified
through natural language or linguistic variables. Also, the existing fuzzy-based
search algorithms lack the applicability of context parameter requirement of the
user and the sensor attributes. In this regards, we have devised a search system
based on context-oriented weighted fuzzy c-means clustering algorithm to reduce
346 S. Pattar et al.
the sensor search space in the IoT and thereby decrease the communication over-
head for efficient query resolution. Our approach is also capable of handling the
queries given by the user through linguistic variables.
In this section, we outline the sensor search problem considering user queries
and sensor properties. In the first subsection, we formulate the problem math-
ematically along with an example. We discuss the importance of fuzzy logic
applicability for IoT search systems in the second subsection.
sensors (such as sensor type, data output, geographical location, life-time, accu-
racy, error rate, bandwidth, reliability, etc). These properties are used as param-
eters in the search process to select an appropriate set of resultant sensors. They
are classified into two types namely, content-properties and context-properties.
The former set of properties consists of the parameters of the data emitted by
the sensors (represented by the set O), while the latter set describes the situ-
ation based properties (context) of the sensors (represented by the set C). It
is necessary for us to find out an optimum subset P ⊆ S such that it satisfies
the conditions specified by the user. Some of the attributes of set A change over
time. Thus, a sensor s ∈ S can be represented as,
s: T → F (1)
where, T represents the set of time intervals within which sensor reading are
extracted and F is the set of attribute instances of a sensor, where F ⊆ A i.e.,
the sensor readings can either belong to output-data set (O) or context-based
attribute set (C). An IoT application has to capture user requirements in the
form of a query in order to perform search and select operation on the sensor
set S. We use query Q to represent the user-requirements, that in-turn specifies
the sensor properties x ∈ A required by the user along with the relationships
among them. This query is used to discover the “optimal subset” of pn sensors
that match the user-requirements. Thus the query Q, can be formulated by the
function as follows,
Q : T × X × pn → F (2)
where, the query poses a request to locate pn sensors from S, such that they
have x ∈ A properties at time-stamp t ∈ T . F ⊆ S defines the subset of “optimal
sensors” obtained after the query has been executed.
Example. Let us consider an IoT application designed for an ITS system with
n sensors. The objective of this application is to offer search service to the user
and help him/her in locating appropriate sensors for their use. Thus, S repre-
sents the set of sensors deployed in the ITS environment. Let the attribute set
defined for the sensor set S be A = {reading − data, latitude, longitude, battery −
lif e, reliability, time − stamp}. For a given set of four temperature sensors, let us
assume that set O = {{(30deg, 1.00 pm), (36deg, 1.02 pm)}, {(40deg, 1.05 pm)},
{(42deg, 1.10 pm)}} consists of temperature measured by the sensors at a partic-
ular time interval. Similarly, C contains such readings for two context-properties,
battery-life and reliability. A user query on such a search space can take form
like, To retrieve sensors that measure temperature (sensor type context-property)
between the time interval 1.00 pm to 1.15 pm (content property). Thus, a query
can have constraints on both the set of content as well as context-properties.
arises due to the uncertainty associated with the problems, making it difficult
to arrive at an accurate solution. These challenges pave the way for fuzzy logic-
based approaches. Fuzzy logic is an effective technique that handles the ambi-
guity present in complex problems, thereby providing the best possible solution.
Unlike binary logic that accepts only either of the two values 0 or 1 (true or
false), fuzzy logic is multi-valued. It deals with approximation rather than pre-
cision. Fuzzy logic also provides a medium to represent linguistic variables (in
natural language) that in-turn express human’s reasoning capabilities [8].
Let us take an example from an IoT search requirement scenario to better
understand the importance of fuzzy logic and how its association with context
parameters uplifts the competence of the search system. Two users Ravi and
Ranjan intend to acquire road traffic congestion data captured by the sensors
in a certain city, say Bengaluru, for different purposes. Ravi prefers to make
use of data collected from highly accurate sensors (say, 95% of accuracy) while
Ranjan is interested in using the information aggregated by the moderately
reliable sensors (say, 70% of reliability). Here, the terms ‘accurate/accuracy’,
‘reliable/reliability’ are referred to as “linguistic descriptor” that captures the
user requirements.
The existing search systems designed for the IoT require queries to be in
a crisp format and thus fails to provide a user-centric query resolution mecha-
nism. Fuzzy logic can be used to handle this challenge. In traditional crisp logic
methodology, the linguistic descriptor is expressed either as ‘accurate’ (value:
1) or ‘not accurate’ (value: 0), while in fuzzy systems, the linguistic variable is
expressed as ‘less-accurate’, ‘accurate’, ‘more-accurate’ and so on, where each
variable is associated with a truth value between 0 to 1 operating on the basis
of membership functions. If a sensor has an accuracy value of 0.9, then 0.9 is
called as the “degree of membership” of the sensor in a fuzzy set that represents
accuracy context-property.
In this work, we design a fuzzy logic-based search mechanism for the IoT sys-
tem. Initially, properties of the sensors are used to create clusters that hold sim-
ilar sensors (through our proposed semantic weight-based fuzzy c-means search-
space clustering algorithm). Then, a user query that expresses search constraints
in linguistic terms is mapped to one of the appropriate cluster (through the
selection of cluster head). Finally, a subset of sensors that matches the given
user requirements are selected and returned. This approach offers user-centric
results by considering the fuzzy representation of both sensor properties and
user constraints. Also, the overall communication overhead for query resolution
is effectively reduced due to the use of clustering approach.
In this section, we describe in detail the execution flow and construction of the
proposed sensor search model delineating the techniques used for query process-
ing, search space structuring and query resolution.
Context-Oriented Fuzzy Based Search System for the IoT 349
Phase - I Phase - II
Search
Result Search within Cluster Head
Cluster Selection
Phase - III
Our proposed sensor search scheme executes in three phases as depicted in Fig. 1.
The processes carried out in each phase is explained in the following.
the search space structuring and cluster selection process, thus resulting in an
efficient search performance.
the membership value to obtain the updated weighted cluster heads. Equation 5
is used for this purpose, where hi is the cluster head for ith cluster, m is the
fuzzification parameter, and sj is a particular sensor among n sensors.
n
sim(j, k) ∗ um ij
hi = n sj , j = 1, 2, . . . n. (5)
t=1,k=j sim(j, t) ∗ um
tj
k=1,k=j
In the third step, the membership values are updated using the below
equation.
−1/(m−1)
sj − hj
uij = z −1/(m−1)
(6)
i=1 sj − hi
The above steps are repeated until the difference between the membership
matrix of current and previous iterations are below a pre-defined tolerance value
(i.e., β).
352 S. Pattar et al.
5.1 Implementation
We have implemented our proposed prototype search technique in Java. The
SSN ontology required for sensor data modeling is handled by Apache Jena
Framwork (Version 3.12.0) [3] and SPARQL querying language [5]. We have also
made use of libraries like Commons Math from Apache [2] to calculate semantic
weights and euclidean distance. For our experimental purposes, we have used
environmental datasets MesoWest [4] and Air Quality Sensor Dataset [1] that
capture the air quality measurement reading recorded from sensors. As at present
there does not exists public dataset that houses context-properties of the sensors,
we have randomly generated ten context-properties for each of the sensors in the
above datasets. These properties include location, accuracy, type, battery life,
life-time, error rate, bandwidth, reliability, energy, and availability. To evaluate
and measure the performance of the proposed search system, we used a computer
with Intel(R) Core i5-4200M 2.50 GHz processor with 8 GB RAM. The dataset
constructed consists of a total 1,00,000 readings.
After, conducting several empirical tests we have set the optimal values of
the parameters as ζ = 0.7, z = 10, m = 1.75 and β = 0.5. In the following, we
discuss them: (i) The balancing factor (ζ) in semantic score calculation controls
the contribution of sensor attributes to the cluster formation, thus it decides
the quality of the clusters. In this work as we used ten context-properties which
mostly have three linguistic terms, the ontology thus has a tree with a maximum
depth of 10 and at most 30 nodes (concepts). Here, a factor of 0.7 covers all the
context-properties with these values of the ontology’s depth and the number of
nodes. (ii) As the number of context-properties in the dataset is limited to 10, the
number of clusters (z) at this value achieves better search results. (iii) The fuzzi-
fication parameter (m) controls the magnitude of membership value. A higher
value will magnify the addition of context-property, while a lower value decreases
the role of the context-property to clustering process. It is observed that a 1.75
value balances the membership matrix with all the context-properties. (iv) The
tolerance value (β) controls the number of iterations required for cluster forma-
tion. We noted that at 0.5 value of β, the difference in membership updation is
no longer decreasing and thus effectively stops the clustering process.
Processing Time. It is the time taken by the search system to resolve a given
query (i.e., search space structuring, query processing, and resolution). As seen
in Fig. 2, the proposed technique takes comparatively less time than AntClust.
This is attributed to the fact that AntClust involves a two-phase cluster for-
mation process where at first level sensors are grouped based on their type and
further at second level they are clustered based on their attributes. However,
in our approach, only one phase is involved where semantic-based clustering
technique groups the sensors with similar descriptive (linguistic) properties, we
thus eliminate the two-way cluster formation and search. Due to this, the overall
communication overhead in the search system is drastically reduced.
our clustering approach can easily identify the sensor’s affinity to a particular
cluster as all the matching context-properties of that cluster are considered.
6 Conclusions
Searching for relevant sensors among a large pool of connected devices in the IoT
has become a significant task with the creation of several prominent applications
across various sectors of the society. Thus, numerous techniques are proposed in
the past to address this problem, but they do not consider the user’s context
and also take query in rudimentary quantity values. In this paper, we have
devised a fuzzy-based context-oriented search algorithm that aids in forming
clusters through semantic weights. The submitted user query is directed to the
cluster that is most appropriate to the context of the query. Experimental results
demonstrate that the proposed search space structuring technique drastically
decrease the search processing time owing to the effective semantic-based cluster
head selection in comparison to the meta-heuristic approach. It also achieves
lower communication overhead through the one-way inter-cluster search. Also,
when compared with the existing work our method yields effective results and
thus provide a superior user-centric search system.
References
1. Air Quality Sensor Dataset. https://github.com/BathHacked/documentation/
wiki/Bath-Air-Quality-Sensor-Data. Accessed 20 Aug 2019
2. Apache Commons Math. https://commons.apache.org/proper/commons-math/.
Accessed 20 Aug 2019
3. Apache Jena. https://jena.apache.org/. Accessed 20 Aug 2019
4. MesoWest Dataset. http://mesowest.utah.edu. Accessed 20 Aug 2019
5. SPARQL Query Language. https://www.w3.org/TR/sparql11-query/. Accessed
20 Aug 2019
6. Bezdek, J.C., Ehrlich, R., Full, W.: FCM: the fuzzy c-means clustering algorithm.
Comput. Geosci. 10(2–3), 191–203 (1984)
7. Desai, P., Sheth, A., Anantharam, P.: Semantic gateway as a service architecture
for IoT interoperability. In: IEEE International Conference on Mobile Services, pp.
313–319. IEEE (2015)
8. Dilli, R., Argou, A., Pilla, M., Pernas, A.M., Reiser, R., Yamin, A.: Fuzzy logic and
MCDA in IoT resources classification. In: Proceedings of the 33rd Annual ACM
Symposium on Applied Computing, pp. 761–766. ACM (2018)
9. Ebrahimi, M., ShafieiBavani, E., Wong, R.K., Fong, S., Fiaidhi, J.: An adaptive
meta-heuristic search for the Internet of Things. Future Gener. Comput. Syst. 76,
486–494 (2017)
10. Hung, C.C., Kulkarni, S., Kuo, B.C.: A new weighted fuzzy c-means clustering
algorithm for remotely sensed image classification. IEEE J. Sel. Top. Sign. Proces.
5(3), 543–553 (2011)
11. Lefort, L., et al.: Semantic sensor network. XG final report (2011)
12. Pattar, S., Buyya, R., Venugopal, K.R., Iyengar, S.S., Patnaik, L.M.: Searching for
the IoT resources: fundamentals, requirements, comprehensive review, and future
directions. IEEE Commun. Surv. Tutorials 20(3), 2101–2132 (2018)
356 S. Pattar et al.
13. Pattar, S., et al.: Progressive search algorithm for service discovery in an IoT
ecosystem. In: 2019 International Conference on Internet of Things (iThings)
and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber,
Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData),
pp. 1041–1048 (2019)
14. Pattar, S., et al.: Ontology based service discovery for intelligent transport systems
using Internet of Things. In: Fourteenth International Conference on Information
Processing (ICInPro), pp. 223–225 (2018)
15. Perera, C., Vasilakos, A.V.: A knowledge-based resource discovery for Internet of
Things. Knowl. Based Syst. 109, 122–136 (2016)
16. Perera, C., Zaslavsky, A., Christen, P., Georgakopoulos, D.: Context aware com-
puting for the Internet of Things: a survey. IEEE Commun. Surv. Tutorials 16(1),
414–454 (2014)
17. Perera, C., Zaslavsky, A., Liu, C.H., Compton, M., Christen, P., Georgakopoulos,
D.: Sensor search techniques for sensing as a service architecture for the Internet
of Things. IEEE Sens. J. 14(2), 406–420 (2014)
18. Roopa, M., Pattar, S., Buyya, R., Venugopal, K.R., Iyengar, S., Patnaik, L.: Social
Internet of Things (SIoT): foundations, thrust areas, systematic review and future
directions. Comput. Commun. 139, 32–57 (2019)
19. Truong, C., Römer, K.: Content-based sensor search for the Web of Things. In:
Proceedings of the GLOBECOM - IEEE Global Communications Conference, pp.
2654–2660 (2013)
20. Truong, C., Römer, K., Chen, K.: Fuzzy-based sensor search in the Web of Things.
In: Proceedings of International Conference on the Internet of Things, (IoT), pp.
127–134 (2012)
21. Zhai, J., Liang, Y., Yu, Y., Jiang, J.: Semantic information retrieval based on fuzzy
ontology for electronic commerce. JSW 3(9), 20–27 (2008)
22. Zhang, C.: FuzWare: a fuzzy-based middleware for context-aware service. In: IEEE
2nd Advanced Information Technology Conference (IAEAC) on Electronic and
Automation Control, pp. 1181–1185 (2017)
Implementation of a Smart Embedded
System for Passenger Vessel Safety
1 Introduction
Even though one of the most commonly used means of transport in Bangladesh
is the passenger’s vessel, the rate of fatal accidents has increased at an alarming
rate according to the department of shipping. In 38 years (1977 to 2013), the
number of death totaled to 4420, 520 people were injured and 400 people are still
missing. Overloading being the leading cause accounts for the maximum share
of 56% of the accidents in water routes. In order to earn maximum profit, oper-
ators overload their vessels with passengers and cargos. Very common scenery is
a passenger vessel loaded with cargos in the upper deck as well as letting passen-
gers in the upper deck. As a result, the top weight increases making the vessel
unstable. In most cases, vessels met accidents due to being 4/5 times overloaded
than its capacity. Sometimes passenger vessel operators load passengers from
riverbanks and waterways in order to earn some extra money making the vessel
overweighted. To detect and prevent accidents due to overweight, an IoT based
embedded system has been developed called Safety measuring model for a pas-
senger’s vessel. The prototype of this embedded system sends the weight status
and current location of the vessel to the server via the internet. If the vessel at
any point detects overweighting, the signal is sent to the server which allows the
automated embedded system to let the coast guards know of the vessel’s condi-
tion and take action accordingly. In this system, two types of sensors are used.
First is the float switch sensor to get the reading of the weight of the vessel in
the water. Another is a GPS tracker to trace the current location of the vessel. A
microcontroller receives the raw data and sends it to a Raspberry pi through the
internet from further processing. The Raspberry pi analyzes the received data
to make a decision by identifying the weight status of the vessel. Based on that
decision the coast guard is notified if any action is to be taken. With this embed-
ded system, vessel overweight incident can be prevented and countless lives can
be spared. The rest of the paper discusses the literature review in Sect. 2, the
proposed system methodology, the mathematical exploration consists of Sect. 3,
in Sect. 4 discussed each component, in Sect. 5 broadly discussed the embedded
system with graphical visualization and implementation set.
2 Literature Review
In recent years, a number of research papers were published showing the types
of vessels, vessel maintenance, and architecture and production costs. Yet very
few kinds of research are done to ensure safety, measuring the weight of vessels
to avoid accidents caused by overweight.
The primary reason for each accident study is to give decision-makers sub-
stantial and solid data so as to settle on educated and ideally better choices. In
this paper, mishap data for passenger vessels gathered from Bangladesh Inland
Water Transport Authority (BIWTA) is introduced and an investigation is com-
pleted to decide the normal reasons for accidents on passenger vessels [1].
The purpose of this paper is to structure the applied model concerning an
advancement overloaded sensor that can be utilized to distinguish and reduce the
overburden issue of passenger ships. To beat this issue, an inventive innovation
joins the lift ideas utilizing the load sensor (HCC-High Capacity Compression)
and utilize the batching controller to set the base and most extreme limits are
prescribed to be installed at the entrance dot of the passenger’s vessel [2].
In this paper, it has been seen that the regular reasons for accidents are
impacts and contacts, fire and blasts, overburdening, harsh climate, and so on.
At long last, in light of this investigation, various proposals have been proposed
so as to guarantee more prominent security of passengers [3].
The exploration disclosed that the significant reasons for waterway accidents
are impact, Nor’wester or tempest, overloading, dependability disappointment,
Implementation of a Smart Embedded System for Passenger Vessel Safety 359
unnecessary current, base harm. The investigation demonstrates that cargo ves-
sels and passenger’s vessels are confronting more accidents contrasted with dif-
ferent kinds of vessels. It was likewise seen that numerous mishaps and appli-
cable data frequently stay unreported. Besides, the recorded waterway accident
information is likewise very short data for leading standard research works. At
last, a few suggestions are advanced with a vision to develop a more secure
waterway transportation system for Bangladesh and for further research and
investigations [4].
The general point is to acquire an expanded comprehension of measurements
in maritime security culture, its impact on wellbeing the executives and its
connection to the International Safety Management Code. This paper presents
results from two security culture studies directed on two sorts of vessels. The
outcomes indicated contrasts among officials and team in how wellbeing society
measurements were reported [5].
Addressing the research gap in the researches done previously on passenger’s
vessels, an embedded system is proposed with the approach to make sure the
wellbeing of the passenger’s vessel. The embedded system is developed in such a
way that it will provide the overweight status and location of the vessel to help
prevent any accidents before it happens. The developed embedded system will
help prevent vessels from loading more than its capacity thus avoiding accidents
due to overweight and save lives every day.
3 Proposed System
In the proposed embedded system, the first concern is to detect or take the read-
ing of the value of the current water level. This collected data will be analyzed.
A breakpoint needs to be determined to measure if the water level or the weight
of the vessel is exceeding beyond the expected level. To determine the breaking
point, the weight of the vessel is to be calculated. It all comes down to a rule
called “buoyancy” and can be clarified by what is known as “Archimedes’ Prin-
ciple.” What this rule fundamentally states is that a floating object uproots a
measure of fluid equivalent in weight to the heaviness of the object. This implies
the measure of water uprooted by a boat, regardless of how full that vessel is,
weighs as much as the boat. The weight is identified with the volume of water
displaced by,
This is the place the dimensions of the boat become significant. The volume
of water uprooted by the boat is equivalent to the volume involved by the piece
of the boat that is underneath the surfaces of the water. When the volume of
dislodged water is more prominent than the all total out volume of the boat, it
implies that the highest point of the boat will never again be above the water
and the boat will sink [6].
360 F. M. J. M. Shamrat et al.
We can locate the absolute upward force on the can be applied by water
pressure (which we call the buoyant force Fbuoyant ) by basically taking the dis-
tinction between the magnitudes of the upward force Fup and downward force
Fdown ,
Fbuoyant = Fup − Fdown (1)
“Archimedes Principle Formula” is recognized as [7],
F = ρgV (2)
Where, Buoyant force of a given body = F,
Volume of the displaced fluid = v,
acceleration due to gravity = g.
As we see that density ρρ = ρf − ρg. Here ρf is the density of fluid and ρg
is the density of the body. Accordingly, the principle can also be expressed as,
F = (ρf − ρg)gV (3)
Archimedes theory equation is useful in finding the buoyant force, the volume
of the dislodged body, the density of fluid or density of body if a portion of these
numeric values are known or firstly need to find out these numeric values to
implement Archimedes formula.
In Figs. 1 and 2 we have shown the flowchart diagram and process of the
proposed model of the embedded system that we have actually done in the
embedded system.
To build up the mentioned embedded system, water level raw data are gathered
by water float switching sensors. The sensors do the job of unmistakable sorts of
critical merit. The sensor collects the data by applying the algorithm and sends
it to the Raspberry Pi server (Node-Red), to make the decisions. The sensor will
detect the current water level like max, min and normal.
4.1 Raspberry pi B
like Scratch and Python. It contains 1 GHz single-focus CPU, 512 MB RAM,
Mini HDMI port, Micro USB OTG port, Micro USB control, HAT-impeccable
40-stick header, Composite video and reset headers, CSI camera connector [8].
In Fig. 3 we showed the image of Raspberry pi zero.
Fig. 3. Raspberry pi B.
The switch can measure or monitor the water level by moving up and down in
its row. It’s basically made only for counting water level data. In Fig. 5 we have
shown the floating switch [10].
Give Ultimate GPS V3 a boost with this outer dynamic antenna. This GPS
radio wire draws about 10mA and will give an extra 28 dB of increase. It has a
5-m long cable so it will effortlessly reach any place that needs it to. The antenna
is magnetic so it will adhere to the highest point of a vehicle or truck. Used to
detect the location of the specific object [11]. In Fig. 6 we have shown the GPS
antenna.
To plan the programming fragment of the proposed model, every one of the
duties is summarized and put in a progression of procedures with the goal that
one can interpret them in programming code. An algorithmic methodology for
the arrangement of activities is offered below (Fig. 8):
6 Implementation
The proposed procedure has been designed and developed in default IDE and
furthermore can develop Anaconda application. A passenger’s vessel is selected
for the implementation and test this embedded system, where max capacity
is defined 1500 kg by the help of the “Archimedes” formula. There 35 passen-
gers were available in the vessel. They all are of different weights, collected the
weight data in a variety of ways since the passenger arrives. Data is collected
by the water floating switch sensor and location share is implying with the GPS
antenna. In excess of a couple of experiments have been observed to test the
algorithm of the proposed system. Evaluate these outcomes of the experiments
with each other. The outcome of the introduced system is bringing into being in
a predictable range.
6.1 Experiment
An overall 22 experiments are done for the measurement of the performance
accuracy of the embedded system. By installing the embedded system, read the
Implementation of a Smart Embedded System for Passenger Vessel Safety 367
weight of the passenger’s vessel that is displayed in Table 1 and the relating
graphical portrayals are exhibited as graphical visualize (1, 2) and are shown in
Figs. 9 and 10.
Using the estimation of the table recognizes by the sensors and the output
gets by using our calculation. In Figs. 9 and 10 graphical visualize of the infor-
mation that is collected by sensors and output has been exhibited.
In several experiments of the embedded system, water floating sensors switch
are installed at the breakpoint of the vessel (see Fig. 7). It reads the water level
if the vessel is overweighed than the water level of the vessel will increase. We
get weight raw data by utilizing sensors from a vessel and save it in Table 1
and deploy our proposed Algorithms 1 and 2 and on it. We get the result of
overweight issues when the weight overcomes the max capacity (weight ≥ 1500)
of the vessel, the embedded system shows the overweight positive, either negative
(Table 1). It only passes data overweight “yes” when vessel weight exceeds the
368 F. M. J. M. Shamrat et al.
max weight. We have shown the graphical visualize of Table 1 data in Figs. 9 and
10. Sensors will send the data request and decision of the system to the server
(Node-RED) through the internet, raspberry pi B will read the data from the
server and visualize the data and location of the vessel to the output monitor.
Implementation of a Smart Embedded System for Passenger Vessel Safety 369
The monitoring man will notify the coast guard to stop the vessel from safe
thousands of lives.
7 Conclusion
The automated embedded systems in the passenger’s vessels are very urgent
at the present time for the safety of both the passengers and the vessel due
to the overloading issue. With the help of some methodology, the issue can
be dealt with. The framework that is implemented in the system is created,
structured and actualized using the python programming language. To examine
the methodology and to determine an agreeable level, a number of investigations
were performed. Implementation of this framework can help thousands of lives,
families, and dreams every year from meeting a fatal end.
References
1. Rahman, S.: An analysis of passenger vessel accidents in Bangladesh. In: 10th
International Conference on Marine Technology, MARTEC (2016)
2. Abdul Rahman, N.S.F., Rosli, H.Z.: An innovation approach for improving pas-
senger vessels safety level: overload problem. Int. J. Bus. Tourism Appl. Sci. 2(2),
61–74 (2014)
3. Awal, Z.I., Rafiqul Islam, M., Hoque, M.: Analysis of passenger vessel accidents in
the inland waterways of Bangladesh. In: Session I-B-1 Analysis of Passenger Vessel
Accidents in the Inland Waterways of Bangladesh (2006)
4. Rashid, K., Islam, M.R.: Reasons and remedies of inland passenger vessels accidents
in Bangladesh. In: AIP Conference Proceedings, vol. 1919, p. 020036 (2018)
5. Ek, Å.: A study of safety culture in passenger shipping. In: The 3rd Safety and
Reliability International Conference, vol. 3, pp. 99–106 (2003). The Publishing and
Printing House of Air Force Institute of Technology
6. Archimedes law and buoyant force information. https://www.khanacademy.org/
science/physics/fluids/buoyant-force-and-archimedes-principle/a/buoyant-force-
and-archimedes-principle-article. Accessed 26 Apr 2019
7. Vessel floating and displaced water information. http://scienceline.ucsb.edu/
getkey.php?key=620. Accessed 26 Apr 2019
8. Raspberry pi zero. https://learn.adafruit.com/introducing-the-raspberry-pi-zero?
view=all. Accessed 26 Apr 2019
9. Microcontroller. https://learn.adafruit.com/programming-microcontrollers-using-
openocdon-raspberry-pi?view=all. Accessed 26 Apr 2019
10. Water floating switch sensor. https://www.quora.com/How-does-a-ship-float-on-
water. Accessed 26 Apr 2019
11. GPS antenna. https://learn.adafruit.com/adafruit-ultimategps/external-antenna.
Accessed 26 Apr 2019
12. Gypa, I., Boulougouris, E., Vassalos, D.: Decision support tool for modelling secu-
rity scenarios onboard passenger ships. Taylor & Francis Group, London (2018).
ISBN 978-0-8153-7993-5
13. Reddy, M., Rao, K.: Fire accident detection and prevention monitoring system
using wireless sensor network enabled Android application. Indian J. Sci. Technol.
9(17), 1–5 (2016)
370 F. M. J. M. Shamrat et al.
14. Shams, R., Hossain, S., Priyom, S., Fatema, N., Shakil, S.R., Rhaman, M.K.: An
automated firefighting system. In: 2015 12th International Conference on Fuzzy
Systems and Knowledge Discovery, FSKD 2015, pp. 2327–2331 (2015)
15. Chen, J.J., Lu, M.X., Wu, T.H., Hwang, Y.S.: Sub-1-V FastResponse hysteresis-
controlled CMOS buck converter using adaptive ramp techniques. IEEE Trans.
Very Large Scale Integr. (VLSI) Syst. 9(21), 1608–1618 (2013)
16. Cheng, Y.F., Hirdaris, S.E.: Improvement of ship safety through stability research
and innovations. In: 11th International Conference on the Stability of Ships and
Ocean Vehicles (2012)
17. Ahola, M., Romanoff, J., Kujala, P., Remes, H., Varsta, P.: Cruise and Ferry Expe-
rience program for future passenger ship designers. In: RINA Conference on Edu-
cation and Professional Development of Engineers, Newcastle, UK (2011)
18. Bouguila, N., Daoudi, K.: A statistical approach for binary vectors modeling and
clustering. In: Theeramunkong, T., Kijsirikul, B., Cercone, N., Ho, T.-B. (eds.)
PAKDD 2009. LNCS (LNAI), vol. 5476, pp. 184–195. Springer, Heidelberg (2009).
https://doi.org/10.1007/978-3-642-01307-2 19
Extended Track
Empirical Analysis on the Effect of Image
Compression and Denoising Using
Different Wavelets on Iris Recognition
1 Introduction
2 Background
2.1 Related Work
2.2 Preliminaries
Wavelet Compression. Wavelet Compression The images are compressed and
compared among three wavelet-based techniques, namely Embedded Zero Tree
Wavelets (EZW), Set Partitioning in Hierarchical Tree (SPIHT) and Spatial-
Orientation Tree Wavelet (STW). For all of the three compression techniques,
the Haar wavelet has been used for decomposition in which the image is filtered
into several sub-bands. The image being of dimension 280 × 320, it was zero-
padded to form 512 × 512 images so that decomposition can be made possible.
Empirical Analysis on the Effect of Image Compression and Denoising 377
Where Sn (t) denotes set of coordinates. i, j are coordinates and Cij is the
coefficient at that coordinate. The list of significant and insignificant pixels
can be used to eliminate redundant data and achieve compression.
– Spatial-Orientation Tree Wavelet. The STW is alike to EZW, only the
difference is how a tree is formed. STW follows the state transition model. For
one threshold value to the following, the position of transform values keeps
on changing from one state to another. The approach used in STW differs in
encrypting the information related to Zero Tree. The model makes easier in
reducing the number of bits required in encryption. It is observed that the
tree structure has more spatial scanning and does not follow a zigzag scanning
as seen in EZW. In Fig. 4 we can observe the STW tree and in Fig. 5 a STW
scanning is shown which is different from EZW.
378 P. Baro et al.
By using the FK filters the quality metrics value of the water-marked image,
i.e., NC (normalized cross-correlation) Q (quality of the image) is computed
like the other wavelet filter. The idea behind having several wavelets is that
the decomposition will be more accurate when the signal (image) looks sim-
ilar to the wavelet. Thus the image and noise are well decomposed for some
wavelets compared to others. The image is decomposed using a 2D stationary
wavelet (SWT) decomposition technique as depicted below in Fig. 6.
In this, the image is filtered row-wise using the wavelet into the low pass and
high pass signals each of which requires further filtration into high pass and
low pass in column-wise. This decomposes the image into four sub-images each
specializing in: approximation, horizontal edges, vertical edges, and diagonal
edges. Additionally, downsampling is performed following each filtration without
loss of information according to NY Quist criteria.
Fig. 6. SWT
Dataset. In this work, the CASIA v1.0 dataset has been used. It is publicly
Dataset. This dataset is available at http://biometrics.idealtest.org/. The images
have been captured using a proprietary Iris camera by the National Labora-
tory of Pattern Recognition (NLPR), Institute of Automation (IA) and Chinese
Academy of Sciences (CAS). There were 108 eyes which consisted of 756 iris
images. For an eye 7 number of images were captured in two sessions, in the first
session 3 images were captured and the remaining 4 were captured in the second
session.
380 P. Baro et al.
3 Proposed Work
The flow diagram of the procedure adopted is shown in Fig. 7. These two flows
can be observed in the given diagram. Both flows have in common: the acquir-
ing of the dataset, application of compression and denoising techniques on the
dataset, recognition of iris and performance metric computation. However, one
flow concentrates on the impact of wavelet compression and others concentrate
on wavelet denoising. The images generally consist of redundant pixels and also
some noises, so, to remove redundancy and disturbance in an image, compression
and denoising using various methods has been used. The used dataset is CASIA
v1.0. The detailed explanation of the flow diagram is as follows:
At first we will read the gray scale images and convert it into discrete pixel
values. For the lossy image compression we have applied EZW, SPIHT and
STW techniques. We have categorized the output into four different ratios. It
comprises of 10:1, 15:1, 20:1, 30:1 compression ratios. For the recognition purpose
we have applied the Libor Masek which is an open-source code. And next we have
computed the values for different parameters like EER, FAR, FRR and Quality
measurement. For Denoising the gray scale images are read and converted into
discrete pixel values. The images are decomposed by 2-D stationary wavelet
using Daubichies, Haar, Bior 1.3 and Fk4 wavelets. Here, we have taken four
different thresholds 6, 8, 10, 12 and calculated the performance metrics as done
in the compression.
Dilation Measure. The DM is the ratio of radius of the Hough circle drawn
around the pupil and that of the iris.
Rp
DM = (6)
Ri
Where Rp denotes the radius of pupil and Ri denotes radius of iris in pixels.
QM = F IM ∗ OM ∗ DM (7)
4 Results
4.1 Compression
Performance Curves
It is observed clearly that accuracy is lightly degraded up to the targeted com-
pression ratio. Higher Compression Ratios will degrade the image qualities and
recognition in relation of PSNR and Quality Metrics. The EER varies with dif-
ferent techniques. The quality of some images may reduce but some may not be.
For the set FAR there is a noticeable change in FRR. The STW with CR of 30:1
performs as good as the original image which can be observed in Figs. 8, 9 and
10. STW performs well for PSNR as well as Quality Metrics and also the EER
value is lower in STW. EZW gives good result at increasing compression ratio.
SPIHT performs as close as that of original image at a lower compression ratio.
4.2 Denoising
The results of denoising applying the different wavelets with variable thresholds
(6, 8, 10 and 12) are shown in Figs. 11, 12 and 13. Due to increase in thresh-
old, smoother images are obtained and fk4 wavelets perform higher smoothing
(Tables 4, 5 and 6).
It is seen that the accuracy slightly changes due to denoising. The FAR
remains unchanged but the FRR values changes with different denoising thresh-
olds. With an increase in threshold, the curves of EER move towards right ini-
tially and after the threshold of 8 towards the left. This indicates the improve-
ment of recognition at higher thresholds. This could be due to more elimination
384 P. Baro et al.
Fig. 11. Threshold vs Fig. 12. Threshold vs QM Fig. 13. Threshold vs EER
PSNR
of noise at higher thresholds. The graph in Fig. 13 shows that bior1.3 gives the
minimum EER value. The metrics show that an increase in threshold leads to
a fall in PSNR indicating loss of signal strength. In Fig. 11 we can see that the
PSNR value is higher at lower threshold value. Haar, db1 and bior1.3 acts sim-
ilarly with an increase in threshold value. Different techniques show different
trends in quality. For db1 and Haar wavelets, it gives the similar curve. bior1.3
gives the best result in terms of achieving the better quality at thresholds of 12
respectively.
5 Conclusion
The impact of wavelet compression and denoising on iris images were studied
separately. The analysis is done on the basis of different compression ratios and
performance is observed on the set compression ratios, thresholds. The wavelet
Empirical Analysis on the Effect of Image Compression and Denoising 385
based techniques could be useful for compressing the iris images without much
loss of recognition performance. STW performed well at higher compression rate.
It is observed that the wavelet compression techniques and wavelet denoising
techniques has not deteriorated the recognition performance. As an outcome,
huge dataset can be compressed to a small size and the matching is affected
minimally. The denoising using various wavelets and hard thresholds indicate
that both haar and db1 wavelets perform similarly. In terms of PSNR, QM and
EER we have seen that bior 1.3 gives the best output compared with other
wavelets. And in future we are trying to extend our work on colored image
compression along with denoising and see the impact on Iris Recognition.
References
1. Trokielewicz, M., Czajka, A., Maciejewicz, P.: Iris recognition after death. IEEE
Trans. Inf. Forensics Secur. 14(6), 1501–1514 (2019)
2. Gupta, R., Sehgal, P.: Non-deterministic approach to allay replay attack on iris
biometric. Pattern Anal. Appl. 22(2), 717–729 (2019)
3. Hamd, M.H., Ahmed, S.K.: Biometric system design for iris recognition using intel-
ligent algorithms. Int. J. Mod. Educ. Comput. Sci. 11(3), 9 (2018)
4. Shen, J.J., Yeh, C.H., Jan, J.K.: A new approach of lossy image compression based
on hybrid image resizing techniques. Int. Arab J. Inf. Technol. 16(2), 226–235
(2019)
5. Strela, V., Heller, P.N., Strang, G., Topiwala, P., Heil, C.: The application of
multiwavelet filterbanks to image processing. IEEE Trans. Image Process. 8(4),
548–563 (1999)
6. Paul, A., Khan, T.Z., Podder, P., Ahmed, R., Rahman, M.M., Khan, M.H.: Iris
image compression using wavelet transform coding. In: 2nd International Confer-
ence on Signal Processing and Integrated Networks (SPIN), pp. 544–548. IEEE
(2015)
7. Rakshit, S., Monro, D.M.: An evaluation of image sampling and compression for
human iris recognition. IEEE Trans. Inf. Forensics Secur. 2(3), 605–612 (2017)
8. Zemliachenko, A., Kozhemiakin, R., Vozel, B., Lukin, V.: Prediction of compres-
sion ratio in lossy compression of noisy images. In: 13th International Conference
on Modern Problems of Radio Engineering, Telecommunications and Computer
Science (TCSET), pp. 693–697. IEEE (2016)
9. Goyal, B., Dogra, A., Agrawal, S., Sohi, B.S.: Two-dimensional gray scale image
denoising via morphological operations in NSST domain & bitonic filtering. Future
Gener. Comput. Syst. 82, 158–175 (2018)
10. Funk, W., Arnold, M., Busch, C., Munde, A.: Evaluation of image compression
algorithms for fingerprint and face recognition systems. In: Proceedings from the
Sixth Annual IEEE SMC Information Assurance Workshop, pp. 72–78. IEEE
(2005)
11. Hedaoo, P., Godbole, S.S.: Wavelet thresholding approach for image denoising. Int.
J. Netw. Secur. Appl. (IJNSA) 3(4), 16–21 (2011)
386 P. Baro et al.
12. Dehkordi, A.B., Abu-Bakar, S.A.: Noise reduction in iris recognition using mul-
tiple thresholding. In: 2013 IEEE International Conference on Signal and Image
Processing Applications, pp. 140–144. IEEE (2013)
13. Rodriguez, N., Barba, L.: Fejer-Korovkin wavelet based MIMO model for multi-
step-ahead forecasting of monthly fishes catches. Polibits 56, 71–76 (2017)
14. Daugman, J., Downing, C.: Effect of severe image compression on iris recognition
performance. IEEE Trans. Inf. Forensics Secur. 3(1), 52–61 (2008)
15. Shapiro, J.M.: Embedded image coding using zerotrees of wavelet coefficients. IEEE
Trans. Signal Process. 41(12), 3445–3462 (1993)
16. Ives, R.W., Bishop, D.A., Du, Y., Belcher, C.: Iris recognition: the consequences
of image compression. EURASIP J. Adv. Signal Process. 2010(1), 680845 (2010)
17. Ives, R.W., Broussard, R.P., Kennell, L.R., Soldan, D.L.: Effects of image compres-
sion on iris recognition system performance. J. Electron. Imaging 17(1), 011015
(2008)
18. Varanis, M., Pederiva, R.: The influence of the wavelet filter in the parameters
extraction for signal classification: an experimental study. Proc. Ser. Braz. Soc.
Comput. Appl. Math. 5(1) (2017)
19. Ives, R.W., Bishop, D.A., Du, Y., Belcher, C.: Effects of image compression
on iris recognition performance and image quality. In: 2009 IEEE Workshop on
Computational Intelligence in Biometrics: Theory, Algorithms, and Applications,
pp. 16–21. IEEE (2009)
20. Mishra, K.N.: An efficient technique for online iris image compression and per-
sonal identification. In: Tiwari, B., Tiwari, V., Das, K.C., Mishra, D.K., Bansal,
J.C. (eds.) Proceedings of International Conference on Recent Advancement on
Computer and Communication. LNNS, vol. 34, pp. 335–343. Springer, Singapore
(2018). https://doi.org/10.1007/978-981-10-8198-9 35
21. Rai, H.M., Chatterjee, K.: Hybrid adaptive algorithm based on wavelet transform
and independent component analysis for denoising of MRI images. Measurement
144, 72–82 (2019)
22. Dua, M., Gupta, R., Khari, M., Crespo, R.G.: Biometric iris recognition using
radial basis function neural network. Soft Comput. 23(22), 11801–11815 (2019)
Segmentation of Chart Images Using
H-Means Algorithm
1 Introduction
With the growing advancements in computer vision, clustering techniques
applied on datasets depending on their characteristics and features have
improved even more and gained importance too [9]. Clustering algorithms have
been a vital means scaled to various application areas apart from computer
vision, i.e., data mining, statistical analysis, image processing, quantization,
and video processing [8]. Image segmentation has to be efficient in categoriz-
ing objects in images from the background. Therefore, many methods for image
segmentation have already been proposed. Clustering is one of the common and
broadly used segmentation approach because of its simplicity and efficiency.
There are different clustering algorithms like K-means, fuzzy c-means, spectral
clustering, expectation, and maximization [9]. Variants of clustering algorithms
were proposed having different perspectives which were designed for different
purposes [10]. However, existing clustering algorithms requires user-specified
parameters as input.
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 387–397, 2020.
https://doi.org/10.1007/978-981-15-3666-3_31
388 P. Mishra et al.
2 Literature Survey
algorithm which improved the way of initializing centre value. Proposed method
gives a novel way of determining the centroids without increasing the complexity
of algorithm. Although no other parameter like threshold value is used, but ini-
tializes the k i.e. the number of clusters. Similar work was suggested by [11] who
proposed an enhanced algorithm for better accuracy and efficiency by taking dif-
ferent centroid values and determining its data points. [1] used histogram before
segmentation to enhance the region of interest by equalizing RGB channel sep-
arately. [6] suggested a mechanism for image segmentation using K means and
morphological processing resulting into efficient segmentation accuracy. While
[12] gave segmentation technique by merging similar regions. As per the avail-
able literature it was notices that k means is considered to be the simplest
algorithm giving efficiently accurate results of segmentation.
3 Proposed Work
We proposed a method for finding the correct and dynamic value of K using peak
histogram values and segment the image using MATLAB functions. Input chart
image is converted into greyscale, of which histogram is obtained. The number
of highest peaks from the histogram is taken as K, i.e., number of centroids.
Image is segmented by using K means clustering algorithm, details of which are
shown in Fig. 1.
Every color image has tri band i.e., red, green and blue channel based his-
togram. Thus for a chart images having different color styles, it is difficult to
determine and discriminate the data from the histogram which can be used for K
clusters. Thus, color images are converted into greyscale. We proposed a method
for finding the correct and dynamic value of K using peak histogram values and
segment the image using MATLAB functions. Input chart image is converted
into greyscale, of which histogram is obtained. The number of highest peaks
from the histogram is taken as K, i.e., number of centroids. Image is segmented
by using K means clustering algorithm, details of which are shown in Fig. 1.
390 P. Mishra et al.
Fig. 2. (a) Original image, (b) Histogram of RGB Channel, (c) Histogram of Input
Image
Every color image has tri band i.e., red, green and blue channel based his-
togram. Thus for a chart images having different color styles, it is difficult to
determine and discriminate the data from the histogram which can be used for
K clusters. Thus, color images are converted into greyscale. After that, we can
easily find that there are some “peaks” on the histogram, which are the local
maxima of the greylevel intensity frequency. The number of peaks will be used
to decide how many K-values will be used in the image segmentation process.
Because the peaks mean that the pixel count with those tone values was high,
which means there are many pixels that have the same RGB values refer to the
original color image (Fig. 2).
likeness of the images. It measures the quality of the output image. A smaller
value of PSNR means that the output image has poor quality
⎡ ⎤
⎢ max(r(x, y))2 ⎥
P SN R = 10. log1 0 ⎢
⎣ nx −1 ny −1
⎥
⎦ (2)
1 0 0 [(r(x,y))]2
nx ny nx −1 ny −1 [(r(x,y))−(t(x,y))]2
0 0
The number of K-value can be determined correctly but not randomly general-
ized. The number of clusters will be suitable for any images for the segmentation.
From Table 2, it can be observed that the proposed method has the small values
of MSE and high values of PSNR.
The values of the RMSE and PSNR measures indicate that the higher value
of RMSE provides good significant about visual data of chart images. On the
other hand, the smaller value of PSNR measure depicted that the output images
are not up to significant enough to describe the visual content of complex images
of chart classes. Therefore, it can be concluded that the output image resulted
from the proposed method are of good quality.
Figures 3 and 4 shows the comparison of RMSE and PSNR values generated
for tested images. From these figures, it can be seen that resulted images are of
good quality in comparison to that of standard K means algorithm. Figure 5
shows the sample of tested images on the proposed method. The proposed
method is tested on a bar chart, area chart, stacked donut chart, pie chart,
line chart, scatter chart, and donut chart as shown in Fig. 6. Figure 6A is an
Segmentation of Chart Images Using H-Means Algorithm 393
original image, while Fig. 6B shows the segmented area of the targeted image
and Fig. 6C depicts the binary representation of segmented images. It can be
seen that the proposed method has finely segmented the required regions of
interest. Every small sub-region of the target image is precisely segmented from
the background. It is visible from area chart, pie chart and stacked donut chart
that every sub-area is properly segmented from other corresponding areas. How-
ever, the proposed systems generated desired results, but in the case of the pie
chart, it lacked in segmenting low-intensity areas from other sub-areas of chart
images.
Fig. 6. Results obtained of chart images. (A) Original input (B) Segmentation result
(C) Binary form of segmentation result
Image Total peaks Highest 25% Highest 50% Highest 75% Highest 50%
peaks selected peaks selected peaks selected peaks selected
Image 1 54 14 27 41 54
Image 2 101 26 51 76 101
Image 3 130 33 65 98 130
Image 4 254 64 127 191 254
5 Conclusion
A new image segmentation method is proposed based on the histogram and k-
means. Method uses histogram to determine peak value which is used as value
of K for segmentation. Apart from this, method proves to be robust towards
various types of chart images, separating the required RoI from complete chart
image. In this way every piece of chart is segmented proving to be effective
for better chart interpretation and analysis. Evaluation parameters RMSE and
PSNR shows that the method shows the better result accuracy in comparison
to standard algorithms.
References
1. Dhanachandra, N., Manglem, K., Chanu, Y.J.: Image segmentation using K-means
clustering algorithm and subtractive clustering algorithm. Procedia Comput. Sci.
54, 764–771 (2015)
2. Ishak, A.B.: A two-dimensional multilevel thresholding method for image segmen-
tation. Appl. Soft Comput. 52, 306–322 (2017)
3. Yuheng, S., Hao, Y.: Image segmentation algorithms overview. arXiv preprint
arXiv:1707.02051 (2017)
4. Bora, D.J.: Importance of image enhancement techniques in color image segmen-
tation: a comprehensive and comparative study. arXiv preprint arXiv:1708.05081
(2017)
5. Purohit, P., Joshi, R.: An efficient approach towards k-means clustering algorithm.
Int. J. Comput. Sci. Commun. Netw. 4(3), 12 (2013)
6. Zheng, X., Lei, Q., Yao, R., Gong, Y., Yin, Q.: Image segmentation based on
adaptive K-means algorithm. EURASIP J. Image Video Process. 1, 68 (2018)
7. Yedla, M., Pathakota, S.R., Srinivasa, T.M.: Enhancing K-means clustering algo-
rithm with improved initial center. Int. J. Comput. Sci. Inf. Technol. 1(2), 121–125
(2010)
8. Gao, J., Zhou, Y., Barner, K.E.: View: visual information extraction widget for
improving chart images accessibility. In: 2012 19th IEEE International Conference
on Image Processing, Florida, USA, pp. 2865–2868. IEEE (2012)
9. De, P.: Automatic data extraction from 2D and 3D pie chart images. In: 2018 8th
IEEE International Advance Computing Conference (IACC), India, pp. 20–25.
IEEE (2018)
Segmentation of Chart Images Using H-Means Algorithm 397
10. Prasad, V.S.N., Siddiquie, B., Golbeck, J., Davis, L.S.: Classifying computer gener-
ated charts. In: 2007 IEEE International Workshop on Content-Based Multimedia
Indexing, France, pp. 85–92. IEEE (2007)
11. Nazeer, K.A., Sebastian, M.P.: Improving the accuracy and efficiency of the k-
means clustering algorithm. In: Proceedings of the World Congress on Engineering,
pp. 1–3. Association of Engineers, London (2009)
12. Sharma, A., Sharma, V.K., Srivastava, D.K.: Region merging based segmentation
approach for color image using maximal similarity index. In: 2nd International
Conference on Trends in Electronics and Informatics (ICOEI), India, pp. 403–408.
IEEE (2018)
Hand Vein Biometric Recognition Using
Local Binary Pattern
1 Introduction
Biometric becomes popular among researcher since last two decades due to its
different applications in common life of people. Biometrics help to improve the
applications of different technology with reliability, identification and secure
access of users. Human biometric traits includes palmprints, facial images, hand
vein images, signature, fingerprint, retinal information etc. The distinctive infor-
mation in biometric images uses as a feature vector to identification verification.
Dorsal Hand vein identification has come-up with important biometrics traits.
It becomes a very important and effective biometrics trait. Vein structure is an
example of intrinsic biometric. It is more accurate than extrinsic as the vein
pattern is a bank of blood vessels situated directly beneath the person’s hand.
Study of this biometric feature has been started since the past two decades [1,2].
Many researchers had been contributed towards an accurate recognition system
or optimizing the matching time.
Eigen veins have been used to represents vein patterns and reduces its dimen-
sion of vein pattern through principal component analysis (PCA) [1]. FRR and
FAR have been reduced and matching of features is done by Euclidean distance.
Finger webs have been used to find the region of interest (ROI) and feature points
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 398–406, 2020.
https://doi.org/10.1007/978-981-15-3666-3_32
Hand Vein Biometric Recognition Using Local Binary Pattern 399
The present work has been encompasses the use of dorsal hand vein in person
indentification. The proposed approach has been based on Local binary pattern
of input image. The orientation of local binary pattern has been used as a feature.
Hamming distance has been used for calculation of difference of the distances
between the images.
2 Proposed Approach
The process of the dorsal hand vein biometric system consists of three major
steps, namely preprocessing of vein images, feature extraction and verification of
features. These steps of proposed approach is discussed in detail in the following
paragraphs.
2.1 Preprocessing
To identify the unique feature of vein extraction, hand vein images are applied
to preprocessing steps for changing image into acceptable formats. Input image
is taken as shown in Fig. 1.
Local binary pattern has been used to enhance the detection rate on different
types of data sets. It works as visual descriptor which used for classification
purpose. The LBP feature vector has been created by applying the proposed
approach.
The selected window has been divided into different cells. The every pixel
of cell is compared to each of its 8 neighbours. The neighbours of each pixels
of a cell is pixels of its left-top, left-bottom, left-middle, left top, right-middle,
right-top etc. The each pixels followed along circle. It may be clock wise or anti-
clockwise. The numeric value of each pixels has been replaced by “0” or “1”. If
the value of center pixel is greater than the neighbour’s pixel value then write
“0” otherwise “1”. 8-digit binary number comes as a results of this comparison.
It can be expressed as follow (Fig. 3).
Bitwise matching is performed using hamming distance between the image given
by the user and image present in the dataset. It creates a large gap between these
two. Normalized hamming distance can be found as follows: If the calculated
distance is below a given hamming distance threshold, it is considered as correct
matching. And if the calculated distance is above a given hamming distance
threshold, it is considered as incorrect matching (Figs. 4 and 5).
402 K. K. Singh et al.
3 Experiment
The proposed method has been validated on a standard dataset taken from the
Department of Biomedical and Systems Engineering, Cairo University. The data
set contains 600 images collected from 50 people. All simulations were done using
MATLAB 8.1.0.604 (R2017a).
Equal Error Rate (ERR): The acceptance and rejection errors are of equal
error rate is ERR. ROC curve is used to find the value of the EER. Correct
Recognition Rate (CRR): It is defined as the proportion of the number of sam-
ples that were correctly classified to the whole number of test samples. Results
of proposed approach have been verified in two different ways. The identification
and verification both have been conducted for different test samples. The effec-
tiveness of identification part of part of proposed approach has been calculated
in terms of correct recognition rate.
If it is obtained from the same person, then recognition has been considered
correct. The false acceptance rate (FAR) and false rejection rate (FRR) has
been used to calculate successness in verification mode of proposed approach.
FAR and FRR values have been used to plot the curve of Receiver Operating
Characteristics (ROC). ROC states graph between FAR and FRR for all feasi-
ble hamming distance thresholds and are generally used to calculate the Equal
Error Rate. The proposed method have been used 600 vein images from the
database for calculating the CRR. Out of 600 vein images 586 images have been
classified correctly and obtained 97.66% of the accuracy. Different hamming dis-
tance threshold values have been selected for matching of vein pattern to get
the false rejection rate and false acceptance rate. The corresponding threshold
vs. False Acceptance Rate (FAR) and threshold vs. False Rejection Rate (FRR)
404 K. K. Singh et al.
graphs has been plotted in Figs. 7 and 8 respectively. The resulting Equal Error
Rate plot has been given in Fig. 9. Figure 7 shows that when we increases the
threshold value the false acceptance rate is reducing. It becomes minimum when
values of threshold is 99. From Figs. 7 and 8, it has been concluded that at a
Hamming distance threshold of 0.017 the best FAR and FRR values are 0.045
and 0.023 respectively. In Fig. 9, ERR obtained is 0.011. From the results we can
say that the proposed approach gives very less Equal Error Rate compared with
other algorithms for detection of vein. Table 1 shows the comparative study of
different algorithms for detection of hand vein. The values in the table clearly
shows that results obtained by using Principal Component Analysis (PCA), EP
Tree algorithm, Convolution Neural Network method and independent compo-
nent analysis. The quantitative parameters have been calculated for all these
approaches. The different number of users data set has been used for calculation
of ERR, FRR, FAR and CRR.
Hand Vein Biometric Recognition Using Local Binary Pattern 405
5 Conclusions
The proposed algorithm developed a biometric technique for dorsal vein recog-
nition. The results clearly shown that the proposed approach can adequately
represent dorsal vein features and a high accuracy of 97.66 matching then some
classification of dorsal vein patterns can be added.
References
1. Subramanian, R.K.: Low dimensional representation of dorsal hand vein features
using principle component analysis (PCA). World Acad. Sci. Eng. Technol. Int. J.
Comput. Electr. Autom. Control Inf. Eng. 3, 1 (2009)
2. Lin, C.-L., Fan, K.-C.: Biometric verification using thermal images of palm dorsal
vein patterns. IEEE Trans. Circ. Syst. Video Technol. 14(2), 199–213 (2004)
3. Lee, J.-C., Lee, C.-H., Hsu, C.-B., Kuei, P.-Y., Chang, K.-C.: Dorsal hand vein
recognition based on 2D Gabor filters. Imag. Sci. J. 62(3), 127–138 (2013)
4. Li, X., Huang, D., Zhang, R., Wang, Y., Xie, X.: Dorsal hand vein recognition
based on Width Skeleton Models, State Key Laboratory of Software Development
Environment, Beihang University, Beijing (2016)
5. Nadiya, K., Gopi, V.P.: Dorsal hand vein recognition based on Orientation of
Local Binary Pattern, 1077 - Computational Intelligence for Multimodal Biomed-
ical Imaging (2015)
6. Wan, H., Song, H., Chen, L., Yang, J.: Dorsal hand vein recognition based on
Convolutional Neural Network (CNN). In: 2017 IEEE International Conference on
Bioinformatics and Biomedicine (BIBM) (2017)
7. Ajay Kumar, K., Prathyusha, V.: Personal authentication using hand vein Trian-
gulation and Knuckle Shape extraction. IEEE Trans. Image Process. 18(9), 2127–
2136 (2009)
8. Wang, Y., Zhang, K., Shark, L.-K.: Personal identification based on multiple key
point sets of dorsal hand vein images. IET Biometrics 3(4), 234–245 (2014)
9. Khan, M.H.-M., Khan, N.M.: Dorsal hand vein biometric using independent com-
ponent analysis. In: 6th International Conference on Internet Technology and
Secured Transactions, pp. 191–195 (2011)
10. Zhon, D., Shao, H., Liu, S.: Towards application of dorsal hand vein reconition
under uncontrolled environment based on biometric graph matching. IET Biomet-
rics 8(2), 159–187 (2019)
11. Gupta, P., Gupta, P.: Multibiometric authentication system using slap fingerprints,
palm Dorsal vein and Hand geometry. IEEE Trans. Ind. Electron. 66(2), 9777–9784
(2018)
406 K. K. Singh et al.
12. Huang, D., Tang, Y., Wang, Y., Chen, L., Wang, Y.: Hand-Dorse vein recognition
by matching local features of multisource keypoints. IEEE Trans. Cybern. 45(9),
1823–1837 (2015)
13. Singh, K.K., Bajpai, M.K., Pandey, R.K.: Reconstruction of original signal from
contaminated signal using fractional order differentiator. In: IEEE International
Symposium on Signal Processing and Information Technology (ISSPIT), pp. 274–
278 (2015)
Distance Calculation of an Object
in a Stereo Vision System
1 Introduction
signal, laser, etc. [2]. Passive measurement is the approach where the distance
of the object from the observer is calculated without sending the signals. In
this experiment, passive measurement is used by gathering information without
the use of a signal [3]. This paper highlights the use of the visual information
obtained from the stereo based system to estimate the distance of an object from
an observer.
Stereo vision builds a three-dimensional description from the captured scene
from multiple points, the stereo vision technique is considered passive as addi-
tional lightning such as a laser beam in the scene is used. Passive stereo vision
is mainly used in many applications such as robotics, 3D object recognition and
navigation of mobile robots [3]. This paper uses a passive stereo vision technique
using a binocular vision formula using two cameras that capture the same scene
from a different viewpoint.
2 Related Works
Stereo vision has a wide range of applications in many fields. Many approaches
using stereo vision are presented in different areas [4]. Distance measurement
using a single camera is widely used in many existing systems but a constraint is
imposed in the result obtained [5]. In [6], it is proposed that the 3D reconstruc-
tion of the scene can be only used to a certain scale factor if a single camera is
used.
In [4], the author proposed that calibrated camera becomes useful for appli-
cations that require spatial data in Euclidean space. In [7], author presented the
2D mapping to 3D which enables to describes the point in 2D corresponds to
the point in the 3D scene. In [8], author demonstrated the calculation of the
distance of an object along with the size measurements using multiple vision
sensors. A new approach for creating a 3D scene for mobile robotics by com-
bining a predictive maintenance (PdM) camera and a stereo camera is proposed
in [9].
In [10], a new method is proposed by using real-time motion estimation
such that depth errors that occurred to camera shift is reduced. This helps in
rectifying the errors in alignment between the cameras. The motion estimation
framework is incorporated with the disparity between the stereo camera.
3 Theory
The stereo vision-based system to reconstruct the three-dimensional description
of the target object is based on triangle theory. Triangle is formed by the two
cameras and the target object. Measuring the distance of an object using stereo
vision is shown in Fig. 1.
Here, A is the target point representing the object, OL and OR represents
the optical centers for the stereo camera with the same focal length, AL and
AR gives the points for projection of the target point A of the stereo camera
respectively, the length of the left and right abscissa corresponding to the AL
Distance Calculation of an Object in a Stereo Vision System 409
and AR in the camera are given by xl and xr respectively. So, the coordinates
representing the object is xl and xr , b denotes the actual distance between the
two cameras, f represents the focal length of the camera as shown in Fig. 1. The
disparity between the two sets of images is given by d = xl − xr and L represents
the actual distance of the target object from the camera, Eq. 1 gives the formula
for calculating the value of L for the basic stereo binocular vision.
b∗f
L= (1)
xl − xr
The coordinate system required to be considered for designing a stereo vision
for calculating the distance for the target object from the camera viewpoints are
given below [11]:
(1) World coordinate system (XW , YW , ZW ), user-defined coordinate system in
3 dimension.
(2) Camera coordinate system (XC , YC , ZC ), the coordinate system with the
camera as the center and z-axis is taken along the optical axis.
(3) Image plane coordinate system (x, y), this plane is on the intersection of
the optical center of the camera axis and the imaging point.
(4) Image coordinate system (u, v), which is in accordance with the pixel coor-
dinate system.
Understanding the relationship between different coordinate systems is
required to build a stereo based distance measurement system for calculating
the distance of the target object in the scene. Relationship between camera
coordinates and world coordinates is described by the translation vector t and
rotation matrix R which is given by:
⎡ ⎤ ⎡ ⎤
Xc Xw
⎢ Yc ⎥ ⎢ ⎥
⎢ ⎥ = RT t ⎢ Yw ⎥ (2)
⎣ Zc ⎦ 0 0 ⎣ Zw ⎦
1 1
410 N. J. Singh et al.
Fig. 2. Geometric representation of the object from the observer in lateral view.
Equation 3 provides the relationship between the image plane coordinate sys-
tem and world coordinate system, where the focal length of the camera is denoted
by f and s is the skew coefficient [12].
⎡ ⎤
⎡ ⎤ ⎡ ⎤
x f 000 Xw
R t ⎢ ⎥
⎢ Yw ⎥
s ⎣y ⎦ = ⎣ 0 f 0 0⎦ T (3)
⎣
0 0 Zw ⎦
1 0 010
1
The relationship between image coordinate and world coordinate is presented
in [12]: ⎡ ⎤
⎡ ⎤ ⎡ ⎤ Xw
u ax 0 u0 0 ⎢ ⎥
Yw ⎥
s ⎣v ⎦ = ⎣ 0 ay v0 0⎦ ⎢
⎣ Zw ⎦ (4)
1 0 0 1 0
1
where, (u0 , v0 ) is the coordinate when the camera axis is projected to image
coordinate and, ax and ay is u, v axis scale factors in image coordinate.
Figure 2 shows the geometric representation of the object from the observer.
Point A represents the observer and BC represents the height of the object. h1
is the distance of the bottom of the object from the observer, h2 is the distance
of the center of the object from the observer, h3 is the distance of the top of the
object from the observer and h is the height of the object.
A new equation is formulated to calculate the height of the object is given
in Eq. 2.
h2 h2
h= + (5)
tan(sin−1 hh23 ) tan(sin−1 hh21 )
4 Proposed Method
The proposed method has five processes, calibration of the camera is the first
step followed by rectifying the image, next process is disparity mapping followed
by 3D reconstruction followed by detecting and estimating the distance of the
target object from the camera. The last step is the measuring of the height of
the target object. The parameters of the cameras are given in Table 1.
Distance Calculation of an Object in a Stereo Vision System 411
is obtained by locating the matching pixels for every pixel in the pair of the
images, finally, to determine the distance of the target object, triangulation is
applied to the matches found. Figure 5 gives the rectified image.
4.4 3D Reconstruction
This subsection explains the construction of a 3D representation of the image
scene. 3D reconstruction of the image scene is achieved by using disparity map-
ping and camera calibration parameters. The input of the disparity map is given
by the pixels which match with the coordinates in 3D representation. The coor-
dinates in the 3D coordinate system relative to the optical center of the left
camera are estimated using the fundamental and essential matrices. Figure 7
gives the reconstructed 3D representation of the scene.
the distance used for the estimation is the euclidean distance. Let the coordinates
of the camera representing the viewpoint be (x, y, z) and the coordinates for the
centroid of the bounding box representing the target object be (xc , yc , zc ), then
the distance is calculated by using the following formula:
Fig. 7. Reconstructed 3D
distance of the bottom point of the detected object from the viewpoint. Figure 9
gives the height of the detected object.
5 Result
Stereo based system for estimating the distance of a target object from the
viewpoint is successfully implemented. The distance is measured for the person
which moves towards the camera. The distance measurement of the object is
shown in Fig. 8. The error percentage is estimated by the following formula:
|D − M |
error% = ∗ 100 (7)
D
where,
D = actual distance.
M = measured distance.
Table 2 gives the percentage of error in estimating the distance of the object
from the camera along with the measured distance and the actual distance of
the object. The distance of an object measured using the stereo vision system
gives a minor error. As the analysis of the image is done frame by frame which
leads to slow processing which in turn provides the error. And most importantly
the timing for starting and stopping for the acquisition of the cameras does
not coincide at the same time, which results in the mismatched in the frame
matching between the pair of images.
6 Conclusion
In this experiment, a stereo vision system to calculate the depth of an object
in an image has been implemented using binocular triangle theory, it is shown
that distance can be calculated from a stereo vision system with minimal error.
Camera calibration is applied to estimate the intrinsic and extrinsic parameters
of the camera. This method can be improved by minimizing flaws in matching
video frames. This method can be applied in finding a safe distance in driving
cars as well as in robotics and autonomous car etc.
416 N. J. Singh et al.
References
1. Hou, A.L., Cui, X., Geng, Y., Yuan, J., Hou, J.: Measurement of safe driving
distance based on stereo vision. In: Sixth International Conference on Image and
Graphics (ICIG), Hefei, Anhui, China, pp. 902–907 (2011)
2. Sukhla, N.: A review on image based target distance and height estimation tech-
nique using laser pointer and single video camera for robot vision. Int. J. Eng. Res.
Rev. 3(1), 128–135 (2015)
3. Nagar, S., Verma, J.X.: Distance measurement using stereo vision. Int. J. Electr.
Electron. Eng. 07(01), 50–53 (2015)
4. Appiah, N., Bandaru, N.: Obstacle detection using stereo vision for self-driving
cars. In: IEEE Intelligent Vehicles Symposium (IV) (2011)
5. Fathi, H., Brilakis, I.: Multistep explicit stereo camera calibration approach to
improve Euclidean accuracy of large-scale 3D reconstruction. J. Comput. Civ. Eng.
30, 04104120 (2014)
6. Pollefeys, M., et al.: Detailed real-time urban 3D reconstruction from video. Int.
J. Comput. Vision 78(2–3), 143–167 (2008)
7. Hsu, T.-S., Wang, T.-C.: An improvement stereo vision images processing for object
distance measurement. Int. J. Autom. Smart Technol. 5, 85–90 (2015)
8. Kuhnert, K.-D., Stommel, M.: Fusion of stereo-camera and PMD-camera data for
real-time suited precise 3D environment reconstruction. In: IEEE/RSJ Interna-
tional Conference on Intelligent Robots and Systems (2006)
9. Santoro, M., AlRegib, G., Altunbasak, Y.: Misalignment correction for depth esti-
mation using stereoscopic 3-D cameras. In: IEEE 14th International Workshop on
Multimedia Signal Processing (MMSP) (2012)
10. Mahammed, M.A., Melhum, A.I., Kochery, F.A.: Object distance measurement by
stereo vision. Int. J. Sci. Appl. Inf. Technol. (IJSAIT) 2, 05–08 (2013)
11. Hold, S., Nunn, C., Kummert, A., Muller-Schneiders, S.: Efficient and robust
extrinsic camera calibration procedure for lane departure warning. In: IEEE Intelli-
gent Vehicles Symposium, pp. 382–387 (2009). https://doi.org/10.1109/IVS.2009.
5164308
12. Steger, C., Ulrich, M., Wiedemann, C., Shaorong, Y.: Machine Vision Algorithms
and Applications, pp. 257–261. Tsinghua University Press, Beijing (2008)
13. Adi, K., Widodo, C.E.: Distance measurement with stereo vision. Int. J. Innov.
Res. Adv. Eng. (IJIRAE) 4(11), 24–27 (2017)
Comparative Assessment of Consumer
Perception on Sports Brands Using Data
Mined from Online Communities
1 Introduction
Sportswear industry and Sports Apparels have become popular among people
and especially in the youth as it makes them more comfortable and relaxed
in sportswear outfits. Due to the increase in the demand of sportswear, big
Sports Brands companies like adidas, Nike, Fila, Reebok and Puma tried to
increase their business with focus of sales of their products of various categories
and marketing in India. Sports Brands have become an important part in the
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 417–429, 2020.
https://doi.org/10.1007/978-981-15-3666-3_34
418 V. Singhvi and P. Srivastava
2 Review of Literature
In order to introduce or popularize a particular product quickly in the market,
Advertisement is one of the best and the most effective ways of branding a
particular product. For more popularity of a product more advertisement is
needed which directly impacts and affects the costing. Since, advertising needs
big amount of investment in the market to have a healthy competition in the
market. Researchers gave the suggestion of using Brand Extension technique
to achieve efficiency of advertising [1]. In addition to this scenario, extension
of brand can also become an important method to showcase new products in
market without advertising them which can save huge amount of money and
lower the cost of new products.
Comparative Assessment of Consumer Perception on Sports Brands 419
Loyal consumers purchase and repurchase a product of the same brand and
accordingly brand loyalty can be measured but the commitment to repurchase a
particular product in addition to its services [2]. It is noted that up to 65% of even
the satisfied customers tend to leave or leave their current service provider with
respect to sports brands because of one of the pleasure lived during the immer-
sion into the service platform [4,5]. This gap leaves researchers with insufficient
information about what actually drives consumers towards Brand Loyalty.
In the year 2005, Langer and Beckman used Netnography to study cross-
consumer online communication about cosmetic surgery on Danish internet mes-
sage board. They further observed that Netnography was very much useful since
it is very difficult to get the information offline. On further observation, Netnog-
raphy provided a deeper insight into consumer’s concerns, opinion and motive
with an opportunity to express opinions freely. Cengiz and Yayala, 2007 com-
pleted their research on effect of marketing mix on positive word of mouth com-
munication. There were sub variables under each p. The project worked and iden-
tified 7 factors which contribute towards the word of mouth. The study focus and
emphasize on what the companies should do to have a positive oral communica-
tion out of which the key factor came out to be excellent service. Another study
reviewed for the study is “Factors influencing purchase of Apparels and Footwear
from organized retail outlet”. Through this study the major factors which affect
consumer perception were style, quality, variety, price, appeal, brand name and
referral group [6].
Some other group of researchers Ko, Kim and Hoon in 2008 tried to compare
consumer behaviors for sports brands in China and Korea. The perceived price
was founded to be negatively influencing the purchase intention in the country
China. One more study in similar field which was reviewed was “The external
factors that influences Taiwanese students shoe purchase intention”. This par-
ticular study was based on the 6 important external factors including brand,
promotion, price, style durability and comfort [8]. Another study in the similar
field which was reviewed for the research is “The external factors that influences
Taiwanese students shoe purchase intention”. This study was based on 6 external
factors which include brand, price, promotion, style, durability and comfort. The
motive of this study was to investigate various factors that influence the Tai-
wanese college students to buy a particular product. The key factors and results
came out of the study were the factors comfort and style as the most important
driving factors for Taiwanese college students shoe purchase intention [7].
Youn, Song and MacLachlan in the year 2008 examined how consumer’s
brand preferences and price sensitivities evolve as their levels of experience
increase in outdoor sports gear categories [9]. Another article reviewed for the
research is “Service quality as a mediator of the relationship between marketing
mix and consumer loyalty”. The 3 factors studied under this are Marketing mix,
Service quality and customer loyalty. The inferences showed that there exists a
strong and important relation between marketing and customer loyalty and a
positive relationship between customer loyalty and service quality. In addition
to above results, there is no intermediate effect of service quality between the
420 V. Singhvi and P. Srivastava
perception of marketing mix and consumer loyalty [10]. Saha, Dey, and Bhat-
tacharya in 2010 in their study “Factors affecting consumer buying behavior of
shoes in Kolkata” tried to identify various aspects in the market that drives the
purchase behavior of customers in city of Kolkata. Four major factors came out
from the study were price, product, place and promotion [12].
Krishna, 2012 focused on the behavior of youth’s buying sports items which
showed different consumers have got different decision making processes. He
also showed that the target goal of the buyer is in getting a qualitative and
quantitative product with low or minimal cost associated with it [13]. Yoh, Mohr
and Gordon in 2012 did an investigation on factors that influence Korean teens
athletic footwear purchase and the research was carried out on 7 basic factors of
style, price, color, brand name, comfort, quality, and celebrity endorsement. The
“Text mining” technique is the data mining technique which is used to automate
the discovery of any kind of knowledge from unstructured dataset or text [14].
4 Methodology
The research study takes key element from Text Mining and Netnography tech-
niques and accordingly proceed further to achieve the objective of the study.
Text Mining technique in software applications is used in the research study for
extracting the comments and important information from online communities for
the research. Netnography will be used by adaptation of the technique of observ-
ing an online community through the data they generate, research sprints and
java coding methods to divide the brand associations and comments based on
BAP. The research study follows 6 steps of Netnography [11] includes Research
Planning, Entrée, Data Collection, Interpretation, Ensuring ethical standards
and Research representation. The research study was carried forward in the fol-
lowing sequential steps:
several Brand Analysis Parameters i.e. Brand Name, Service Quality, Product
Quality, Price, Style, Mobile App/Website etc but only 4 of them which are
Brand Name, Service Quality, Product Quality, Mobile App/Website are con-
sidered in the current research study. Also, 5 popular sports brands in India
were taken into consideration namely adidas, Puma, Nike, Reebok and Fila.
E-commerce platforms of www.flipkart.com, www.amazon.in, www.mouthshut.
com, public tweets from Twitter social media platform, blogs and commenting
forums were taken as data sources from where data collection phase was carried
out [17].
The data collection phase proceeded further after the Identification of data
source was completed and data got accumulated as raw data in this particu-
lar phase. The collected data was divided into primary and secondary data in
which the primary data constituted data from social media platforms and meta
data pertaining to the select a particular sports brands. The next phase would
be data collection using tools like : Data Minor, Web Scraper and using Twitter
API’s. The secondary data constituted the data from published articles from
magazines, online journals, print journals, e-books and newspapers. The data
was collected for the following five Sports Brands: adidas, Puma, Nike, Reebok
and Fila. For the research data collection was done from online e-commerce
platforms, social media sites, blogs and reviews websites. The major e-commerce
platforms used for data collection are: www.flipkart.com, www.amazon.in, www.
mouthshut.com and www.snapdeal.com. Also, the public tweets using twitter
APIs were extracted using JAVA [18] program for the extraction of tweets from
twitter and segregation of comments or reviews was done 4 BAP being used
in the research i.e. Brand Name, Product Quality, Service Quality and Mobile
App/Website. Online comments and Reviews were collected from the mentioned
data sources for each sports brands and finally the data model in terms of
Tables 1, 2, 3, 4 and 5 was created after the data processing step. Tools which
used to collect the data during the research are includes Data scrapper - Data
Minor, Web scrapper and Twitter APIs. Data collection has been done from vari-
ous e-commerce websites, blogs, using twitter APIs etc. and accordingly based on
the Data processing approach data has been divided into positive and negative
comments after reading them. Thus, before reading the comments data has to
be flattened out for which the JAVA programs has been written and accordingly
comments were segregated. Few samples of data collected from various sites are
shown in Figs. 1, 2, 3 and 4.
422 V. Singhvi and P. Srivastava
Fig. 1. A positive comment for Puma sports brand on product quality brand analysis
parameter on an e-commerce site.
Fig. 2. A positive comment for Reebok sports brand on product quality brand analysis
parameter on an e-commerce site.
Fig. 3. A positive comment for Nike sports brand on product quality brand analysis
parameter on an e-commerce site.
Fig. 4. A negative comment for Nike sports brand on brand name and price brand
analysis parameter on MouthShut.com.
to convert the raw data into readable format and the second JAVA coded pro-
gram is being used to count the segregate the sentences containing the comments
and reviews of positive and negative nature. Further using statistical approach
and reading the reviews and comments, each and each meaningful comment and
review is divided into positive and negative comment with respect to the 4 BAP
of Brand Name, Product Quality, Service Quality and Mobile App/Website.
424 V. Singhvi and P. Srivastava
The above processing steps were carried for all the 5 Sports Brands of adidas,
Reebok, Nike, Puma and Fila w.r.t. the 4 mentioned BAP. In the end, the pro-
cessed finished data is collected and stored in MySQL [16] database with the
column as positive (+) and negative (−) for each consumer perception BAP.
Tableau Tool is being used as an Analytic tool for the representation of the pro-
cessed finished data, review and comments for all 5 sports brands with respect
to the 4 BAPs [15]. Tableau public provides strong computational research and
analytic visualization in terms of graphs, charts, diagrams, tables etc. helps in
assessment and doing comparative analysis on the data being mined from online
communities.
In the data representation phase, the data which got collected in Sect. 4.2
and further processed during Sect. 4.3, the same processed data is being used
for Data Representation using the research and analysis tool called Tableau.
The Data which got collected during data collection phase was created into
MySQL database and further using Data processing technique and SQL queries,
the data was flattened out to be in tabular format as shown in the Tables 1,
2, 3, 4 and 5 for Brand analysis parameters (BAP) of Brand Name, Product
Quality, Service Quality and Mobile App/Website. Tableau is a well known BI
Data Analytic tool prominently used in the IT industry for graphical, tabular,
representations and a lot more. The representation of the data is done in the
percentage format after the summation of the each BAP data for all positive and
negative comments total. The data is being transformed into percentage format
to make the base as common for all the parameters for all the sports brands
finally a comparative assessment can made of consumer percentage for selecting
a sports brand. Graphs from 1 to 4 represents the tabular data are shown for
each Sports Brands with each of their BAP in percentage format on the X-Axis.
Tables and Graphs. The Tables 1, 2, 3, 4 and 5 contains the processed data
comments from each of the sample collected from various e-commerce sites, twit-
ter APIs, blogs and forums. Graph 1 to 5 represents the graphical representation
of the tabular data contained in Tables 1, 2, 3, 4 and 5 in percentage format for
each brand.
The 5 tables (Tables 1, 2, 3, 4 and 5) shows the data collection and processed
results w.r.t. 5 sports brands adidas, Puma, Nike, Reebok and Fila with on four
Brand Analyses Parameters/factors i.e. Brand Name, Product Quality, Service
Quality and Mobile App/Website. Data has been collected from various social
networking, e-commerce sites, blogs, reviews post and using search engine tech-
nique results. The ‘+’ symbol represents Positive Reviews/Comments and ‘−’
symbol represents Negative Reviews/Comments.
n
n
P aram P ositive% = ( (P aram P ositivek ) ∗ 100)/( (P aram P ositivek )
k=1 k=1
n
+ (P aram N egativek )) (1)
k=1
Table 1. Adidas data collection results (w.r.t. 4 brand analyses parameters i.e. Brand
Name, Product Quality, Service Quality and Mobile App/Website).
Table 2. Puma data collection results (w.r.t. 4 brand analyses parameters i.e. Brand
Name, Product Quality, Service Quality and Mobile App/Website).
Table 3. Nike data collection results (w.r.t. 4 brand analyses parameters i.e. Brand
Name, Product Quality, Service Quality and Mobile App/Website).
Similarly, the generic % calculation formula for Brand Analysis Negative Param-
eter would be:
n
n
P aram N egative% = ( (P aram N egativek ) ∗ 100)/( (P aram P ositivek )
k=1 k=1
n
+ (P aram N egativek )) (2)
k=1
where n = Number of data samples for the specific parameter either +ve or
−ve In the same way, calculation is performed for the positive and negative
% values for the 6 brand analysis parameters i.e. brand name, product quality,
service quality and mobile App/Website. The 4 graphs (Figs. 5, 6, 7 and 8) shows
averaging percentage from Tables 1, 2, 3, 4 and 5 for Brand Name w.r.t. 5 Sports
Brands. The X-Axis percentage value has been calculated on the basis of formulas
(Param Positive% and Param Negative%) shown in the previous section of the
paper. The 4 graphs include Brand Name Parameter Graph (Fig. 5), Product
Quality Parameter Graph (Fig. 6), Service Quality Parameter Graph (Fig. 7)
and Mobile App or Website Parameter Graph (Fig. 8). All the 4 graphs shows
the comparative assessment of respective graph w.r.t. five Sports Brands which
are adidas, Nike, Puma, Reebok and Fila.
Table 4. Reebok data collection results (w.r.t. 4 brand analyses parameters i.e. Brand
Name, Product Quality, Service Quality and Mobile App/Website).
Table 5. Fila data collection results (w.r.t. 4 brand analyses parameters i.e. Brand
Name, Product Quality, Service Quality and Mobile App/Website).
Fig. 5. Bar graph showing brand name BAP w.r.t. positive and negative percentages
of five sports brands which are adidas, Nike, Puma, Reebok and Fila.
Fig. 6. Bar graph showing Product Quality BAP w.r.t. positive and negative percent-
ages of five sports brands which are adidas, Nike, Puma, Reebok and Fila.
Fig. 7. Bar graph showing Service Quality BAP w.r.t. positive and negative percent-
ages of five sports brands of adidas, Nike, Puma, Reebok and Fila.
Fig. 8. Bar graph showing Mobile App/Website BAP w.r.t. positive and negative per-
centages of five sports brands of adidas, Nike, Puma, Reebok and Fila.
428 V. Singhvi and P. Srivastava
6 Conclusion
The Research study compares 4 Brand Analysis Parameters (BAP) which
includes Brand Name, Product Quality, Service Quality and Mobile
App/Website for driving the consumer perception w.r.t. 5 Sports Brands which
are adidas, Puma, Nike, Reebok and Fila. The paper focusses on the four BAP
factors for measuring the consumer perception in which consumer reviews and
comments from various social media platforms, e-commerce websites, twitter
accounts were taken into consideration. The consumer comments once divided
into positive and negative comments depending on the meaning of the sentence
and using data processing steps. Finally, a comparative assessment was done
for the 4 BAP i.e. Brand Name, Product Quality, Service Quality and Mobile
App/Website and all five sports brands were compared with respect to each
other on every BAP. The data collected and processed was represented in tabu-
lar format which was further converted into percentage values for baseline and
was represented in form of graphs with the help of Tableau business intelligence
and analytic tool. The research study used the steps of Netnography and Text
Mining methodology to collect the data, process it and in the end represented
the processed data using BI & A Tableau tool. Overall the research study, did
the comparative assessment of consumer perception for selecting Sports Brands
using Netnography and Text Mining with the help of data mined from Online
Communities.
References
1. Smith, D.C., Park, C.W.: The effects of brand extensions on market share and
advertising efficiency. J. Mark. Res. 29, 296–313 (1992)
2. Ryan, M.J., Rayner, R., Morrison, A.: Diagnosing customer loyalty drivers. Mark.
Res. 11(2), 18–26 (1999)
3. Chaudhuri, A., Holbrook, M.B.: The chain of effects from brand trust and brand
affect to brand performance: the role of brand loyalty. J. Mark. 65(4), 81 (2001)
4. Arvidsson, A.: Brands: a critical perspective. J. Consum. Cult. 5(2), 235–259 (2005)
5. Arvidsson, A.: Brands: Meaning and Value in Media Culture, p. 168. Routledge,
London (2006)
6. Cengiz, E., Yayala, H.E.: The effect of marketing mix on positive word of mouth
communication: evidence from accounting offices in Turkey. Innov. Mark. 3(7), 486
(2007)
7. Chen, Y.T.: External factors influencing Taiwanese students shoe purchase inten-
tion (2008)
8. Ko, K.H.: A cross cultural study of antecedents of purchase intention for sports
shoes in Korea and China. J. Glob. Acad. Mark. Sci. 18(1), 157–177 (2008)
9. Youn, N., Song, I., MacLachlan, D.: A multi-category approach to modeling con-
sumer preference evolution: the case of sporting goods, June 2008
10. Hu, Y.J.: Service quality as mediator of the relationship between marketing mix
and customer loyalty. Int. J. Organ. Innov. 2, 282 (2009)
11. Kozinets, R.V.: Netnography: Doing Ethnographic Research Online. SAGE Pub-
lications Inc., London (2010)
Comparative Assessment of Consumer Perception on Sports Brands 429
12. Saha, S., Dey, M., Bhattacharya, S.: Factors Affecting consumer buying behavior
of shoes in Kolkata: a case study. IUP J. Manag. Res. 9(4), 39–60 (2010)
13. Krishna: Assessing youth’s buying behaviour towards sports shoes (a case study
of Nike). Int. J. Res. Commer. Manag. 3(9) (2012)
14. Yoh, T., Mohr, M., Gordon, B.: The effect of gender on Korean Teens’ athletic
footwear purchasing. Sport J. 15 (2012)
15. Tableau. https://www.tableau.com. Accessed 9 Oct 2019
16. MySQL. https://www.mysql.com/. Accessed 15 Aug 2019
17. Twitter Developer APIs. https://developer.twitter.com/en/docs. Accessed 10 Oct
2019
18. Oracle JAVA 8. https://www.java.com/en/. Accessed 16 Sept 2019
Analytical Modelling of Internet
of Vehicles (IoV) by IoT Cloud
for Dipping Traffic Congestion and Safety
Alert: A Review
1 Introduction
driver stress, traffic breaking regulations, etc. Road safety and vehicle conges-
tion are the main difficulties facing the transport industry. Due to more traffic
congestion, the efficiency of transport scheme has been reduced and as a result
it increase fuel consumption and travelling time also.
So few methods are require to decrease vehicle traffic congestion is needed to
alleviate the negative environmental and economic impacts imposed by transport
inefficiency [2]. Most vehicle accidents are caused by human errors that can be
avoided and inappropriate driving methods. By using direct or indirect Vehicle
to Vehicle (V2V), Vehicle to Infrastructure (V2I) a distributed system sharing
sensor information from vehicles can decrease accidents. Vehicle-connected sen-
sor technology enables drivers to enhance their driving experience. Self-driving,
linked vehicles and autonomous vehicles are becoming increasingly practicable
with latest developments in sensing techniques. The utilization of area based cau-
tions, which give drivers traffic related data alarms by means of different versatile
and online stages, for example, mapping administrations are utilized by traffic
the board frameworks trying to diminish vehicular blockage [3,4]. These frame-
works are restricted by the compelling inclusion territory just as the expenses of
executing the framework utilizing over the top cell systems, especially in remote
regions. In this way, a devoted traffic location and the executives framework that
improves the use and throughput of street framework is required for improving
the traffic stream by giving the drivers with ongoing area based alarms identi-
fying with suggested elective courses as well as driving paces.
The abrupt traffic stoppage particularly in quick looking over streets and
roadways portrayed by a rare imperceptibility is one of the significant reasons for
mishaps among vehicles. It tends to be brought about by different mishaps, work-
in-advance on streets, tops of traffic, etc. Some IoT Cloud framework is gone for
traffic observing and ready warning for drivers. Portable sensors are spoken to
by GSM/GPRS/GPS TK103 tracker based frameworks introduced in vehicles
that gather geo-confinement and speed information that are sent to an OpenGTS
server that stores them in a SQL database, giving a continuous Open- StreetMap
representation of the traffic situation. Furthermore approaching unstructured
geolocation information are sent to a GeoJSON [6] parsing microservice so as to
be embedded in a MongoDB database.
The utilization of vehicle mounted sensors are actualized by current street
security frameworks for helping drivers by giving vulnerable side help, keeping
up safe after separations what’s more, path help however are unequipped for
recognizing non viewable pathway components of impacts. The Internet of vehi-
cles (IoV) is an Internet of Things (IoT) based arrangement focused at meeting
the street wellbeing and traffic the board objectives of Intelligent transportation
System (ITS) for vehicular customers which has picked up progress because of
the expanded dependability of remote correspondence and sensor technologies.
The IoV arrangement comprises of a WSN containing a focal server which pro-
cedures the got information continuously to create relevant measures, vehicular
customer hubs and Roadside Units (RSU) which capacity as waypoints for gath-
ering the required tactile information and spread of wellbeing and traffic related
cautions [5].
432 S. Debnath and A. Saha
2 Related Works
In this technology era, for developing any IoT infrastructure, its very needful to
know about all connected things with limited computation power and efficiency.
There are numerous methodologies towards a progressively compelling street use
and an increasingly wise method for expanding the limit of the street organize,
since structure new framework is never again a fitting choice. Few examples of
advanced traffic control frameworks are, e.g., intelligent speed limits, versatile
incline metering or dynamic steering. These models depend on an incorporated
traffic the executives, which controls the activity and the reaction as per a given
traffic circumstance.
Nowadays the transportation systems get influenced only for the development
of society, economy and innovation, and also for the family unit qualities. The
adjustment in vehicle framework causes an adjustment in vehicle conduct and
locational example of the framework. The adjustment in family unit attributes,
transport conduct, locational design, and other development impacts bring about
the development of traffic. Traffic congestions included the change or improve-
ment in street limit is just as the after effect of progress in the transportation
framework and subsequently at long last a circumstance emerges where the traffic
request is more prominent than the limit of the roadway. Few autonomous vehicle
are equipped with some controls, so that they can calculate their neighbor vehi-
cle distance using distance algorithm which may reduce few traffic congestion.
To prevent accident distance ranging principal are one of the most important
technique.
As IoT Cloud [6] are used to vehicular traffic monitoring and vigilant notifica-
tion continuously, so its generate a large amount of data. So data storing, data
processing and data recovery are the major issues as the exponential growth
of vehicular traffic data. During accident, its need to generate data immedi-
ately and send immediately that correspondent node where security authorities
resides. Moreover in hilly areas such traffic congestion are high due to narrow
zig zag road, landslides during rainy season, flood effect etc.
An IoT Cloud [6] techniques were used to processing of big vehicular traffic
data which generated using MongoDB and OpenGTS. Considering the constant
time delay for any vehicle to Cloud Server and that to Client, it shows data
retrieval is so fast. So it may reduce traffic congestion and give a quick alert
notification to correspondent node. The same way the system can alert to all
neighbor node so that they can aware of the incident and may reduce traffic
congestion. All the mobile sensor node are characterized by few tracker based
system like GPRS, GSM, TK 103 etc. Vehicular Traffic vehicular data are col-
lected through unstructured way and send it to GeoJSON for data parsing and
thereafter insert into MongoDB database storage. Due high speeds and limited
degrees of independence in vehicle node, sometimes its hard to locate their posi-
tion. The data communication between different nodes are performing by 4G
network. Therefore micro-services and OpenGTS server are employed through
Docker containers to take benefit of the virtualization of resources.
Analytical Modelling of Internet of Vehicles (IoV) by IoT Cloud 433
from accident that can be brought about by careless driving, inappropriate driv-
ing practices and absence of keeping up of safe driving separation. Moreover area
based alarm may installed for road safety related track.
devices for vehicles and OpenGTS. Such communication features enable vehicles
to send those information via a fastest connection to an OpenGTS server [6].
The server supports the visualization of all location based data through a col-
laborative road map. All unstructured dta now store through SQL database in a
structured manner. GeoJSON is require for alert notification. Assuming latency
of communication network is constant for transferring data from Vehicle node
to infrastructure and vice versa. Here the analytical model require data pars-
ing and insertion and query time processing for that. As number of subsequent
request increasing so more data needs to store and retrieve that same. Vehicle
node speed are not beyond control so sometime few request may loss for that
speed. But considering a maximum speed, if the alert message can send, it will
help to forward the message to drivers app and prevent accident and as a result
traffic congestion also reduced.
Fig. 1. System design for IoT Cloud system for traffic monitoring and alert
notification [6]
of IoV. It provides not only vehicle communication assuredly but as well as vehi-
cle intelligent also. IoV is an integrated network system with a high managerial
level, usability, operationalization and credibility. It consisting of several clients,
several vehicles, several nodes and several networks. The bellow figure shows
heterogenous network structure of Internet of Vehicle. Internet of Vehicles (IoV)
architecture shows seven different layer [17] named as user vehicle interface layer,
Data Acquisition Layer, Data Filtering and Preprocessing Layer, Communica-
tion Layer, Control and Management Layer, Processing Layer, Security Layer.
First one shows the interaction between vehicle node with other correspondent
node. Data acquisition layer are responsible for collecting data from various inter-
connected vehicle nodes. The same phenomenon seen in IoT Cloud system with
different way where Open GTS server are responsible for collecting all unstruc-
tured data. Third layer are accountable to avoid the transmission of unrelated
content and to reduce network traffic Congestion. Communication Layer pro-
vides the appropriate network to send the information by using congestion and
some other parameter (Fig. 2).
Traffic management, traffic engineering, and packet transmission are the few
policies which maintained by Control and management layer. Huge number of
data are generating on such scenario which need to maintained properly by
database software. Data processing are the major issues which done processing
layer. Query data processing also helps to send alert message and notification to
all correspondent node with an useful time. Data authentication, data integrity,
data nonrepudiation, confidentiality and several access control among different
nodes are the functions of security layer. All seven layers for IoV architecture
are shown in bellow Fig. 3.
Analytical Modelling of Internet of Vehicles (IoV) by IoT Cloud 437
Several open source discrete simulation platform use for IoV simulation for eval-
uating performance [10]. Few wireless communication technologies like ZigBee,
WiFi, DSRC are studied here for short range communication. IoV simulations
are affecting through model measurement, time delay, packet size, traffic model,
different communication channel. For short range communications using ZigBee,
WiFi, DSRC technologies needs a proper model dimension. During simulation
the sender node moves at particular speed to the receiving node. Therefore a
deterministic movement approach used in a straight line for Vehicle movement.
Simulation time can be calculated at fixed speed by the node to cover the chan-
nel’s range of wireless communication technology. There is no propagation delay
as packet covers a total distance which is less than 1 km. The time required for
transmission delay is 100 ms. Transmission delay can be computed by the ratio of
packet length and data rate of the channel. Therefore to find maximum amount
of packet which can be switched for ZigBee, WiFi and DSRC can be calculated
by the following formula [10].
P acketSize(l)
T ransmissionDelay(T ) = (1)
DataRate(r)
438 S. Debnath and A. Saha
In an IoV system the timing of data packet exchange surveys the memo-
ryless exponential Poisson distribution. The time of generating packets follows
an exponential distribution. During Simulation the mean frequency range varies
from 1 Hz to 1 kHz. Another comparison can be made for long range wireless
communication for Vehicle to vehicle communication and vehicle to infrastruc-
ture communication [18,19]. The ranges which support such communication is
of at least 1 km (Table 1).
Technologies Carrier fre- Series (R) Association Through Power con- Maximum Topology
quency (f) time (At ) put (Th ) sumption number of
(P) nodes (n)
WiFi 2.4/5 GHz 100 m–1 Km 4s 54 Mb/s >400 mA No limit Star and
Mesh
ZigBee 2.4 GHz 100 m 30 ms 250 Kb/s 30 mA >65000 Mesh
Bluetooth 2.4 GHz 10 m 600 ms 1 Mb/s 40 mA 8 Star and
Mesh
As vehicle speed is not controlled by any other node, so packets are sent from
moving nodes. Considering Packet size is constant for wireless communication
technologies WiFi, Zigbee and DSRC [10]. Packets are coming from moving
vehicle so few packet may loss during transmission. The highest packet loss
by DSRC technologies [10]. Due to lower bandwidth leads to lower congestion,
ZigBee can exchanged low frequency of message. The reason behind packet loss
not only traffic congestion, it might be due to gesture or sometimes for Doppler
Effect. When both the wave’s source and the wave’s receiver are in movement,
then only effect found. As a result it produce alteration of the received frequency.
The received frequency can be calculated using carrier frequency, differences
in speed and with the speed of light. The equation for variations in received
frequency for Doppler effect are as follows [10]
Δv
F requency = f c(1 ± ) (2)
c
Most of the packets sent are still effectively obtained as the frequency loss for
speeds in a particular value. Whereas the value is lower than the channel band-
width for WiFi, Zigbee and DSRC wireless communication techniques. Higher
data rates confirm that transmission delay with Wi-Fi technologies is lower than
DSRC technology. Thus congestion is minimized for increasing throughput of
the link.
Analytical Modelling of Internet of Vehicles (IoV) by IoT Cloud 439
References
1. Singh, S.K.: Road traffic accidents in India: issues and challenges. Transp. Res.
Procedia 25, 4708–4719 (2017). Science direct
2. Darwish, T., Abu Bakar, K.: Traffic density estimation in vehicular ad hoc net-
works: a review. Ad Hoc Netw. 4, 337–351 (2016). https://doi.org/10.10007/
1234567890
3. Zhu, Z., Loo, J., Chen, Y., Chai, K.K., Zhang, T.: Recent advances in connected
vehicles via information-centric networking. In: IET International Conference on
Intelligent and Connected Vehicles (ICV 2016), Chongqing, pp. 1–8 (2016)
4. Hirose, K., Ishibashi, K., Yamao, Y., Hirayama, Y., Sawada, M.: Low-power V2M
communication system with fast network association capability. In: 2015 IEEE 2nd
World Forum on Internet of Things (WF-IoT), Milan, pp. 204–209 (2015)
5. Djajadi, A., Putra, R.J.: Inter-cars safety communication system based on Android
smartphone. In: 2014 IEEE Conference on Open Systems (ICOS), Subang, pp. 12–
17 (2014)
6. Celesti, A., Galletta, A., Carnevale, L.: An IoT cloud system for traffic monitoring
and vehicular accidents prevention based on mobile sensor data processing. IEEE
Sens. J. 18(12), 4795–4802 (2018)
7. Guerrero-ibanez, J.A., Zeadally, S., Contreras-Castillo, J.: Integration challenges
of intelligent transportation systems with connected vehicle, cloud computing, and
internet of things technologies. IEEE Wirel. Commun. 22, 122–128 (2015)
8. Lin, F., Huang, Q., Tan, Q.: Distributed storage video cloud of intelligent trans-
portation system based on the backward cloud generator. Adv. Transp. Stud. 2,
15–22 (2015)
9. Abhale, A., Khandelwal, S.: Intelligence transportation service using vehicular
cloud network. In: 2016 IEEE International Conference on Advances in Electronics,
Communication and Computer Technology, ICAECCT 2016, pp. 277–282 (2017)
440 S. Debnath and A. Saha
10. Thakur, A., Malekian, R.: Internet of vehicles communication technologies for
traffic management and road safety applications. Wirel. Pers. Commun. 109(1),
31–49 (2019). https://doi.org/10.1007/s11277-019-06548-y. LLC, part of Springer
Nature, Springer Science Business Media
11. Barrachina, J., Sanguesa, J., Fogue, M., Garrido, P., Martinez, F., Cano, J.: V2Xd:
a vehicular density estimation system that combines V2V and V2I communications.
In: 2013 IFIP Wireless Days (WD) (2013)
12. Eze, E.C., Zhang, S.J., Liu, E.J., Eze, J.C.: Advances in vehicular ad-hoc net-
works (VANETs): challenges and road-map for future development. Int. J. Autom.
Comput. 13(1), 1–18 (2016)
13. Thakur, A., Malekian, R., Bogatinoska, D.C.: Internet of Things based solutions
for road safety and traffic management in intelligent transportation systems. In:
Trajanov, D., Bakeva, V. (eds.) ICT Innovations 2017. CCIS, vol. 778, pp. 47–56.
Springer, Cham (2017). https://doi.org/10.1007/978-3-319-67597-8 5
14. Du, Z., HomChaudhuri, B., Pisu, P.: Coordination strategy for vehicles passing
multiple signalized intersections: a connected vehicle penetration rate study. In:
2017 American control conference (ACC) (2017)
15. Celesti, A., Mulfari, D., Fazio, M., Villari, M., Puliafito, A.: Exploring container
virtualization in iot clouds. In: 2016 IEEE International Conference on Smart Com-
puting (SMARTCOMP), pp. 1–6, May 2016
16. Cheng, J.J., Cheng, J.L., Zhou, M.C., Liu, F.Q., Gao, S.C., Liu, C.: Routing in
internet of vehicles: a review. IEEE Trans. Intell. Transp. Syst. 16(5), 2339–2352
(2015)
17. Golestan, K., Soua, R., Karray, F., Kamel, M.: Situation awareness within the
context of connected cars: a comprehensive review and recent trends. Inf. Fusion
29, 6–83 (2016)
18. Hirose, K., Ishibashi, K., Yamao, Y., Hirayama, Y., Sawada, M.: Low-power V2M
communication system with fast network association capability. In: 2015 IEEE 2nd
World Forum on Internet of Things (WF-IoT), Milan, pp. 204–209 (2015)
19. Hannan, M., Habib, S., Javadi, M., Samad, S., Muad, A., Hussain, A.: Inter-vehicle
wireless communications technologies, issues and challenges. Inf. Technol. J. 12(4),
558–568 (2013)
20. Thakur, A., Malekian, R., Bogatinoska, D.C.: Internet of things based solutions
for road safety and traffic management in intelligent transportation systems. In:
Trajanov, D., Bakeva, V. (eds.) ICT Innovations 2017. CCIS, vol. 778, pp. 47–56.
Springer, Cham (2017). https://doi.org/10.1007/978-3-319-67597-8 5
Object Detection Using Point Feature
Matching Based on SURF Algorithm
1 Introduction
Object detection plays a crucial part in image processing field. For implementa-
tion of object detection, an image is acquired called test image. Some descriptor
points are detected on the test image and a comparison is carried out with the ref-
erence image. Point feature detection is used to detect a specified target i.e. one
specific object instead of that kind of objects in a cluttered scene. The algorithm
in simple words is based on comparing and analyzing feature correspondences
between the reference target image and the given cluttered image.
The suggested approach is associated with trending and competent matching
methods. Specifically, to objects, which consider only images and their gradients
to detect objects. This algorithm gives more accuracy for objects that show
texture patterns that are non-repeating, which results in generation of matches
for unique features. Whereas for homogeneously colored objects and for objects
bearing repeating patterns, this method is not likely to work [2]. The Speeded Up
Robust Features (SURF) algorithm is used for classifying objects of a specific cat-
egory. Speeded Up Robust Features (SURF) algorithm can be adopted for many
applications such as object recognition since it is a local feature and descriptor
algorithm. “SURF” considers quite abundant number of feature descriptors from
original image. This, in turn, results in a reduction of the errors caused by local
variation in the average of all feature matching. As SURF has feature descrip-
tor which is consistent to scale, partially consistent in illumination changes and
orientation, so it can identify and recognize objects robustly even in case of
clutter and partial occlusion. Speeded UP Robust Features “SURF” algorithm
incorporates the following three steps:
Detection Step: This is the first step in which distinguishing locations such as
blobs, corners, and other interest points are selected in the original image.
Description Step: It is the second step in which interest points selected in the
previous step should be allocated unique identifiers. These identifiers should be
in-dependent of features scale and rotations [3]. These identifiers thus achieved
are called descriptors. The descriptors are vectors that contain information about
the points itself and the surroundings.
Matching Step: The third step in SURF algorithm, where there is a comparison
of descriptor vectors between the object image and the new input or original
image. The matching score is calculated concerning the distance between vectors
e.g. Euclidian distance and the sign of Laplacian.
Object detection systems are extensively used in the fields of video surveil-
lance, and image retrieval. It has been observed that real-time object detection
and learning are demanding and high priority tasks. Domains of object detection
that have been quite exposed, comprises of face detection and pedestrian detec-
tion. Object detection has relevance in several progressive fields of computer
vision, such as robotics and industrial automation.
2 Literature Review
The paper by the authors Ms. Vimal Sudhakar Bodke, Prof. Omkar S Vaidya
mainly focus on repeatability of same features in two different images. The algo-
rithm proposed is SURF algorithm as feature detection [2].
The proposed paper by Faishal Badsha, Rafiqul Islam, Mohammad Farhad
Bulbul [5] has stated the comparison between SIFT algorithm and SURF algo-
rithm. While for the later, Hessian matrix is used for feature detection and Haar
wavelet is explained for assignment of descriptors. An architecture of the working
system has also been diagrammatically stated by the paper.
Authors Vishweshwar Todkari, Akash Pawar, Yogeshwar Shanbhag, Gan-
garam Rajput, T. S. Bhoye have done a survey on various object detection tech-
niques [7]. The algorithm discussed is SIFT (Scale Invariant Feature Transform)
algorithm which is the basis for SURF algorithm, for evaluating color descriptors
using photometric changes.
Object Detection Using Point Feature Matching 443
Other methods such as dual background elimination has also been discussed.
In another paper proposed by Rasool Reddy, Krishna, Ravi Kumar [8] SURF
has been described to be robust for rotations up to certain degree.
Hemalatha and Santhosh Kumar [9] in the proposed paper, have explored
the usage of SURF algorithm in the field of security. The paper is focused on
copy-move forgery based on application of image segmentation. By comparing
the key features in an image using SURF algorithm the forgery is detected.
3 SURF Algorithm
The SURF Algorithm stands for Speeded up Robust Features. It has been devel-
oped by Bay et al. The SURF algorithm uses a geometric transform to distinguish
the locale of the object which is to be detected in the given image. As mentioned
above, it consists of 3 basic steps:
• Extraction,
• Description, and
• Matching
Where, Lxx (x, σ) is the convolution result of the second order derivative of
Gaussian filter (δ 2 g(σ))/(δx2 ) with the image I in point X and similarly for
Lxy (x, σ) and Lyy (x, σ).
Equation 1 Shows the Gaussian second orders partial derivatives in y-
direction and xy-direction [2,5] and can be given as under:
The circular region that lies around the interest point provides information.
Based on this set of information a reproducible orientation is fixed. This is the
foundation of SURF descriptor. The SURF descriptor is obtained by constructing
a square region aligned to the chosen position. A unique and efficient feature
description can be achieved by extraction and definition of descriptors. The
SURF descriptor is formed on Haar wavelet responses in order to be invariant
to rotation in x and y direction [3,5]. Haar wavelet’s mother wavelet Ψ (t) can
be mathematically stated as:
⎧
⎪
⎨1, if 0 t < 1/2
Ψ (t) = −1, if 1/2 t < 1 (4)
⎪
⎩
0, otherwise
Fig. 2. Flow diagram for point feature matching using SURF algorithm
446 M. Dey et al.
5.1 Experiment
Initially, the algorithm reads the reference image that consists of the object of
interest (see Fig. 3). The next step is reading the target image containing more
than one object or the scene in which the object is to be detected (see Fig. 4).
Then, the feature points in both the images are detected and the strongest fea-
ture points are visualized in the image considered for reference (see Fig. 5) as well
as target image (see Fig. 6). Extraction of feature descriptors and their match-
ing is done using the attained descriptors at the interest points in both images.
Resulting in displaying of putatively matched features (see Fig. 7). Subsequently
displaying localized Object in the scene, done using presumed matches and elim-
inating outliers (see Fig. 8). Next figure (see Fig. 9) displays the detected object
by getting the bounding polygon of the reference image.
Next is the detection of other image, which can be done using the previously
mentioned steps (see Figs. 10 and 11). Finally, displaying the detected object by
point feature matching algorithm (see Fig. 12) [6,8].
Fig. 3. Reading the reference image that contains the object of interest
Fig. 4. Reading the target image containing more than one object
Object Detection Using Point Feature Matching 447
Fig. 10. (a) Reading an image consisting of a second object of interest (b) Detecting
and visualizing feature points
recognized (see Fig. 12). It can be observed that this algorithm is robust to small
amount of out-of-plane rotation as in the above figures. The constraints that are
to be applied to this algorithm is illumination, orientation and presence of other
similar objects. This also doesn’t work for uniformly colored objects (see Fig. 7)
where, two objects were detected to have similar features [2]. Hence, the accuracy
of this system can be enhanced in the matching section of the algorithm.
6 Measurement of Accuracy
6.1 Table Depicting the Objects and Their Detection Using SURF
Algorithm
See Fig. 13.
– Sensitivity:
Sensitivity = TP/(TP + FN) = (Number of true positive assessments)/
(Number of all positive assessments) [13]
Here, Sensitivity = 7/8 = 0.875
– Specificity:
Specificity = TN/(TN + FP) = (Number of true negative assessments)/
(Number of all negative assessments) [13]
Here, Specificity = 1/2 = 0.5
– Accuracy:
Accuracy = (TN + TP)/(TN + TP + FN + FP) = (Number of correct
assessments)/Number of all assessments) [13]
Here, Specificity = 8/10 = 0.8
Thus, by the above assessment, the SURF algorithm gives an accuracy of 80%
when implemented on a set of input images. These images were captured from
different angles and distances.
7 Conclusion
The SURF algorithm is built on comparing and analyzing point correspondences
between the reference image and the target image. Use of integral images helps in
mathematically determining summations over image sub grids rapidly. Hessian
matrix used for feature detection is known for accuracy and performance. To
implement method of smoothing i.e. reducing the amount of intensity variation
between one pixel and the next, of images box filters are used. The proposed
system can be implemented with certain constraints in it. Where, the accuracy
of matching can be increased, although the system results in good overall per-
formance when the conditions of illumination, coloring, texture are looked after.
SURF Algorithm is somewhat influenced by the scale-invariant feature trans-
form (SIFT) algorithm. The difference being that SURF is most of the instances
has proved to be faster than SIFT and is said to be less fragile against diverse
image transformations than SIFT [5].
References
1. Yang, Z.L., Guo, B.L.: Image mosaic based on SIFT. In: International Conference
on Intelligent Information Hiding and Multimedia Signal Processing, pp. 1422–1425
(2008)
2. Bodke, V.S., Vaidya, O.S.: Object recognition in a cluttered scene using point
feature matching. IJRASET 5(IX), 286–290 (2017)
3. Bay, H., Tuytelaars, T., Van Gool, L.: SURF: speeded up robust features. In:
Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3951, pp. 404–
417. Springer, Heidelberg (2006). https://doi.org/10.1007/11744023 32
452 M. Dey et al.
4. S. U. R. F. (SURF): Herbet bay, andreas ess, tinne tuytelaars, luc van gool. Elsevier
pre-print (2008)
5. Viola, P.A., Jones, M.J.: Rapid object detection using a boosted cascade of simple
features. In: IEEE CVPR (2001)
6. Todkari, V., Pawar, A., Shanbhag, Y., Rajput, G., Bhoye, T.S.: A survey on sus-
picious object detection. IJARIIE 4(1), 628–631 (2018). ISSN(O)-2395-4396
7. Lowe, D.: Distinctive image features from scale-invariant key point. Int. J. Comput.
Vis. 60, 1–28 (2004)
8. Rasool Reddy, K., Krishna, K.V.S., Ravi Kumar, V.: Detection of objects in clut-
tered scenes using matching technique. IJECT 5(spl – 3), 42–44 (2014)
9. Hemalatha, S., Santhosh Kumar, S.: Image forgery detection using key-point
extraction and segmentation. IJPT 8, 13219–13229 (2016)
10. Saleem, S., Bais, A., Sablatnig, R.: A performance evaluation of SIFT and SURF
for multispectral image matching. In: Campilho, A., Kamel, M. (eds.) ICIAR 2012.
LNCS, vol. 7324, pp. 166–173. Springer, Heidelberg (2012). https://doi.org/10.
1007/978-3-642-31295-3 20
11. Karami, E., Prasad, S., Shehata, M.: Image matching using SIFT, SURF, BRIEF
and ORB: performance comparison for distorted images. In: Proceedings of the
2015 Newfoundland Electrical and Computer Engineering Conference (2015)
12. Athani, S., Tejeshwar, C.H.: Performance analysis of key frame extraction using
SIFT and SURF algorithms. IJCSIT Int. J. Comput. Sci. Inf. Technol 7(4), 2136–
2139 (2016)
13. Zhu, W., Zeng, N., Wang, N.: Sensitivity, Specificity, Accuracy, Associated Con-
fidence Interval and ROC Analysis with Practical SAS. R Health Care and Life
Sciences, NESUG (2010)
Diabetes Mellitus Prediction Using
Ensemble Machine Learning Techniques
1 Introduction
At Present, probably the most common, complex, chronic illness named as dia-
betes. It is such a syndrome that affects the capability of a human body to pro-
duce or use insulin. Diabetes Mellitus (DM) is a clinical syndrome characterized
by persistent hyperglycemia due to an absolute deficiency of insulin [1]. There
are many types of DM but the most common types are Type-1 Diabetes (T1D)
and Type-2 Diabetes (T2D). T1D happens due to idiopathic, whereas T2D hap-
pens when the human body can not produce adequate insulin, or when the body
cells are not able to use insulin accurately, which is called insulin resistance [2].
Diabetes immune system attacks and destroys the beta cells in the pancreas of
c Springer Nature Singapore Pte Ltd. 2020
A. Saha et al. (Eds.): ICCISIoT 2019, CCIS 1192, pp. 453–467, 2020.
https://doi.org/10.1007/978-981-15-3666-3_37
454 Md. T. Islam et al.
the human body which produce insulin. Diabetes represents a critical budgetary
weight to people and society [3]. According to the statistics of International Dia-
betes Federation (IDF), in every seven seconds, someone is estimated to collapse
from diabetes or its entanglements, with half of those deaths (per year in total
4 million) are occurring under the age of 60. This is totally unusual and never
witnessed before in the background of a worldwide diabetes prevalence of 8.8%
(95% confidence interval 7.2% to 11.3%) of the world population in 2017 [4]. In
Bangladesh, the percentage of diabetes affected patients has increased substan-
tially to 10% or 8.4 million population of 14.8 million total population in 2011
while it was only 4% in 1995 to 2000, 5% in 2001 to 2005 and 6% in 2006 to 2010
[5]. It indicates that the current percentage of diabetic patients in Bangladesh is
the highest among all the neighboring countries. A report of Daily Science stated
that, if it was possible to prevent diabetes in an earlier stage, the possibility of
minimizing the destructive effects of diabetes would be more. There are several
Machine Learning (ML) techniques such as classifications, clustering and associ-
ation which are developed and broadly used in the medicinal forecast. Different
models of algorithms provide different accuracy, specificity, sensitivity. So, it’s
easy to find out which algorithm is the best among the existing algorithms.
We have analyzed to find out the accuracy of two ML algorithms, Bagging and
Decorate by analyzing our collected dataset and compared their performances.
The other parts of the manuscript are arranged as follows: in Sect. 2, Sect. 3
the related works and methodology have been elaborated with a distinguish-
ing destination to the justness of the classier algorithms respectively. In Sect. 4
the experimented results of this analysis has been clarified with the impulsion
to justify the novelty of this exploration work. Finally, this research paper is
terminated with Sect. 5.
2 Related Works
An analysis was conducted by a research team on big data of health-care to pre-
dict diabetes accurately. The dataset they used includes total 9 features which
have numerical and nominal attributes. They targeted to make a classifier sys-
tem by using several ML techniques. In that analysis, they found 79.13% accu-
racy using Support Vector Machine (SVM) [6]. Similarly, a group of researchers
applied ML techniques to classify diabetes. They applied a few ML algorithms, in
particular; Naı̈ve Bayes, SVM, Random Forest (RF) and Simple CART. Among
these algorithms, they found 76.5% accuracy using RF in terms of classification
of diabetes [7]. Dutta et al. have worked on critical features to predict diabetes.
They worked with 9 variables and RF algorithm where they found RF to give
84% accuracy [8]. Sreekanth Rallapalli et al. have proposed a predictive model
using CART model & Scalable RF to classify diabetes based on various fac-
tors. They used a dataset contains 1500 instances with having 6 attributes. In
that research, they found that the Scalable RF algorithm gave 87.5% accuracy
whereas the normal RF algorithm gave 75% accuracy [9]. Soumayadeep Manna
and his cooperators have proposed a system to predict important factors that
Diabetes Mellitus Prediction Using Ensemble Machine Learning Techniques 455
cause diabetes. They used a dataset which has 3075 instances and each instance
has 8 features. They have used Logistic Regression (LR) and RF, whereas RF
gave 86.70% accuracy and LR gave 89.17% accuracy [10]. In the same man-
ner, research has been conducted based on ML algorithms where researchers
used SVM, AdaBoost, Bagging, K-NN and RF algorithms with a dataset of 506
instances with having 30 features for each instance. They got 75.49% accuracy
for AdaBoost, 76.28% for Bagging, 72.33% for K-NN, 75.30% for RF, 72.72% for
SVM [11]. Likewise, another research team proposed a predictive model using RF
algorithm based on some variables like age, weight, hip, waist, height, etc. In the
study, they performed the analysis based on 4 groups of datasets where RF gave
84.19% for the group 4 dataset [12]. Kinge et al. analyzed to determine the per-
formances of several algorithms named DT(J48), Naive Bayes, RF, AdaBoost,
Bagging, Multilayer Perceptron (MLP), Simple Logistic to predict diseases using
data mining and ML techniques. A dataset of heart disease that has total 303
instances with 74 raw attributes was taken and only one 14 significant features
were used among them. Accuracy for DT(J48), Naı̈ve Bayes, RF, AdaBoost,
Bagging, Multilayer Perceptron (MLP), Simple Logistic algorithms are 78.15%,
82.59%, 83.15%, 81.59%, 81.59%, 79.41% and 83.1% respectively [13]. Deepika
Verma and Nidhi Mishra conducted a study to identify Diabetes by using a
dataset on Naı̈ve Bayes, J48, Sequential Minimal Optimization (SMO), MLP,
and decreases Error Pruning Tree (REP-tree) algorithms. In their study, they
found the best accuracy, 76.80% from SMO on diabetes dataset [14].
3 Methodology
The total workflow of our study showed in Fig. 1. We have categorized our study
into four different sections. They are,
– Collection of data
– Preprocessing of data
– Training of data
– Application of Machine Learning Algorithms
Start
Data Preprocessing
Dataset training
Apply
Classification algorithms
Bagging Decorate
Determine
Statistical Matrics
Compare Performance
End
After that, combined these into two balancing process to get an ensemble
system,
1
P
sP (.) = wl (.) (3)
P
l=1
There are many ways to aggregate the dynamic models fitted in parallel. The
working procedure of Bagging algorithm is given below:
are trained on actual training data as well as some artificial data. Artificial
training instances in every iteration are generated from data distribution; here
the number of instances to be generated is specified as a fraction, resize, of the
given training set size. The marks for these generated training instances are
chosen to vary maximally from the present ensemble’s predictions. The making
procedure of the artificial data is explained broadly in the following section. We
refer the marked artificially initiated training as to the diversity data. A new
classifier has been trained on the composition of the actual training data and
the diversity data. If after attaching this new classifier to the present ensembles
increase more ensemble training error, we immediately reject this classifier, else
it is added to the present ensemble. This process is continuously repeated until
getting our expected committee size or surpass the highest number of iterations.
For classifying an unmarked instance x, we use the following procedure. Every
base classifier, Ci , in the ensemble C provides probabilities for class membership
of x. If PCi,y (x) becomes the probability of instance x belongs to class y as
stated in the classifier Ci , then we calculate the class membership for the whole
ensemble as [23]:
∗ Pci,y (x)
Py (x) = ci c ∗ (5)
|C |
Age Year
Height cm
Weight kg
BMI kg/m2
Systolic BP mmHg
Diastolic BP mmHg
Blood sugar mmol/L
before meal
Blood sugar mmol/L
after meal
Symptom Year
duration
X
0 50 100 150 200
Figure 2 demonstrates the mean and standard deviation of several major numer-
ical features for instance, mean age of the patients in our dataset is 48.944 while
the standard deviation is 16.047. Again, the mean height in our dataset is 158.31
and the standard deviation is 7.005.
Age Year
Height cm
Weight kg
BMI kg/m2
Systolic BP mmHg
Diastolic BP mmHg
Blood sugar mmol/L
before meal
Blood sugar mmol/L
after meal
Symptom Year
duration
X
0 50 100 150 200
Minimum Maximum
80
60
Percentages
40
20
0 X
high BP
Hunger
Headache for
extremities
Sex
Drug history
Weight loss
Thirst
Relatives
Physical activity
Smoking
Tabaco chewing
Burning
Weakness
Diabetes
Fig. 4. Percentages of several nominal features in terms of positive and negative
The values of Precision, Recall and F-measure are same at 0.985 while MCC is
0.975. For Decorate algorithm the outcomes of ROC and PRC Area are same,
0.999.
Diabetes Mellitus Prediction Using Ensemble Machine Learning Techniques 463
Table 7. Comparison between our proposed system and several existing systems
Comparisons between the two algorithms we used in this study have been
shown in Table 6. After comparing the outcomes of the performance parameters
for both algorithms it is clear that Decorate performed better than Bagging. In
this analysis, Bagging has given 95.59% accuracy. On the other hand, Decorate
algorithm has given 98.53% accuracy. The Specificity of Bagging is 94.9% while
it is 98.3% for Decorate.
Figure 5 illustrates the accuracy of Bagging and Decorate algorithm in per-
centage according to our experiments.
We have compared our proposed model with some other existing systems
based on their accuracy and the dataset they used. The comparison has been
shown in Table 7 which clarifies that our proposed system has given the best
performance among the compared models.
Diabetes Mellitus Prediction Using Ensemble Machine Learning Techniques 465
(a) (b)
Fig. 5. Percentages of different urine colors (a) after meal, and (b) before meal
5 Conclusion
In this contemporary world, a notable portion of people are suffering from dia-
betes all over the world and the most shocking matter is that most of the patients
are unaware of it. Even they don’t know which type of diabetes they have. If
the types of diabetes can be determined at an early stage and can take proper
treatment, diabetes can be controlled and can’t go to an alarming stage. There
are many effective studies have been done to find the easiest way to predict
human diseases using ML which have been very helpful for patients. So, This
is very convenient to use ML to perform this analysis. Our aim was to build
a more accurate model than the existing models using Bagging and Decorate.
During the study, there were many limitations, for example, limited data of the
patients, limitation of instances, features and the most important one is miss-
ing information which have been handled using ML techniques in WEKA tool.
Despite having these limitations, the analysis has been performed properly. How-
ever, the study can also be conducted using other advanced ML algorithms such
as ANFIS and more advanced ensemble techniques which combine Neural Net-
work, Fuzzy System, etc. with more instances and a wider range of attributes. By
comparing these models, an expert system can be developed to predict diabetes
and its types more accurately and effectively.
References
1. Insulin and diabetes, Diabetes UK (2019). https://www.diabetes.org.uk/guide-to-
diabetes/managing-your-diabetes/treating-your-diabetes/insulin. Accessed 02 Jan
2019
2. Diabetes definition causes and symptoms. Davita.com (2019). https://www.davita.
com/education/kidney-disease/risk-factors/diabetes. Accessed 05 Jan 2019
466 Md. T. Islam et al.
20. Han, J., Kamber, M., Pei, J.: Data Mining, 3rd edn, pp. 370–382. Elsevier, Ams-
terdam (2011)
21. Kandan, H.: Bagging the skill of Bagging(Bootstrap aggregating). Medium (2018).
https://medium.com/@harishkandan95/bagging-the-skill-of-bagging-bootstrap-
aggregating-83c18dcabdf1. Accessed 29 Jun 2019
22. Decorate, Weka.sourceforge.net (2019). http://weka.sourceforge.net/doc.stable/
weka/classifiers/meta/Decorate.html. Accessed 31 Dec 2018
23. Melville, P., Mooney, R.: Constructing diverse classifier ensembles using artificial
training examples. In: Proceedings of the 18th International Joint Conference on
Artificial Intelligence, IJCAI 2003, Mexico, Acapulco, pp. 505–510 (2003)
Author Index