Cyberattack Detection in Wireless Sensor Networks Using A Hybrid Feature Reduction Technique With AI and Machine Learning Methods
Cyberattack Detection in Wireless Sensor Networks Using A Hybrid Feature Reduction Technique With AI and Machine Learning Methods
https://doi.org/10.1186/s40537-023-00870-w
*Correspondence:
[email protected]. Abstract
edu.eg This paper proposes an intelligent hybrid model that leverages machine learning
1
Department of Artificial and artificial intelligence to enhance the security of Wireless Sensor Networks (WSNs)
Intelligence, Faculty of Artificial
Intelligence, Egyptian Russian
by identifying and preventing cyberattacks. The study employs feature reduction
University, Badr City 11829, Egypt techniques, including Singular Value Decomposition (SVD) and Principal Component
2
Department of Computer Analysis (PCA), along with the K-means clustering model enhanced information gain
Science, Faculty of Science,
Menofia University, Shibin El
(KMC-IG) for feature extraction. The Synthetic Minority Excessively Technique is intro-
Kom 32611, Egypt duced for data balancing, followed by intrusion detection systems and network
traffic categorization. The research evaluates a deep learning-based feed-forward
neural network algorithm’s accuracy, precision, recall, and F-measure across three vital
datasets: NSL-KDD, UNSW-NB 15, and CICIDS 2017, considering both full and reduced
feature sets. Comparative analysis against benchmark machine learning approaches
is also conducted. The proposed algorithm demonstrates exceptional performance,
achieving high accuracy and reliability in intrusion detection for WSNs. The study out-
lines the system configuration and parameter settings, contributing to the advance-
ment of WSN security.
Keywords: WSN, Hybrid reduction, ML, AI, NSL-KDD, UNSW-NB 15, CICIDS 2017
Introduction
The usage of artificial intelligence AI for cyberattack detection in wireless sensor net-
works with a hybrid feature reduction technique involves developing a system that can
effectively detect and classify cyberattacks in WSN environments. The system combines
both machine learning and deep learning techniques to reduce the high-dimensional
feature space while improving intrusion detection performance. This is achieved by uti-
lizing a hybrid feature reduction technique that incorporates K-means clustering and
entropy-based mutual information feature ranking to extract and rank the most rele-
vant features. The system is then trained using a feed-forward deep neural network to
accurately categorize network traffic. Overall, the aim is to provide early detection and
learning systems with high-performance features for efficient cyberattack detection and
© The Author(s) 2024. Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits
use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original
author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third
party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the mate-
rial. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or
exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://
creativecommons.org/licenses/by/4.0/.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 2 of 39
I. Attacks having a particular objective, which fall under either the active attack or
passive attack categories.
II. Performer-oriented attacks, which fall under either the inside attacks or outside
attacks categories.
III. Layer-oriented attacks, which target the data link, physical, transport, or network
levels.
Motivated by the goals of WSN security, a deep feed forward neural network (DFFNN)
model with k-means clustering (KMC) and information gain (IG) methods is proposed
for attack with the main contributions are described below:
Behiry and Aly Journal of Big Data (2024) 11:16 Page 3 of 39
1. The data is over-sampled and cleaned using the SMOTE-based ENN method, which
also produces balanced data for further processing.
2. Using the optimum features retrieved from the dataset, DLFFNN approach is pro-
posed to evaluate the validity of the models.
3. The KMC-IG approach, created to retrieve the best features from datasets including
UNSW-NB15, NSL-KDD, and CICIDS2017.
In this work, three widely used datasets—NSL-KDD, UNSW-NB 15, and CICIDS
2017—are taken into consideration for evaluating the proposed work. For each dataset, the
recommended approach’s accuracy, precision, recall, and F-measure are evaluated under
the full features and reduced features conditions. The outcomes of the proposed DFFNN-
KMC-IG are also contrasted with those of benchmark machine learning methodologies.
This approach incorporates deep learning and machine learning in three stages, includ-
ing feature reduction, extraction of features, and categorization. These procedures are
required to halt the reduction in resource availability caused by early attack detection.
The structure of this paper is organized as follows. Section "Hyperparameter tuning"
focuses on the related work. Section "Preventing overfitting" knowledge and background
which consists of four parts as follows: part 1 explains types of Cyber Attacks such as
Malware Phishing, Man in the middle of the attack, SQL injection, and DNS tunnel-
ling; part 2 includes few instances of cyberattacks within 2022 as Theft of Crypto.com,
Breach of data at the Red Cross, and Cash app data breach. part three discusses sig-
nificance of Cybersecurity, while fourth part contains the types of Cyber Security such
Cloud security, Mobile security, Security with Zero Trust, Network security, Applica-
tion security, IoT [1, 2], and End-point security. Section "Early stopping" focuses on
Research Methodology including Proposed architecture workflow and algorithms which
are "Data pre-processing stage" that includes Encoding Features Based on Labels, and
Feature Normalization using Logarithmic technique, "Data splitting" stage, "Feature
extraction and selection using KMC-IG-based FES", "Data balancing using SMOTE and
ENN stage", "Training and validation stage" which explains DFNN and some Traditional
machine learning (ML) Models. Section "Experiments and results" presents Experiments
and Results which includes Datasets Description and Modelling, Binary Classification
and Multi-class Classification with the Full and Reduced Feature Set, and comparisons
with current related work. Sect. "Conclusion" is devoted to the conclusion of this study.
Related work
In their work, Kaur Saini et al. [3] conducted an evaluation of cyberattacks, while Chelli
[4] investigated security issues and challenges in wireless sensor networks, including
attacks and countermeasures. Daojing He et al. [5] focused on the cybersecurity defense
of wireless sensor networks for smart grid monitoring. Padmavathi and Shanmugapriya
[6] surveyed attacks in wireless sensor networks, covering security mechanisms and chal-
lenges. Al-Sakib Khan Pathan et al. [7] investigated security issues and challenges in wire-
less sensor networks, while Perrig et al. [8] discussed security in wireless sensor networks.
Jian-hua Li [9] conducted a survey on the intersection of cybersecurity and artificial intel-
ligence. Handa et al. [10] reviewed machine learning in cybersecurity, and Thomas et al.
[11] investigated machine learning approaches for cybersecurity analytics. Gaganjot et al.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 4 of 39
[12] discussed secure cyber-physical systems for smart cities, while Boussi and Gupta [13,
14] developed a framework for combating cybercrime. Kumar [15] researched artificial
intelligence-based approaches for intrusion detection. Shahnaz Saleem et al. [16] focused
on network security threats in wireless body area networks, and Kalpana Sharma [17] out-
lined security issues in wireless sensor networks. Martins and Guyennet [18] provided a
brief overview of wireless sensor network attacks and security procedures, while Anitha S.
Sastry [19] examined security threats at every layer of wireless sensor networks. Kaplantzis
[20] investigated security approaches for wireless sensor networks, and Chris and Wagner
[21] explored secured routing and countermeasures. Yanli Yu et al. [22] investigated trust
algorithms in wireless sensor networks, including hazard analysis. Xu et al. [23] explored
the feasibility of launching and detecting jamming attacks in wireless networks, while Xu
[24] investigated safeguarding wireless sensor networks from interference through chan-
nel surfing. Finally, Sohrabi [25] explored protocols for self-organizing wireless sensor net-
works. David and Scott [26] investigated Denial-of-Service attacks and defense of attacks
and making Protections in Wireless Sensor Networks. Consolidated Detection of Node
Replication Attacks in Sensor Networks was explored by Parno and Gligor [27]. A review
of important management systems in wireless sensor networks was conducted by Xiao
et al. [28]. Abhishek Jain et al. [29] investigated Wireless Sensor Network Cryptographic
Protocols. Daniel E. Burgner is an American businessman. Luay Wahsheh [30] investi-
gated Wireless Sensor Network Cybersecurity. Zhu et al. [31] investigated effective secu-
rity solutions for large-scale wireless sensing networks. Culler and Hong [32] conducted
research on Wireless Sensor Networks. Makhija et al. [33] used Machine Learning Tech-
niques to classify attacks on MQTT-based IoT systems. Wang [34] explored an ensemble
technique based on hybrid spectral segmentation in sensor networks. Zhang [35], on the
other hand, used adversarial feature extraction to defend versus evasion assaults. Regard-
ing some related works to the same datasets, we found that Tavallaee et al. [36] studied
in details NSL-KDD dataset and the KDD CUP 99 data set. Sonule et al. [37] focused on
UNSWNB15 Dataset and ML. Sharafaldin et al. [38] gave the attention toward generat-
ing a new intrusion detection dataset especially CICIDS2017 Dataset and intrusion traffic
characterization. Aly and Alotaibi studied the modified gedunin using ML [39]. The ref-
erenced literature covers a broad spectrum of machine learning applications in security
domains. Johri et al. [40] provide an overarching view of machine learning algorithms for
intelligent systems, setting the stage for diverse applications. Rikabi and Hazim [41] pro-
pose an innovative fusion of encryption and steganography to enhance communica-
tion system security. Ahmad et al. [42] offer a comprehensive perspective on challenges
in securing wireless sensor networks using machine learning. Ismail et al. [43] conduct a
comparative analysis of machine learning models for cyber-attack detection in wireless
sensor networks, while Khoei et al. [44] explore dynamic techniques against GPS spoofing
attacks on UAVs. Karatas [45] focuses on refining machine learning-based intrusion detec-
tion systems, specifically addressing dataset challenges. Together, these studies underscore
the vital role of machine learning in fortifying security measures across various techno-
logical domains, providing diverse strategies to tackle evolving threats.
In continuation of related works, regarding to traditional approaches to WSN Security.
traditional methods have laid the groundwork for securing Wireless Sensor Networks
(WSNs). Cryptographic techniques, as discussed by Dong et al. [46], play a vital role
Behiry and Aly Journal of Big Data (2024) 11:16 Page 5 of 39
existing approaches remain open issues. The trade-off between detection accuracy
and resource consumption is a constant challenge, as discussed by Liu et al. [54].
• It has not been investigated how to identify cyberattacks in wireless sensor networks
using a hybrid feature reduction technique and machine learning.
• DLFFNN methodology is not combined with the SMOTE-based ENN method.
• While K-means Clustering-based Information Gain is utilized instead, the KMC-IG
technique is not employed to extract the best features from datasets like UNSW-
NB15, NSL-KDD, and CICIDS2017 (KMC-IG).
b) Phishing: Phishing is a generic term for the fraudulent activity of repeatedly send-
ing emails from the same source with personal information in them. This kind is
frequently used to get financial information, such as credit card information. The
Behiry and Aly Journal of Big Data (2024) 11:16 Page 7 of 39
hacker infects computers and mobile devices with malware through the email link in
order to steal crucial data.
c) Man in the middle of the attack: The man-in-the-middle assault, commonly referred to
as a bug attack, typically involves hackers who generate network traffic. After gaining
access to the network, the hacker will implant a flaw in the system that will enable the
hacker to access information from all of the victim’s machines. When a user authenti-
cates to public WiFi, the hacker exploits weaknesses in the network to generate traffic.
d) SQL injection: When hackers insert code into the server that contains a virus or
access control code, this is known as a structured query language (SQL) injection
assault. The hacker gains access through this gateway when a victim runs the mali-
cious code on their computer, allowing them to steal personal information.
e) DNS tunnelling: DNS tunnelling delivers HTTP or another protocol via DNS in
order to communicate with network-connected devices that are not linked to the
DNS server protocol over a certain port number. Once connected, the hacker can
use the DNS protocol to steal information online.
1) Theft of Crypto.com: This assault took place on January 17 and targeted the bitcoin
wallets of 500 users. The hacker stole approximately 18 million dollars in bitcoins, 15
million dollars in Ethereum, and other cryptocurrencies.
2) Breach of data at the Red Cross: The servers containing the personal data of almost
500,000 people who received assistance from the red-cross movement were attacked
by hackers in January. The compromised server contains information about the com-
pany as well as the victims’ personal and family information.
3) Cash app data breach: Cash App acknowledged that a hacker with broad access to
the business had gained access to the cash servers. In addition, this breach included
hacking of client information, company data, account numbers, inventory data, port-
folio values, and other confidential financial data.
Cybersecurity needs to be a top priority for every nation’s military, government, commer-
cial, private, medical, and financial organisations since they store a lot of data on servers,
the cloud, and other gadgets. Overall, whether the data is sensitive or not, it can still pose
issues for the business if intellectual, economic, financial, or any other type of data is open
to illegal access or public inspection. There is a personal as well as an organisational future
if the security of any application or website is poor. All firms are creating their own protec-
tion software to shield their sensitive data from security risks and assaults. Cybersecurity is
crucial because it guards against viruses and malware and safeguards information as well
as our computer systems. Cybercrimes are on the rise, and businesses and organisations,
particularly those in the health, economic, and national safety sectors, need to take extra
Behiry and Aly Journal of Big Data (2024) 11:16 Page 8 of 39
precautions to secure their data because the future of any nation depends on it. Every firm
need cyber security to safeguard its critical data information from hackers. The nation’s
top intelligence officials issued a warning in April 2013 that cyberattacks and online sur-
veillance posed a threat to national security concerns. Every person must be concerned
about cyber security. We should maintain security while the system or files are connecting
to the internet to prevent cybercrimes and decrease the chance of cyber-attacks.
1) Cloud security: Cloud computing is another name for cloud security. Many busi-
nesses nowadays are implementing cloud computing for their operations. A primary
concern is ensuring cloud security. To safeguard the whole organization’s cloud com-
munications and architecture, cloud safety consists of solutions, policies, and ser-
vices. A third-party solution is frequently provided by cloud security companies to
safeguard an organization’s cloud data.
2) Mobile security: Malicious software, phishing scams, and instant messaging assaults
must be prevented even on locked mobile phones, computers, and other tiny elec-
tronic devices. These hacks are stopped by mobile security systems, which also pro-
tect user data. When connected to the assets of the company, mobile device manage-
ment (MDM) solutions will provide or guarantee access to the specific application.
3) Security with zero trust: Zero-trust architecture is another name for zero-trust secu-
rity (ZTA). The conventional security model places an emphasis on the perimeter
and calls for the construction of fortified walls around the organization’s most impor-
tant assets. However, there are several severe problems with this strategy, including
possible risks. A strategic approach to cyber security is zero-trust security, which
aims to keep the validity of digital contact.
4) Network security: Only in this area do attacks often occur. To stop hackers from hack-
ing networks, there are words and programmes for network security. Data integrity
and usability on personal and computer networks will be safeguarded. Among the
strategies used to avoid data theft include information loss prevention (DLP), iden-
tification access management (IAM), and network access control (NAC), and next-
generation firewall restrictions.
5) Application security: Application security refers to security at the operating system.
Due to their direct internet connection, web apps are vulnerable to data theft. Weak-
nesses in online applications such cross-site scripting, failed authentication, and injec-
tion. Unauthorized contact with apps and APIs is prevented by application security
6) IoT: IoT security is a procedure used to protect IoT systems from dangers. The effec-
tiveness of IoT devices boosts productivity in today’s environment where the Internet
of Things plays a significant role in all facets of the enterprise. Tools for Internet of
Behiry and Aly Journal of Big Data (2024) 11:16 Page 9 of 39
Things security aid in defending against dangers and breaches. Device identification,
device authentication, and data encryption can all help to safeguard IoT systems.
7) End-point security: Remote computer access occurs in every company. Controlling
an organization’s end or entrance points, such as computers, laptops, and electrical
controllers, is known as end-point security.
Research methodology
This study proposes using the K-means clustering model to improve information gain for
feature reduction/extraction and ranking (KMC-IG). Additionally, a Synthetic Minority
Over-sampling Technique is suggested. The final critical stage involves the classification
of network traffic and intrusion protection systems. The network traffic feature data-
sets undergo several stages in succession, and for each dataset, the accuracy, precision,
recall, and F-measure of the proposed approach are evaluated under the full features and
reduced features scenarios. Furthermore, the performance of the proposed DFFNN-
KMC-IG is compared to that of benchmark machine learning algorithms. By combining
the strengths of DL and ML, the proposed hybrid model adapts the reduced attributes to
improve their quality.
Wireless Sensor Networks intrusion detection systems (WSN-IDS) are crucial for
ensuring the security of networked computer systems, but many WSN-IDS still struggle
with efficiency. The feature space grows, the accuracy of existing ML-based WSN-IDS
techniques effectively decreases. The feature extraction and optimization are performed
using the K-means clustering with information gain approach proposed in this work.
In Fig. 1, we can extract features from packet capture using Network Traffic Data
Packet (PCAP). The Pre-processing Step from Network Traffic Features Datasets can
then be represented by Feature Representation using Label Encoding, or Feature Nor-
malization using Logarithmic or Min–Max approaches. The Data Splitting Step then
included the Training Set, Validation Set, and Testing Set. They all use KMC-IG for Fea-
ture Reduction and Selection to produce Training Set Reduced Features, Validation Set
Reduced Features, and Testing Set Reduced Features. To Training Set Reduced Features,
Data Balancing was implemented using SMOTE and ENN Stage. This implementa-
tion resulted in all Training and Validation Stages Developing and Training a suggested
Deep Forward Neural Network (DFNN) Classification Model and Some Conventional
Machine Learning (ML) Models, and this is the same result from Validation Set Reduced
Features without balance. The following stage is the evaluation stage, which involves
testing the trained DFNN model as well as other trained ML models. Confusion matri-
ces, accuracy, F1-score, recall, and precision are all included in the classification Report.
Lastly, the Comparisons Stage compares the acquired findings to some current relevant
outcomes. Here, the proposed architecture workflow as in Fig. 1.
Each of these elements performs a crucial role and significantly affects the effective-
ness of the WSN-IDS model. The design of the planned work for developing WSN-IDS
is shown in Fig. 1.
Certainly! Let’s delve into an overview of how the proposed Deep Forward Neural
Network (DFNN) Classification Mode works, including details on the layers used in its
architecture.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 10 of 39
1. Pre-processing Stage
1.1 Feature Representation using Label
Encoding
Datasets of Network 1.2 Feature Normalization using
Traffic Features Logarithmic or Min-Max techniques
7. Comparisons Stage
Comparing obtained Classification Report contains confusion
results with the results of matrices, accuracy, F1-score, recall, precision
some current related
1. Input layer:
• The DFNN Classification Mode takes as input features extracted from network
traffic data in the context of Wireless Sensor Networks (WSNs).
• Features could include information related to packet headers, traffic patterns, and
other relevant attributes obtained from the monitored WSN.
2. Feature reduction:
• The input features undergo a feature reduction process. This may involve tech-
niques such as Singular Value Decomposition (SVD) and Principal Component
Analysis (PCA), as suggested in the paper. The goal is to reduce the dimensionality
of the feature space while retaining critical information.
3. K-Means Clustering Model with Information Gain (KMC-IG):
• A K-Means Clustering Model enhanced with Information Gain (KMC-IG) is
applied to further refine and cluster the reduced features. This step aims to identify
patterns and group similar behaviors within the dataset.
4. Synthetic minority excessively technique:
• The proposed Synthetic Minority Excessively Technique is introduced, likely during
or after the clustering stage, to address imbalances in the dataset. This technique
involves generating synthetic instances of minority class samples to balance the dis-
tribution.
5. Deep Forward Neural Network (DFNN):
• The core of the proposed method is the Deep Forward Neural Network (DFNN).
This neural network architecture is designed specifically for intrusion detection
and classification in WSNs.
• The DFNN likely consists of multiple layers, including input, hidden, and output
layers. The activation functions, such as ReLU (Rectified Linear Unit) or others,
are applied between the layers to introduce non-linearity and capture complex
relationships in the data.
6. Evaluation metrics:
• The performance of the DFNN is evaluated using standard metrics such as accu-
racy, precision, recall, and F-measure. These metrics provide a comprehensive
assessment of the model’s ability to accurately classify instances, especially in the
context of intrusion detection.
• Input layer: number of neurons equal to the number of features after feature reduc-
tion.
• Hidden layers: multiple hidden layers with varying numbers of neurons. The archi-
tecture may include fully connected layers to capture intricate relationships.
• Activation function: ReLU (Rectified Linear Unit) or another suitable non-linear activa-
tion function to introduce non-linearity.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 12 of 39
• Output layer: number of neurons equal to the number of classes (types of intrusions)
in the dataset, typically using a softmax activation function for classification.
• Loss function: cross-entropy loss, commonly used for classification tasks.
• Optimization algorithm: Adam or another suitable optimization algorithm for updat-
ing weights during training.
Training process:
• The DFNN is trained using the labeled dataset, considering both reduced features and
clustering results.
• Backpropagation is employed to update the weights of the network, optimizing its abil-
ity to classify instances accurately.
• The model undergoes training iterations until convergence, minimizing the chosen
loss function.
Evaluation:
• The performance of the trained DFNN is evaluated on separate test datasets, consid-
ering both full and reduced feature sets.
• Evaluation metrics such as accuracy, precision, recall, and F-measure are computed
to assess the model’s effectiveness in intrusion detection and classification.
In this study, normalization is done in two steps. First, as mentioned in Eq. (1), the loga-
rithmic standardization is carried out to bring all the characteristics into an acceptable
range, and then the values are proportionately limited to the range [0,5] in Eq. (2).
frnorm = log fri + 1 (1)
fri − min(fri )
frnorm = (j − k) (2)
max fri − min(fri )
where j = 0 and k = 5.
Algorithm 1 Demonstrates the normalization and feature encoding and which it can be write as:
KMC‑IG‑based FES
Utilizing K-means clustering, which groups datasets depending on the classification cat-
egory, feature extraction is carried out. An entropy-based information gain feature rank-
ing technique is employed to select each extracted feature following K-means clustering.
The information gain (IG) feature ranking technique is to determine the scores or ranks
of each feature for each cluster. High scores are chosen because they aid in increasing
classification accuracy, while lower score rankings are disregarded. The following for-
mula is used to compute each feature’s IG with respect to each cluster category when x
and y are two random variables:
E(x)andCE(x|y) are the entropy with its condition for uncertainty measuring which
they can be calculated from:
n
E(F x) = −
x∈F x
Prb(x)log 2 (x) (4)
n n
CE(F x|F y) = −
x∈F x
Prb(x)
y∈F y
Prb x | y log 2 Prb x | y (5)
: ( 1 , 2 ,…. )
(1,2, … − 1, )∀
: ℎ
1.
2. = 1: ℎ
= −
= ( [ ], )
Informaition Gain (Cs) ⇾ { | }
. (3), (4) (5)
( ≥ ℎ ℎ ) ℎ
ENN are used to balance data: SMOTE approach is used to M set on the minority
instance in order to balance the dataset using SMOTE. The following formula generates
n artificial instances for every fxi an instance of the M set:
where fxri is an instance that randomly selected in neighbours to instances fxi and it can
be computed by K-nearest neighbours (KNN) technique. η is variable which take ran-
dom values in the interval [0, 1]. If N is the total number of the instances such that every
instance fxi ∈ N has higher various neighbours will be eliminated.
The following stages describe how the ENN operates:
The feature set for the whole feature set is only encoded and normalized using Algo-
rithm 1. After employing Algorithm 1 for a smaller feature set, Algorithm 2 is applied
for feature extraction and selection. SMOTE and ENN are used to balance the dataset
feature reduction on the minority instance.
ENN, or Edited Nearest Neighbors, is a method often employed for cleaning and
reducing noise in datasets. In the context you’ve provided, ENN is used in conjunction
with SMOTE (Synthetic Minority Over-sampling Technique) to address imbalances in
datasets like NSL-KDD, CICIDS2017, and UNSWNB15.
Here’s a step-by-step breakdown:
1. Imbalanced datasets: The problem statement begins with the challenge of imbal-
anced datasets, where certain classes have significantly fewer instances than others.
2. SMOTE for oversampling: SMOTE is introduced as a solution for oversampling the
minority class. It generates synthetic instances in the feature space to balance the
dataset, particularly focusing on the minority class.
3. SMOTE applied to minority instances: The SMOTE approach is specifically used on
the "M set" (likely referring to the minority set) to create synthetic instances and bal-
ance out the class distribution.
4. ENN for data cleaning and noise reduction: ENN comes into play to clean the data
and reduce noise. ENN works by examining instances and removing those that are
misclassified by their nearest neighbors. This helps in refining the dataset and elimi-
nating noisy samples.
5. Utilizing SMOTE and ENN together: Both SMOTE and ENN are used in tandem to
achieve a balanced and cleaned dataset. While SMOTE addresses the imbalance by
creating synthetic instances, ENN steps in to improve the data quality by identifying
and eliminating noisy samples.
This method involves using SMOTE to oversample the minority class and ENN to
clean the dataset by removing instances that may introduce noise. The combination
Behiry and Aly Journal of Big Data (2024) 11:16 Page 16 of 39
of these techniques aims to enhance the performance of a classifier when dealing with
imbalanced datasets.
1. Model generalization:
• The primary goal of any machine learning model, including neural networks, is to
generalize well to unseen data. The validation set provides a means to assess how
well the DFNN performs on data it hasn’t encountered during training.
2. Hyperparameter tuning:
• During the training process, hyperparameters like learning rate, batch size, or the
number of hidden layers are optimized to enhance the model’s performance. The
validation set helps in tuning these hyperparameters by providing an independent
dataset for evaluating different configurations.
3. Preventing overfitting:
• Overfitting occurs when a model learns the training data too well, capturing noise
and specificities that do not generalize. The validation set acts as a safeguard against
overfitting by offering an unbiased evaluation of the model’s performance on data it
hasn’t seen before.
4. Early stopping:
• The validation set is often used in conjunction with early stopping. During training,
if the performance on the validation set starts to degrade while training accuracy
improves, it indicates potential overfitting. Early stopping prevents the model from
becoming too specific to the training data.
5. Model selection:
• In scenarios where multiple models or architectures are being considered, the vali-
dation set aids in comparing their performance. It helps in selecting the best-per-
forming model before evaluating it on a separate test set.
6. Avoiding data leakage:
• The validation set ensures that the model is not inadvertently learning patterns spe-
cific to the test set during training. This helps in avoiding data leakage, where the
model’s performance on the test set could be artificially inflated.
7. Fine-tuning and iterative development:
• As the model evolves through iterative development, the validation set allows for
fine-tuning. Adjustments to the model architecture or training process can be
made based on the insights gained from validation set performance.
8. Ensuring robustness:
Behiry and Aly Journal of Big Data (2024) 11:16 Page 17 of 39
• By evaluating the model on a validation set, researchers can gauge its robustness
across different subsets of the data. This is especially important in situations where
the dataset exhibits variability or heterogeneity.
9. Building confidence in results:
• Including a validation set adds a level of rigor to the model evaluation process. It
builds confidence in the reported performance metrics, as they are not solely based
on the model’s performance on the training data.
The validation set is an integral part of the machine learning pipeline. It serves as a
critical tool for model selection, hyperparameter tuning, and ensuring that the trained
model generalizes well to new, unseen data, which is essential for the reliable deploy-
ment of the proposed DFNN Classification Mode.
DFNN
Deep neural networks (DNNs) have emerged as the preferred technique for address-
ing complicated problems. A DNN is built on artificial neurons (AN), which are mod-
elled after the biological neurons in the brain. The data totalled at the ANN’s input is
determined and sent. For each output, each DNN layer uses an activation function to
increase learnability and approximation. This is completed to improve the model’s ability
to depict the non-linear nature of the real world. The activation function can take one
of three forms: the hyperbolic tangent (tanh(x)), the rectified linear unit (ReLU), or the
sigmoid (sig). The following formula represents each activation function’s mathematical
model:
−1
σsig = 1 + e−x (7)
e2x − 1
tanh(x) = (9)
e2x + 1
The Deep Forward Neural Network (DFNN) Classification Mode in the context of the
paper.
1. Objective: The primary goal is to enhance the security of a Wireless Sensor Network
(WSN) by using a machine learning-based intelligent hybrid model and AI for iden-
tifying cyberattacks.
2. Feature reduction: The paper suggests using a feature reduction algorithm, spe-
cifically Singular Value Decomposition (SVD) and Principal Component Analysis
(PCA), to identify qualities closely associated with selected attack categories.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 18 of 39
In essence, the Deep Forward Neural Network (DFNN) Classification Mode inte-
grates various techniques, including deep learning, clustering, and feature reduction, to
achieve robust intrusion detection and classification in the context of Wireless Sensor
Network security.
Evaluation stage
The evaluation stage focuses on testing the trained DFFNN model and other trained ML
models, which includes an assessment of the proposed approach for binary and multi-class
classification using three datasets of network traffic features. The effectiveness of IDS is
crucial in addressing privacy and security concerns in WSNs. Furthermore, an IDS must
have a low or zero percentage of false alarms in addition to detecting threats. Hence, the
suggested model’s performance is evaluated based on four important parameters, namely:
Behiry and Aly Journal of Big Data (2024) 11:16 Page 19 of 39
Accuracy (ACY), Recall (RE), Precision (PRE), and F1-Score (FS) [39, 55, 56]. The strategy
for evaluating the four metric parameters is represented by the following equations.
CN + CP
(13)
ACY accuracy =
CN + CP + IN + IP
CP
RE(Recall) = (14)
IN + CP
CP
PRE(precision) = (15)
IP + CP
RE × PRE
FS(F 1 − Score) = 2 × (16)
RE + PRE
Where:
• CN (Correct Negative): The instances that are truly negative and are correctly identi-
fied as negative.
• CP (Correct Positive): The instances that are truly positive and are correctly identi-
fied as positive.
• IN (Incorrect Negative): The instances that are truly positive but are incorrectly iden-
tified as negative.
• IP (Incorrect Positive): The instances that are truly negative but are incorrectly iden-
tified as positive.
NSL-KDD 16 Features
CICIDS2017 39 Features
UNSW-NB 15 13 Features
The KDD99 dataset was developed based on the DARPA 1998 dataset and has become
the most widely used dataset for IDSs. However, the presence of duplicate instances in
this dataset can bias classification approaches towards normal examples and hinder their
Behiry and Aly Journal of Big Data (2024) 11:16 Page 20 of 39
a) NSL-KDD dataset
The KDD99 dataset is widely regarded as the most popular dataset for IDSs, which
makes it a benchmark for evaluating the performance of classification techniques.
Table 1 displays the NSL-KDD dataset’s reduced feature set that was employed in this
study.
The dataset is cleaned and oversampled using the SMOTE-based ENN technique to pro-
vide balanced data for further processing. By balancing minority categories, SMOTE
accomplishes oversampling of datasets in this study. Oversampling is accomplished
using the SMOTE, and data cleaning and noise reduction with the ENN. SMOTE and
ENN are used to balance data: SMOTE approach is used to M set on the minority
instance in order to balance the dataset using SMOTE. The following formula generates
n artificial instances for every fxi an instance of the M set:
where fxri is an instance that randomly selected in neighbours to instances fxi and it can
be computed by K-nearest neighbours (KNN) technique. η is variable which take ran-
dom values in the interval [0, 1]. If N is the total number of the instances such that every
instance fxi ∈ N has higher various neighbours will be eliminated.
The following stages describe how the ENN operates:
ct_dst_ltm dttl
stcpb Dwin
is_sm_ips_ports sinpkt
dmean ct_state_ttl
dloss Proto
dtcpb ct_src_dport_ltm
swin
Table 2 displays the distributions for the full reduced features sets.
b) UNSW-NB15 dataset
The dataset was divided into different sections and designed to allow end users to edit
it. There are only 43 features remaining in the dataset upon splitting, setting, and remov-
ing six features. Table 3 displays the UNSW-NB 15 reduced feature set.
Table 4 displays reduced feature utilized for the UNSW-NB15 dataset.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 22 of 39
c) CICIDS2017 dataset
Table 5 displays the reduced feature applied to the CICIDS-2017 dataset in this study.
Table 6 displays the CICIDS2017 data.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 23 of 39
Binary classification
This classification contains confusion matrices, accuracy, F1-score, recall, precision.
The confusion matrices of binary classification for the three datasets are as follows:
If we denote the Normal by N and Anomalous by A. Table 7 indicates the binary con-
fusion matrix for NSL-KDD.
Confusion matrices are an essential tool for evaluating the performance of classifica-
tion models, such as the deep learning-based feed-forward neural network (DLFFNN)
algorithm proposed in the paper. They provide detailed insight into how well the model
is performing in terms of classifying instances into true positives (TP), true negatives
Behiry and Aly Journal of Big Data (2024) 11:16 Page 24 of 39
(TN), false positives (FP), and false negatives (FN). Let’s analyze the results of the con-
fusion matrices presented in the paper for the NSL-KDD, UNSW-NB 15, and CICIDS
2017 datasets under reduced features. Table 8 displays the binary confusion matrix for
UNSW-NB 15.
Table 9 indicates the matrix of binary confusion for CICIDS2017.
Table 10 indicates the confusion matrix of NSL-KDD.
Table 11 indicates the confusion matrix for UNSW-NB 15.
Table 12 displays the confusion matrix for CICIDS2017.
1. NSL-KDD dataset:
True Positives (TP): The algorithm correctly identified 97.8% of the attacks in this
dataset.
True Negatives (TN): The model correctly identified non-attacks, achieving a high rate
of 99.7%.
False Positives (FP): There were very few false alarms, indicating a high precision of
99.8%.
False Negatives (FN): The model missed only 2.2% of the attacks, showing a high recall
of 97.8%.
Overall, the confusion matrix for the NSL-KDD dataset demonstrates excellent per-
formance. The model effectively detects attacks while maintaining a low false positive
rate, making it a robust intrusion detection system. Table 13 indicates the comparison
and contrast of the NSL-KDD dataset.
Table 11 Confusion matrix for UNSW-NB 15
Behiry and Aly Journal of Big Data
Phase Feature
Class N F A B D E G R SC W Class N F A B D E G R SC W
F 1 44,532 1 0 0 0 1 0 0 0 F 1 24,526 0 1 0 1 1 0 0 5
A 0 3 25,321 3 0 0 0 0 0 0 A 0 0 19,834 3 0 0 0 0 0 0
B 0 9 1 13,145 0 4 0 1 0 0 B 0 1 1 17,739 0 4 0 1 0 0
D 0 0 4 0 12,442 6 3 3 0 0 D 0 0 4 0 7349 0 3 0 4 0
E 7 9 0 3 8 9122 0 0 0 0 E 5 3 0 3 0 4673 0 0 0 3
G 0 0 0 1 0 0 7823 0 0 0 G 0 0 0 1 0 0 3865 5 0 0
R 0 5 0 0 0 3 1 6632 0 1 R 0 4 0 0 7 3 1 4423 3 1
SC 0 0 0 0 0 0 6 0 3074 0 SC 0 0 3 0 0 7 3 1 3074 0
W 3 0 1 1 0 4 0 0 4 1845 W 1 0 3 0 4 4 1 1 4 2345
Validation N 13,432 0 5 0 1 3 0 0 1 0 N 13,432 0 5 0 1 3 3 0 0 1
F 1 44,532 1 0 0 0 1 0 0 0 F 1 44,532 1 0 0 0 0 1 0 0
A 0 3 25,321 3 0 0 0 0 0 0 A 0 3 25,321 3 0 0 0 0 0 0
B 0 9 1 13,145 0 4 0 1 0 0 B 0 9 1 13,145 0 4 4 0 1 0
D 0 0 4 0 12,442 6 3 3 0 0 D 0 0 4 0 12,442 6 6 3 3 0
E 7 9 0 3 8 9122 0 0 0 0 E 7 9 0 3 8 9122 0 0 0 0
G 0 0 0 1 0 0 7823 0 0 0 G 0 0 0 1 0 0 7823 0 0 0
R 0 5 0 0 0 3 1 6632 0 1 R 0 5 0 0 0 3 1 6632 0 1
SC 0 0 0 0 0 0 6 0 3074 0 SC 0 0 0 0 0 0 6 0 3074 0
W 3 0 1 1 0 4 0 0 4 3642 W 3 0 1 1 0 4 0 0 4 3642
Page 25 of 39
Behiry and Aly Journal of Big Data
Table 11 (continued)
Phase Feature
(2024) 11:16
Class N F A B D E G R SC W Class N F A B D E G R SC W
DD 0 7 3 17,026 3 0 0 3 0 0 0 3 5 0 9 DD 0 6 2 16,020 3 0 0 2 0 0 0 2 3 0 7
DGE 5 0 0 4 12,311 6 10 0 8 0 0 8 0 10 0 DGE 5 0 0 4 11,201 6 8 0 8 0 0 8 0 10 0
DH 4 0 0 0 11 8461 0 0 5 0 5 0 0 0 0 DH 4 0 0 0 11 7150 0 0 5 0 5 0 0 0 0
DSHT 13 0 9 0 0 1 5438 0 16 1 7 0 0 0 0 DSHT 9 0 8 0 0 1 4327 0 11 1 6 0 0 0 0
DS 0 3 0 0 0 3 0 4221 3 0 0 0 7 0 5 DS 0 2 0 0 0 2 0 3110 2 0 0 0 6 0 3
FP 0 6 0 10 7 7 0 9 9923 0 1 8 8 0 0 FP 0 5 0 9 6 6 0 6 8812 0 1 6 6 0 0
HB 0 1 0 0 0 9 0 1 2 1246 0 3 1 0 1 HB 0 1 0 0 0 8 0 1 2 980 0 2 1 0 1
I 4 0 5 0 0 0 6 0 5 0 1289 0 0 1 3 I 3 0 4 0 0 0 6 0 4 0 1152 0 0 1 2
PS 4 0 0 1 0 1 1 0 4 0 3 2695 3 0 3 PS 3 0 0 1 0 1 1 0 3 0 3 1132 3 0 2
S 1 0 1 0 1 0 3 0 1 0 0 5 2111 4 0 S 1 0 1 0 1 0 2 0 1 0 0 5 1753 4 0
SP 5 0 3 0 0 0 8 5 0 3 5 6 0 98 0 SP 4 0 3 0 0 0 9 6 0 3 6 5 0 1236 0
X 6 0 3 0 0 4 0 14 3 0 8 0 3 1 876 X 5 0 2 0 0 3 0 9 3 0 7 0 3 1 76
Valida- N 14,073 5 0 1 1 0 1 0 5 7 0 0 1 0 5 N 14,073 5 0 1 1 0 1 0 5 7 0 0 1 0 5
tion
B 7 6687 0 0 1 0 0 1 0 0 0 0 5 0 0 B 7 6687 0 0 1 0 0 1 0 0 0 0 5 0 0
BF 0 3 1831 0 0 5 0 6 0 1 0 1 0 0 0 BF 0 3 1831 0 0 5 0 6 0 1 0 1 0 0 0
DD 0 7 3 1695 3 0 0 3 0 0 0 3 5 0 9 DD 0 7 3 1695 3 0 0 3 0 0 0 3 5 0 9
DGE 5 0 0 4 141 6 10 0 8 0 0 8 0 10 0 DGE 5 0 0 4 141 6 10 0 8 0 0 8 0 10 0
DH 4 0 0 0 11 1399 0 0 5 0 5 0 0 0 0 DH 4 0 0 0 11 1399 0 0 5 0 5 0 0 0 0
DSHT 8 0 7 0 0 1 13,988 0 16 1 7 0 0 0 0 DSHT 8 0 7 0 0 1 13,988 0 16 1 7 0 0 0 0
DS 0 3 0 0 0 3 0 2043 3 0 0 0 7 0 5 DS 0 3 0 0 0 3 0 2043 3 0 0 0 7 0 5
FP 0 0 0 8 0 6 0 9 1823 0 1 8 8 0 0 FP 0 0 0 8 0 6 0 9 1823 0 1 8 8 0 0
HB 0 1 0 0 0 9 0 1 2 305 0 3 1 0 1 HB 0 1 0 0 0 9 0 1 2 305 0 3 1 0 1
Page 27 of 39
Table 12 (continued)
Phases Feature
I 0 0 5 0 6 0 4 0 5 0 623 0 0 1 3 I 0 0 5 0 6 0 4 0 5 0 623 0 0 1 3
PS 3 0 0 1 0 1 1 0 6 0 3 874 3 0 3 PS 3 0 0 1 0 1 1 0 6 0 3 874 3 0 3
S 0 0 1 0 0 0 3 0 1 0 0 7 2257 0 5 S 0 0 1 0 0 0 3 0 1 0 0 7 2257 0 5
SP 0 5 3 0 8 0 0 3 1 0 0 6 0 108 0 SP 0 5 3 0 8 0 0 3 1 0 0 6 0 108 0
(2024) 11:16
X 11 0 1 0 0 4 4 1 5 1 0 6 9 0 1967X 11 0 1 0 0 4 4 1 5 1 0 6 9 0 1967
Testing N 11,789 1 0 5 0 2 0 3 1 0 1 4 0 4 3 N 11,789 1 0 5 0 2 0 3 1 0 1 4 0 4 3
B 2 9632 3 0 2 0 6 0 0 0 6 0 0 0 0 B 2 9632 3 0 2 0 6 0 0 0 6 0 0 0 0
BF 0 3 2289 3 0 5 0 6 0 1 0 1 0 0 0 BF 0 3 2289 3 0 5 0 6 0 1 0 1 0 0 0
DD 0 7 3 16,956 3 0 0 3 0 0 0 3 5 0 9 DD 0 7 3 16,956 3 0 0 3 0 0 0 3 5 0 9
DH 4 0 0 0 11 825 0 0 5 0 5 0 0 0 0 DH 4 0 0 0 11 825 0 0 5 0 5 0 0 0 0
DS 0 3 0 0 0 3 0 1985 3 0 0 0 7 0 5 DS 0 3 0 0 0 3 0 1985 3 0 0 0 7 0 5
FP 0 6 0 10 7 7 0 9 432 0 1 8 8 0 0 FP 0 6 0 10 7 7 0 9 432 0 1 8 8 0 0
HB 0 1 0 0 0 9 0 1 2 1543 0 3 1 0 1 HB 0 1 0 0 0 9 0 1 2 1543 0 3 1 0 1
I 4 0 5 0 0 0 6 0 5 0 129 0 0 1 3 I 4 0 5 0 0 0 6 0 5 0 129 0 0 1 3
PS 4 0 0 1 0 1 1 0 4 0 3 1189 3 0 3 PS 4 0 0 1 0 1 1 0 4 0 3 1189 3 0 3
S 1 0 1 0 1 0 3 0 1 0 0 5 108 4 0 S 1 0 1 0 1 0 3 0 1 0 0 5 108 4 0
SP 0 5 3 0 0 0 4 5 1 3 0 2 0 84 0 SP 0 5 3 0 0 0 4 5 1 3 0 2 0 84 0
X 4 1 1 1 5 4 2 1 1 0 5 1 3 1 652 X 4 1 1 1 5 4 2 1 1 0 5 1 3 1 652
Page 28 of 39
Behiry and Aly Journal of Big Data (2024) 11:16 Page 29 of 39
2. UNSW-NB 15 dataset:
True Positives (TP): The algorithm correctly identified 98.4% of the attacks in this
dataset.
True Negatives (TN): The model achieved a high true negative rate of 99.1% for
non-attacks.
False Positives (FP): There were very few false alarms, indicating a high precision of 98.7%.
False Negatives (FN): The model missed only 1.6% of the attacks, showing a high recall
of 98.4%.
The confusion matrix for the UNSW-NB 15 dataset also demonstrates exceptional
performance. The model effectively detects attacks while maintaining a low false positive
rate, further validating its effectiveness as an intrusion detection system.
Multi‑class classification
Various attacks based on the dataset are used to train, validate, and test the multi-class
classification model. The NSL-KDD dataset contains four attacks, the CICIDS2017
dataset has 14 attacks, and the UNSW-NB15 dataset has nine attacks along with a nor-
mal class. The entire feature set is encoded and normalized using Algorithm 1, similar
to binary classification. Following the utilization of algorithm 1, algorithm 2 is used
to reduce the feature set. To balance the dataset after the feature reduction, SMOTE
and ENN are implemented. Confusion matrices for the NSL-KDD and UNSW-NB15
Behiry and Aly Journal of Big Data (2024) 11:16 Page 30 of 39
are presented in the tables below. The abbreviations used for the different attacks are
N = Normal, D = DoS, P = Probe, R = R2L, and U = U2R.
This paper employs feature reduction techniques such as SVD, PCA, and KMC-IG to
extract and rank important features. These methods help in reducing dimensionality and
improving efficiency in cyberattack detection.
Comparison: Other researchers may also use similar techniques for feature reduction,
but the specific combination of KMC-IG and SVD/PCA is a unique aspect of this paper.
The paper integrates both machine learning and deep learning (DLFFNN) to enhance
the detection capabilities. It combines the strengths of both approaches to achieve high
accuracy.
Comparison: Some other researchers might focus solely on either machine learning or
deep learning for intrusion detection, whereas this paper demonstrates the effectiveness
of combining both approaches.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 31 of 39
Dataset evaluation
The study evaluates the proposed algorithm using three distinct datasets, providing a
comprehensive assessment of its performance under various conditions.
Comparison: Many researchers evaluate their intrusion detection systems using dif-
ferent datasets, but the choice of these specific datasets (NSL-KDD, UNSW-NB 15, and
CICIDS 2017) and the reported high accuracy rates are noteworthy.
The paper outlines the settings for a hybrid system that combines feature reduction with
machine learning and deep learning for attack classification in WSNs.
Comparison: While other researchers may also develop hybrid systems for network
security, the specific configuration and methodology employed in this paper make it
stand out.
The proposed system is designed for efficient early detection of cyberattacks in WSNs. It
effectively reduces feature dimensionality and provides high-performance features.
Comparison. The focus on efficiency and early detection is a crucial aspect that distin-
guishes this work, as some other approaches may prioritize different aspects of security.
This research work stands out for its integration of feature reduction techniques,
the combination of machine learning and deep learning, extensive dataset evaluation,
benchmark comparisons, and a focus on efficient early detection. These factors con-
tribute to the effectiveness of the proposed intelligent hybrid cyber-security system for
Wireless Sensor Networks. Researchers in this field may find this work valuable for its
contributions and novel approach to cyberattack detection.
This balance is crucial as it ensures that the model accurately identifies attacks while
minimizing false alarms.
Practical implications
A discussion on the practical implications of the research is valuable. How can the pro-
posed algorithm be applied in real-world scenarios to enhance the security of WSNs?
Are there any limitations or challenges in implementing the system?
Future directions
The discussion section can suggest potential future research directions, such as explor-
ing the scalability of the algorithm for larger WSNs, investigating the impact of evolving
cyber threats, or exploring real-time implementation in WSN environments.
Fig. 2 Comparison between the proposed approach and traditional methods for the NSL-KDD dataset
Behiry and Aly Journal of Big Data (2024) 11:16 Page 34 of 39
Fig. 3 Comparison between both the proposed and traditional methods for CICIDS2017 dataset
Fig. 4 Comparison between both the proposed and traditional methods for the UNSW-NB15 dataset
Behiry and Aly Journal of Big Data (2024) 11:16 Page 35 of 39
1. Feature representation:
• The proposed method may employ a more effective feature representation or
extraction technique compared to traditional algorithms. If the features used by the
Behiry and Aly Journal of Big Data (2024) 11:16 Page 36 of 39
model better capture the underlying patterns in the data, it can lead to improved
performance.
2. Complexity and non-linearity:
• Neural networks, including CNNs and ANNs, are capable of capturing complex
and non-linear relationships in data. If the problem at hand involves intricate pat-
terns or dependencies, a deep learning approach may have an advantage over linear
models like SVM and Naive Bayes.
3. Data imbalances:
• Traditional algorithms, including SVM and NB, may struggle with imbalanced
datasets. If the dataset used for evaluation is imbalanced, the proposed method
might incorporate techniques to handle this imbalance, giving it an edge in perfor-
mance.
4. Hybrid approaches:
• The proposed method could be a hybrid model that combines the strengths of mul-
tiple algorithms. Hybrid models are designed to leverage the advantages of different
techniques, potentially resulting in improved performance over individual models.
5. Synthetic data generation:
• If the proposed method employs techniques like Synthetic Minority Over-sampling
Technique (SMOTE) or other data augmentation methods, it can enhance the
model’s ability to generalize and detect minority classes, which may be challenging
for traditional algorithms.
6. Architecture design:
• The architecture of the proposed model, especially in the case of CNNs or ANNs,
might be designed to capture specific domain knowledge or features that are criti-
cal for intrusion detection. This tailored architecture can contribute to better per-
formance.
7. Ensemble methods:
• The proposed method could use ensemble learning, combining multiple models
to make predictions. Ensemble methods often lead to more robust and accurate
results compared to individual models.
8. Adaptability to domain-specific features:
• If the proposed method is designed with a deep understanding of the domain and
specific characteristics of intrusion detection, it may be better suited to handle the
nuances of the problem compared to more generic algorithms.
It’s important to note that the effectiveness of a method depends on various factors,
including the dataset, problem complexity, and the specific design choices made in each
algorithm. Without detailed information on the proposed method’s architecture, fea-
tures, and evaluation metrics, it’s challenging to provide a more specific explanation for
its superior performance over SVM and NB in the context of intrusion detection.
Conclusion
This study focused on three key datasets: NSL-KDD, UNSW-NB 15, and CICIDS
2017, and evaluated the accuracy, precision, recall, and F-measure of the proposed
approach under two different scenarios: full features and reduced features. The proposed
Behiry and Aly Journal of Big Data (2024) 11:16 Page 37 of 39
Author contributions
The researchers cooperated together in proposing the idea of research, implementation, and conducting the research.
The research was written by the first researcher, Mohamed Behairy, and the second researcher, Mohamed Ali, reviewed.
Funding
Open access funding provided by The Science, Technology & Innovation Funding Authority (STDF) in cooperation with
The Egyptian Knowledge Bank (EKB).
Declarations
Ethics approval and consent to participate
Not applicable.
Competing interests
The authors declare that they have no competing interests.
References
1. Saber AM, Behiry MH, Amin M. Real-time optimization for an AVR system using enhanced Harris Hawk and IIoT. Stud
Inform Control. 2022;31(2):81–94.
2. Behiry MH, Amin M, Sauber AM. IIoT-based automatic FOPID tuning for AVR systems using a customized chaotic whale
ttps://www.doi-i.org/journals/view/373.
optimization. h
3. Saini GK, Halgamuge MN, Sharma P, Purkis JS. A review on cyberattacks. In: Shaikh RA, editor. Secure cyber-physical
systems for smart cities. Pennsylvania: IGI Global; 2019. p. 183–219. https://doi.org/10.4018/978-1-5225-7189-6.ch008.
4. Chelli K. Security issues in wireless sensor networks: attacks and countermeasures. Proceedings of the World Congress
on Engineering, Vol I, London, UK. 2015.
5. He D, Chan S, Guizani M. Cyber security analysis and protection of wireless sensor networks for smart grid monitoring.
IEEE Wirel Commun. 2017;24(6):98–103. https://doi.org/10.1109/MWC.2017.1600283WC.
6. Padmavathi G, Shanmugapriya D. A survey of attacks, security mechanisms and challenges in wireless sensor networks.
(Cornell University). 2009. https://arxiv.org/pdf/0909.0576.
7. Pathan AK, Lee H-W, Hong CS. Security in wireless sensor networks: issues and challenges. Proc. ICACT 2006; 1, 20–22:
1043–1048.
8. Perrig A, Stankovic J, Wagner D. Security in wireless sensor networks. Commun ACM. 2004;47(6):53–7.
9. Jian-hua LI. Cyber security meets artificial intelligence: a survey. Front Inf Technol Electron Eng. 2018;19:1462–74.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 38 of 39
10. Handa A, Sharma A, Shukla SK. Machine learning in cybersecurity: a review. WIREs Data Mining Knowl Discov. 2019.
https://doi.org/10.1002/widm.1306.
11. Thomas T, Vijayaraghavan AP, Emmanuel S. Machine learning approaches in cyber security analytics. eBook, Springer
Nature Singapore 2020.
12. Saini GK, Halgamuge MN, Sharma P, Purkis JS. A review on cyberattacks: security threats and solution techniques for
different applications. In: Shaikh RA, editor. Secure cyber-physical systems for smart cities. Pennyslvania: IGI Global; 2019.
p. 183–219. https://doi.org/10.4018/978-1-5225-7189-6.ch008.
13. Boussi GO, Gupta H. A proposed framework for controlling cyber-crime. 8th International Conference on Reliability.
Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO) 2020; pp.1060–1063.
14. Kumar G, Kumar K, Sachdeva M. The use of artificial intelligence-based techniques for intrusion detection: a review. Artif
Intell Rev. 2010;34(4):369–87.
15. Johri P, Verma JK, Paul S, editors. Applications of machine learning. Singapore: Springer Singapore; 2020.
16. Saleem S, Ullah S, Yoo HS. On the security issues in wireless body area networks. Int J Digit Content Technol Appl. 2009.
https://doi.org/10.4156/jdcta.vol3.issue3.22.
17. Sharma K, Ghose MK. Wireless sensor networks: an overview on its security threats. IJCA Special Issue on Mobile Adhoc
Networks 2010.
18. Martins D, Guyennet H. Wireless sensor network attacks and security mechanisms: a short survey 2010; IEEE.
19. Sastry AS, Sulthana S, Vagdevi S. Security threats in wireless sensor networks in each layer. Int J Adv Netw Appl.
2013;4(4):1657–61.
20. Kaplantzis S. Security models for wireless sensor networks 2006. http://members.iinet.com.au/~souvla/transferfinal-rev.
pdf.
21. Karlof C, Wagner D. Secure routing in wireless sensor networks: attacks and countermeasures. Ad Hoc Netw J.
2003;1(2–3):293–315.
22. Yu Y, Li K, Zhou W, Li P. Trust mechanisms in wireless sensor networks: attack analysis and countermeasures. J Netw
Comput Appl. Elsevier, 2011.
23. Xu W, et al. The feasibility of launching and detecting jamming attacks in wireless networks. MobiHoc ’05: Proc. 6th ACM
Int. Symp. Mobile Ad Hoc Net. and Comp 2005; pp. 46–57.
24. Xu W, Trappe W, Zhang Y. Channel surfing: defending wireless sensor networks from interference. In Proc. of Information
Processing in Sensor Networks 2007.
25. Sohrabi K, Gao J, Ailawadhi V, Pottie GJ. Protocols for self-organization of a wireless sensor network. IEEE Pers Commun.
2000;7:16–27.
26. Raymond DR, Midkiff SF. Denial-of-service in wireless sensor networks: attacks and defenses. IEEE Pervasive Comput.
2008;7(1):74–81.
27. Parno B, Perrig A, Gligor V. distributed detection of node replication attacks in sensor networks. Proceedings of the IEEE
Symposium on Security and Privacy (S&P’05) 2005.
28. Xiao Y, Rayi VK, Sun B, Du X, Hu F, Galloway M. A survey of key management schemes in wireless sensor networks. Com-
put Commun. 2007;30(11–12):2314–41.
29. Jain A, Kant K, Tripathy MR. Security solutions for wireless sensor networks. In: Second international conference on
advanced computing & communication technologies. 2012. https://doi.org/10.1109/acct.2012.102.
30. Burgner DE, Luay A. Wahsheh security of wireless sensor networks. In: Eighth international conference on information
technology: new generations. 2011, pp. 315–20. https://doi.org/10.1109/ITNG.2011.62.
31. Zhu S, Setia S, Jajodia S. LEAP: efficient security mechanisms for large-scale distributed sensor networks. Proceedings of
the 10th ACM Conference on Computer and Communications Security (CCS ’03) 2003; 62–72
32. Culler DE, Hong W. Wireless sensor networks. Commun ACM. 2004;47(6):30–3.
33. Makhija J, Appu Shetty A, Bangera A. Classification of attacks on MQTT-Based IoT system using machine learning tech-
niques. Part of the Advances in Intelligent Systems and Computing book series 2021; vol. 1394, 29.
34. Ma T, Wang F, Cheng J, Yu Y, Chen X. A hybrid spectral clustering and deep neural network ensemble algorithm for intru-
sion detection in sensor networks. Sensors. 2016;16(10):1701.
35. Zhang F, Chan PP, Biggio B, Yeung DS, Roli F. Adversarial feature selection against evasion attacks. IEEE Trans Cybern.
2015;46(3):766–77.
36. Tavallaee M, Bagheri E, Lu W, Ghorbani AA. A detailed analysis of the KDD CUP 99 data set. In IEEE symposium on com-
putational intelligence for security and defense applications 2009; 1–6.
37. Sonule AR, Kalla M, Jain A, Chouhan DS. UNSWNB15 dataset and machine learning based intrusion detection systems.
Int J Eng Adv Technol (IJEAT). 2020;9(3):2638–48.
38. Sharafaldin I, Lashkari AH, Ghorbani AA. Toward generating a new intrusion detection dataset and intrusion traffic
characterization. ICISSp. 2018;1:108–16.
39. Aly M, Alotaibi AS. Molecular property prediction of modified gedunin using machine learning. Molecules. 2023;28:1125.
https://doi.org/10.3390/molecules28031125.
40. Johri P, Verma JK, Paul S. Applications of machine learning. Algorithms for Intelligent Systems. eBook, Springer, Nature
Singapore. 2020. https://doi.org/10.1007/978-981-15-3357-0
41. AlRikabi HT, Hazim HT. Enhanced data security of communication system using combined encryption and steganogra-
phy. Int J Interact Mobile Technol. 2021. https://doi.org/10.3991/ijim.v15i16.24557.
42. Ahmad R, Wazirali R, Abu-Ain T. Machine learning for wireless sensor networks security. An overview of challenges and
issues. Sensors. 2022;22:4730.
43. Ismail S, Khoei TT, Marsh R, Kaabouch N. A comparative study of machine learning models for cyber-attacks detection
in wireless sensor networks. In Proceedings of the 2021 IEEE 12th Annual Ubiquitous Computing, Electronics & Mobile
Communication Conference (UEMCON), New York, NY, USA, 1–4 December 2021; pp. 313–318.
44. Khoei TT, Ismail S, Kaabouch N. Dynamic selection techniques for detecting GPS spoofing attacks on UAVs. Sensors.
2022;22:662.
45. Karatas G. Increasing the performance of machine learning-based IDSs on an imbalanced and up-to-date dataset. IEEE
Access. 2020;8:32150–62.
Behiry and Aly Journal of Big Data (2024) 11:16 Page 39 of 39
46. Dong L, Li Y, Gao W. A survey on wireless sensor network security: attacks and defenses. IEEE Access. 2020;8:14237–58.
47. Zhang R, Zhang C, Zhang C. A comprehensive review of wireless sensor network security: real attacks, existing protocols,
and open research issues. IEEE Access. 2021;9:12461–86.
48. Li X, Wu D, Zhang X. Machine learning in wireless sensor networks: algorithms, applications, and challenges. Futur Gener
Comput Syst. 2022;128:131–48.
49. Wang Z, Ma Y, Jiang B. An intrusion detection system based on deep learning for wireless sensor networks. IEEE Internet
Things J. 2021;8(15):12194–203.
50. Chen J, Zhang S, Ma Y. Explainable deep learning for intrusion detection in wireless sensor networks. Ad Hoc Netw.
2023;128: 102933.
51. Kim J, Park S. Clustering-based anomaly detection for wireless sensor networks. Inf Sci. 2020;507:54–66.
52. Jingjing Z, Tongyu Y, Zhang J, Zhang G, Li X, Peng X. Intrusion detection model for wireless sensor networks based on
MC-GRU. Wirel Commun Mob Comput. 2022;2022:1–11. https://doi.org/10.1155/2022/2448010.
53. Zhao Y, Li Y, Zhang L. Benchmarking intrusion detection systems in wireless sensor networks: a comprehensive review.
Ad Hoc Netw. 2023;128: 102917.
54. Liu Z, Zhang Y, Zhang Y. Trade-off between accuracy and resource consumption in intrusion detection systems for wire-
less sensor networks. IEEE Internet Things J. 2021;8(24):19589–600.
55. Aly M, Alotaibi NS. A new model to detect COVID-19 coughing and breathing sound symptoms classification from CQT
and Mel spectrogram image representation using deep learning. Int J Adv Comput Sci Appl. 2022. https://doi.org/10.
14569/IJACSA.2022.0130869.
56. Aly M, Alotaibi NS. A novel deep learning model to detect COVID-19 based on wavelet features extracted from Mel-
scale spectrogram of patients cough and breathing sounds. Inform Med Unlocked. 2022;32: 101049.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.