Adv Net Attack Detection (DoS, Probe) Gemini AI Generated Index From DATAPRO
Adv Net Attack Detection (DoS, Probe) Gemini AI Generated Index From DATAPRO
patterns indicative of malicious activities, adapting to new attack forms that traditional firewalls
and antivirus software might miss. This capability is particularly vital given the dynamic nature
1
of network traffic and the constant emergence of new threats in environments like the Internet of
Things (IoT). DL models, such as Convolutional Neural Networks (CNNs) and Long Short-Term
5
Memory (LSTMs), can automatically extract and enhance features from raw network traffic data,
overcoming the limitations and labour costs associated with manually extracted features. This 2
approach allows for more robust detection of both known and unknown threats, significantly
enhancing network security measures. 2
Other studies using NSL-KDD reported excellent detection accuracy with Random Forest (RF)
and Support Vector Machine (SVM). A comparative analysis on DDoS detection revealed that
15
Probe attacks, often categorized as reconnaissance or port scanning, are also a significant focus.
Machine learning techniques like Naïve Bayes, SVM, and Decision Trees have been applied to
datasets like KDDcup99 for Probe attack detection. For port scanning attacks, a Naive Bayes
7
model achieved 93.5% accuracy. F1-scores for probe detection have been reported as high as
15
98%. 16
Genetic Algorithms (GAs) have garnered significant attention for feature selection in
NIDS due to their adaptability and suitability for dynamic rule generation. Studies have shown
17
GAs can effectively reduce the number of features while maintaining or improving F1-scores. For
example, a GA achieved an F1-score of 76.89% with 40 features on the HIKARI-2021 dataset,
outperforming Chi-squared, PCA, and LDA for feature selection. However, it is important to
17
note that the effectiveness of a feature selection algorithm, such as GA, is not universally superior
but can be contingent on the specific machine learning algorithm it is paired with and the
characteristics of the dataset. For instance, Particle Swarm Optimization (PSO) achieved better
accuracy and precision than GA for J48 and SVM classifiers in one study. this highlights that
17
for determining the score values used in feature selection techniques, such as Information Gain, is
frequently not explicitly mentioned in existing literature. This lack of transparency and
9
standardization impedes the replicability, realism, and robustness of proposed NIDS methods. 3
Without clear methodologies for data validation, pre-processing, and feature selection, comparing
and building upon existing research becomes problematic. This project aims to contribute to
addressing this by providing detailed justifications for its methodological choices, thereby
enhancing the transparency and reproducibility of NIDS research, which in turn improves its
practical applicability and scientific rigor.
of traffic with a large number of features, making it suitable for evaluating the performance of
intrusion detection systems against contemporary cyber threats. Its comprehensive nature and
9
detailed attack categorization make it a valuable resource for training and testing machine
learning models for network security. 2
includes 9 different types of cyber-attacks, notably DoS and Reconnaissance (which encompasses
Probe attacks), alongside normal traffic. The dataset contains over 2.5 million records and 45
1
features, categorized into basic, flow, time, content, and additional generated features. Its design
6
supports the advancement and assessment of machine learning techniques for identifying and
categorizing diverse forms of network attacks. 1
3.3 Data Collection and Feature Description:-
Network intrusion detection datasets are typically derived from monitoring network
activity by inspecting its traffic. This involves collecting raw packet data, flow records, or logs. 3
Features extracted from this data encompass a wide range of attributes, such as protocol type
(e.g., TCP, UDP, ICMP), service (e.g., HTTP, FTP), connection flags, source and destination IP
addresses and port numbers, packet and byte counts, flow inter-arrival times, and various
statistical measures of network behavior. 12
These datasets are often characterized by high
dimensionality, containing numerous features that describe network connections and events. 6
handled, can lead to errors or biased results in machine learning models. Redundant records, often
present in older datasets like KDD Cup 99, can skew training and evaluation metrics. Outliers,
1
which are data points significantly different from others, can also negatively impact model
performance and need to be either removed or transformed. Effective data cleaning ensures that
1
the model learns from accurate and representative data, thereby improving its detection
capabilities.
the array signifies whether a specific feature is retained (1) or removed (0). 17
Fitness Function: A fitness function is defined to evaluate the quality of each feature
subset. This function typically measures the performance of the machine learning model
(e.g., 1DCNN in this study) when trained using only the features selected by that
particular solution. Common metrics for fitness include model accuracy or F1-score.
17
Selection: Based on their fitness scores, fitter individuals are probabilistically chosen to
become "parents" for the next generation. Methods like the roulette wheel approach are
often employed, where individuals with higher fitness have a greater chance of being
selected. 17
Crossover and Mutation: New offspring are generated from the selected parents through
genetic operators:
Crossover: Two parent solutions are combined to produce new offspring, mixing their
feature selections. A common technique is Single-Point Crossover, where a random point
is chosen, and segments of the parent chromosomes are swapped. 17
Sparsity Check: A sparsity check function can be implemented to ensure that the
generated offspring adhere to desired feature reduction levels, adjusting the feature mask
if necessary. 17
The benefits of using GAs include high re-trainability, enhanced adaptability, and
suitability for dynamic rule generation in Intrusion Detection Systems. While the runtime for GA
17
to find an optimal feature set can be longer than other methods, this is typically an offline process
and does not impact the real-time classification performance of the final system. The reduction in
17
the number of features selected by GA can significantly decrease the processing time for
classification without substantially impacting the F1-score. For instance, GA has been shown to
17
achieve high F1-scores (e.g., 76.89% with 40 features) in comparison to other feature selection
techniques. This approach, by reducing dependence on manual feature engineering, makes NIDS
17
development more efficient, scalable, and potentially more robust against evolving attack
patterns, as it can adaptively discover new relevant features.
approximately 87.35% of the entire dataset, with attack types making up the remaining 12.65%. 6
Similarly, CICIDS2017 also presents a high class imbalance, where certain attack types, such as
Infiltration, have very small data portions even after initial relabeling. This imbalance is a critical
9
issue because it can lead to biased model predictions, where the classifier performs well on the
majority (normal) class but poorly on minority (attack) classes, resulting in low detection
accuracy for actual threats. 6
To mitigate this problem, various techniques are employed. Oversampling methods, such
as Adaptive Synthetic (ADASYN) or Synthetic Minority Over-sampling Technique (SMOTE),
are used to generate synthetic samples for minority classes, thereby increasing their representation
in the training data. Conversely, undersampling methods, like cluster-based undersampling using
6
K-means, reduce the number of samples in the majority class. A combined approach of
6
oversampling and undersampling can be used to balance the dataset without excessively
increasing its size. Additionally, during model training, class weights can be computed and
6
applied to the loss function, giving higher importance to the minority classes and guiding the
model to pay more attention to their correct classification.
ensures that all features contribute equally to the model's learning process, preventing features
with larger values from dominating the optimization, thereby improving model convergence and
overall performance. 26
Checking for Attack Representation: Ensuring that DoS and Probe attack instances are
adequately represented and correctly labeled within the dataset. DoS attacks are
characterized by unusually high connection rates or traffic volumes from single or
multiple sources targeting a victim, aiming to exhaust resources. Probe attacks, such as
8
port scans (e.g., Nmap, Portsweep) and IP sweeps, are reconnaissance activities where
attackers scan networks to identify active hosts, open ports, and vulnerabilities. 7
Handling Missing Values and Outliers: As a general preprocessing step, missing values
and outliers are identified and removed or imputed, as they can corrupt the learning
process. 1
Class Imbalance Assessment: UNSW-NB15 has a serious class imbalance, with a large
majority of normal traffic. Validation involves confirming the distribution of DoS and
6
Probe attacks relative to normal traffic and other attack types to inform subsequent
balancing strategies.
Feature Consistency: Ensuring that features relevant to DoS (e.g., packet counts, flow
duration) and Probe (e.g., destination port counts, unique services) are consistent and
accurately reflect the attack behavior. For DoS, validation might involve checking for
unusually high connection rates from single/multiple sources to a target. For Probe, it
might involve identifying systematic port scanning patterns. 7
redundant feature, "Fwd Header Length," is removed, leaving 77 features for analysis. 9
Relabeling for Imbalance: To address the high class imbalance, a new labeling scheme
for attack traffic is introduced. This relabeling helps in grouping similar attacks or
9
focusing on specific ones. However, even after relabeling, some attack types, like
Infiltration, might still have very small data portions, making their detection challenging
for classifiers. 9
Feature Validation for Attack Types: Feature selection techniques like Information
Gain are applied to identify relevant features. Validation checks if these selected features
9
are indeed significant for distinguishing DoS and Port Scan attacks. For example, for DoS
in CICIDS2017, validation might involve examining features related to flow duration,
packet lengths, and flag counts , ensuring they align with known DoS characteristics. For
14
Port Scan, it would focus on features indicating systematic scanning of ports. It has been
8
observed that some classifiers struggle to detect minority classes like Infiltration or Web
Attack even after feature selection, indicating that data volume for these specific attack
types remains a challenge. 9
The detailed data validation process for specific attack types on UNSW_NB15 and
CICIDS2017 is not just about cleaning data, but about ensuring the representativeness and quality
of the attack patterns for model training. The challenge of small data portions for certain attacks
(e.g., Infiltration in CICIDS2017) even after relabeling, means that even with robust pre-
processing, and some attack types remain harder to detect. This directly impacts the model's
performance and the need for careful interpretation of results for minority classes. This highlights
a fundamental limitation of supervised learning in NIDS: its performance is heavily constrained
by the diversity and volume of labeled attack data. For real-world systems, this implies that
models might need to be complemented by anomaly detection techniques or continuous learning
to identify truly novel or rare attacks not well-represented in training data.
The following table provides a consolidated view of key data properties and validation
considerations for the UNSW_NB15 and CICIDS2017 datasets, offering a quick reference for
understanding their suitability and inherent challenges for DoS and Probe detection.
Table 3.1: Key Data Properties and Validation Considerations for UNSW_NB15 and
CICIDS2017
Property UNSW_NB15 CICIDS2017 Validation
Considerations
Year of 2015 20 2017 9 Determines captured threat
Collection landscape and traffic
27
characteristics.
Focus/ General-purpose, designed to General-purpose, realistic Suitability for modern
Scenario address KDD Cup 99 traffic from various threats; generalizability
1 9
limitations. scenarios. across network types.3
Type behavior.
Attack 9 types: DoS, (Probe) 14 types: DoS/DDoS, Port Coverage of specific attack
Traffic Reconnaissance, Fuzzers, Scan (Probe), Brute Force, vectors; diversity of attack
Anony Not explicitly detailed in Not explicitly detailed in Impact on feature utility
mization snippets, but common for public snippets. and privacy concerns.
Status datasets.
Data 2,540,044 records (2,218,761 High volume, 8 CSV files.9 Sufficient data for deep
Volume normal, 321,283 attack).20 learning, but manageability
is key.
Train/Test/ 70% train, 20% test, 10% 70% train, 30% test (for 20% Ensures unbiased model
6 9
Validation validation (entire dataset used). of data). evaluation.
Split
Labels Yes (categorized attacks and Yes ("Benign" and Essential for supervised
Provided normal). 20
"Attacks" with 14 types). learning. 9
General Cleaning (missing values, Combining files, cleaning, Ensures data quality and
Validation outliers), numericalization, relabeling, feature model effectiveness.
Approaches normalization, feature selection (e.g., Information
selection. 1
Gain). 9
Chapter 4: Methodology and Model Development
engineering is not only labour-intensive but can also lead to the loss of crucial information from
the original data, potentially resulting in lower detection accuracy and hindering practical
deployment. 13
The ability of 1DCNNs to learn relevant patterns from the raw data streams effectively addresses
the complexity and diversity inherent in network data, enhancing classifier accuracy by reducing
feature redundancy and mitigating issues related to class imbalance and the detection of rare
attacks. This architectural choice represents a strategic advancement over traditional ML
22
Dropout Layers: Dropout layers are strategically placed after pooling layers (e.g., with a
parameter of 0.2) to prevent overfitting by randomly deactivating a fraction of neurons
during training, thereby improving the model's generalization capability. 6
Flatten Layer: This layer transforms the 2D output of the convolutional and pooling
layers into a 1D vector, preparing it for the fully connected dense layers.
Dense (Fully Connected) Layers: One or more dense layers process the flattened
features, learning complex non-linear relationships.
Output Layer: The final dense layer, typically with a Softmax activation function,
outputs the probability distribution over the different attack classes (e.g., Normal, DoS,
Probe, Other Attacks). 6
predicted probability distribution and the true class distribution, aiming to minimize this
divergence during training.
Optimization techniques are employed to adjust the model's weights and biases to
minimize the chosen loss function. Adam (Adaptive Moment Estimation) and RAdam
(Rectified Adam) are popular adaptive learning rate optimization algorithms. Adam combines the
benefits of AdaGrad and RMSProp, maintaining per-parameter learning rates that are adapted
based on the first and second moments of the gradients. RAdam further improves upon Adam by
rectifying the variance of the adaptive learning rate, which can lead to better convergence and
generalization, especially during the initial stages of training. The "nadam" optimizer, a variant of
Adam incorporating Nesterov momentum, has also been noted for its effectiveness in similar
intrusion detection contexts. 6
rate schedules (e.g., step decay, exponential decay, or cosine annealing) can be implemented to
further control the learning rate's progression over epochs, potentially leading to faster
convergence and improved model performance.
weights, the validation set monitors performance during training to prevent overfitting and guide
hyperparameter tuning, and the test set provides an unbiased evaluation of the final model's
performance on unseen data.
Hyperparameter tuning is the process of selecting the optimal values for parameters that
control the learning process itself, rather than being learned by the model. These include batch
size (e.g., 256, representing the number of samples processed before the model's internal
parameters are updated), the number of epochs (e.g., 100 to 200, representing the number of
complete passes through the training dataset), and the learning rate. 6
Techniques for
hyperparameter tuning range from manual search and grid search to more advanced methods like
random search or Bayesian optimization, aiming to find the combination of hyperparameters that
yields the best performance on the validation set.
in such scenarios, as a model might achieve high accuracy by simply classifying all instances as
the majority class (normal traffic) while failing to detect actual attacks. Therefore, a nuanced
29
correctness.
Formula: Accuracy = (TP + TN) / (TP + TN + FP + FN)
Precision: Proportion of true positive predictions among all instances predicted as
positive. High precision indicates a low rate of false positives, meaning fewer legitimate
6
activities are incorrectly flagged as attacks, which is crucial for preventing alert fatigue in
security operations.
Formula: Precision = TP / (TP + FP)
Recall (Sensitivity or True Positive Rate): The proportion of true positive predictions
among all actual positive instances. High recall indicates the model's ability to correctly
6
identify a large proportion of actual attacks, minimizing false negatives (missed attacks)
which can lead to severe security breaches.
Formula: Recall = TP / (TP + FN)
F1-Score: The harmonic mean of precision and recall. It provides a balanced measure of
6
indicates a good balance between identifying positive cases and minimizing false
positives and false negatives. 31
Accuracy The ratio of correctly predicted instances to the total (TP + TN) / (TP + TN +
number of predictions. 7
FP + FN)
Precision The fraction of records predicted as positive that are TP / (TP + FP)
actually positive. 7
Recall The fraction of actual positive samples that are TP / (TP + FN)
correctly predicted by the classifier. 7
F1-Score The harmonic mean of Precision and Recall, providing 2 * (Precision * Recall) /
a balanced measure of performance. 31
(Precision + Recall)
Once serialized, the model can be loaded back into memory by the Flask application
when it starts up. This eliminates the need to retrain the model every time a prediction is
requested, significantly reducing latency and computational overhead. The Flask application then
uses this loaded model to perform inference on incoming network data, making the machine
learning solution accessible and practical for real-world applications. 32
templates/: A directory storing HTML files (.html) that define the structure and content of
the web pages, such as the home page, upload forms, and results display. Flask uses the
Jinja2 templating engine to render these files dynamically. 34
static/: A directory for static assets like CSS files (.css) for styling, JavaScript files (.js)
for interactivity (e.g., Chart.js for visualizations), and images.
This modular structure promotes organized development and maintainability.
Simplicity and Clarity: The primary goal is to provide a clear and concise overview of
key performance indicators (KPIs) and metrics, avoiding clutter and using clear, jargon-
free language. 36
Visual Hierarchy: Critical security insights should be prioritized and stand out. This can
be achieved through strategic use of size, position, and a limited color palette (e.g., red for
critical, yellow for warning, green for safe). 36
Real-time Updates: Cybersecurity dashboards should process and display live data
streams in real-time, as even a few seconds of delay can mean missed threats. 37
Interactivity: Features like filters, drill-down capabilities, and interactive charts allow
users to explore data in more detail, select specific data ranges, or categorize
information. 36
Adhering to these principles ensures that the GUI is not just functional but also enhances
the human decision-making process, leading to more effective incident assessment and
response.40
2. Flask Processing: The Flask application receives the request, parses the incoming JSON
data, and performs any necessary pre-processing to format the data as expected by the
loaded 1DCNN model.
3. Model Inference: The pre-processed data is then fed to the serialized 1DCNN model,
which generates a prediction (e.g., "DoS Attack," "Probe Attack," or "Normal") along with
a confidence score.
4. Response Generation: Flask constructs an HTTP response, typically in JSON format,
containing the prediction result and its associated confidence score. This response is then
sent back to the client. 32
API keys) is crucial to restrict access to authorized users or systems, preventing unauthorized use
of the prediction service. 41
principles include:
Threat Prioritization: Active threats are highlighted and categorized by severity (e.g.,
Critical, High, Medium, Low) using clear visual cues like color-coding. This ensures that
37
security analysts can immediately identify and respond to the most pressing incidents.
Visual Hierarchy: A clear visual hierarchy ensures that the most critical data stands out,
while secondary information remains accessible but does not clutter the interface. Limited
37
color palettes and effective use of white space improve readability and reduce visual
noise. 36
Interactive Visualizations: The dashboard employs various interactive charts and graphs
(e.g., line charts for trends, bar charts for attack distribution, heat maps for correlations,
attack path visualizations) to represent network data and attack patterns. 36
These
visualizations facilitate pattern detection and relationship identification in large, complex
datasets. 40
Filtering and Drill-Down: Users can apply filters to select specific data ranges or
categories (e.g., attack type, source IP, time period) and utilize drill-down capabilities to
view more detailed data about specific alerts or incidents. This allows for granular
36
investigation.
Real-time Processing: The system is optimized for real-time data processing and updates,
ensuring that alerts and visualizations reflect the current network state with minimal
delay. WebSockets can be employed to enable push-based visualization and automatic
37
refresh. 38
Visualization of Prediction Outcomes and Confidence Scores:-
The GUI clearly displays the prediction outcomes (e.g., "DoS Attack Detected," "Probe
Attack Detected," "Normal Traffic"). A crucial feature is the visualization of confidence scores
for each prediction. A confidence score, typically a decimal number between 0 and 1, represents
the likelihood that the machine learning model's output is correct. 30
Visual Cues: Using graphical elements such as progress bars, color gradients, or varying
icon sizes to intuitively convey the confidence level.
The inclusion of prediction confidence is a critical feature that enhances the actionability
and trustworthiness of the ML-driven IDS. Beyond a simple binary "attack/no attack" decision,
confidence scores allow for a nuanced human-in-the-loop approach. For instance, if a model
predicts an attack with 98% confidence, security analysts can act with greater assurance.
However, if the confidence falls to, say, 55%, it prompts double-checking and further
investigation. This mechanism helps strike the right balance between automation and human
30
intervention, enabling security teams to make informed decisions, especially in ambiguous cases.
It directly addresses the issue of "high false-alarm rates" often associated with NIDS by providing
context for alerts and allowing for prioritization, thereby reducing alert fatigue and focusing
human effort on high-risk, high-confidence threats. 10
severity and attack type, allowing for rapid triage and response. Alerting mechanisms can
37
include on-screen notifications, email alerts, or integration with existing Security Information and
Event Management (SIEM) systems.
Furthermore, the system incorporates mechanisms for user feedback. This allows network
administrators to provide input on the accuracy of predictions, for example, by marking false
positives (legitimate traffic flagged as an attack) or false negatives (missed attacks). This user
feedback loop is invaluable for the continuous improvement and fine-tuning of the machine
learning model. By collecting and analysing this feedback, the model can be periodically retrained
or adapted to better reflect the evolving network environment and attack patterns, enhancing its
long-term effectiveness and reducing future misclassifications.
The following table summarizes the key elements and best practices for designing an
effective IDS prediction dashboard:
Table 5.1: Key Elements and Best Practices for IDS Prediction Dashboards
Element Category Key Elements Best Practices
Key Threat detection rates, alert triage Identify most important metrics for the
Metrics/KPIs progress, incident trends, risk scores. organization; use clear, concise
language.36
Visualization Line charts, bar charts, heat maps, Limit number of charts; use consistent
Types attack graphs, tables. layout; use colour to draw attention to
important information. 36
Interactivity Filters (date, attack type, IP), Enable users to select specific data
Features drill-down capabilities, tooltips. ranges/categories; provide more
detailed data on demand. 36
Confidence Score Numerical score (0-1), categorical Aid human decision-making; allow
Display labels (High/Medium/Low), visual prioritization of alerts; help in double-
cues (progress bars). checking low-confidence predictions. 30
Micro-Averaging: This method calculates metrics globally by summing the true positives,
false positives, and false negatives across all classes. It gives equal weight to each
51
Macro-Averaging: This method calculates the metric (e.g., precision, recall, F1-score) for
each class independently and then takes the unweighted average of these per-class scores. 51
It treats all classes equally, regardless of their support (number of instances), and is
sensitive to the performance on rare classes. A low macro-average F1-score, even with
51
approach accounts for class imbalance by giving more importance to larger classes,
providing a more realistic aggregated score for imbalanced datasets. 51
The proposed 1DCNN model is expected to achieve high overall detection accuracies,
with reported values for multi-class classification on UNSW-NB15 around 97.23% 6
and up to
98.3%. Overall detection accuracy on CICIDS2017 has been reported up to 98.89%.
2 25
References
27 url:https://arxiv.org/html/2502.06688v1
3 url:https://kinit.sk/aiming-to-catchem-all-data-for-network-intrusion-detection/
1 url:https://kuey.net/index.php/kuey/article/download/3936/2619/9001
4 url:https://www.researchgate.net/figure/List-of-attacks-UNSW-NB-15-
dataset_tbl1_325656462
9url:https://scispace.com/pdf/cicids-2017-dataset-feature-analysis-with-information-gain-
5gdh 5inehd.pdf
5 url:https://www.researchgate.net/figure/Attack-Types-in-CICIDS-2017-
dataset_tbl1_339527544
14 url:https://arxiv.org/html/2501.14311v1
15 url:https://infonomics-society.org/wp-content/uploads/Comparative-Evaluation-of-
Machine-Learning-Efficacy-in-DoS-Attack-Detection.pdf
7 url:https://www.arcjournals.org/pdfs/ijrscse/v2-i3/7.pdf
21 url:https://www.kaggle.com/code/maryamanwer/ddos-attack-detection-using-ml
25 url:https://www.mdpi.com/1424-8220/24/14/4516
8 url:https://www.mdpi.com/2079-9292/13/12/2404
43 url:https://www.picussecurity.com/resource/glossary/what-is-attack-path-visualization
44 url:https://www.puppygraph.com/blog/attack-graph
6
url:https://pdfs.semanticscholar.org/726a/0aedfb044d255f9dbad024dcf6160eb4769a.pdf
2url:https://www.researchgate.net/publication/
371900034_Analysis_of_Intrusion_Detection_Systems_in_UNSW-NB15_and_NSL-
KDD_Datasets_with_Machine_Learning_Algorithms
18url:https://repository.unsri.ac.id/59517/1/Article-IJEEI-Important%20Features%20of
%20 CICIDS-2017%20Dataset.pdf
17 url:https://www.mdpi.com/2079-8954/12/7/243
33 url:https://www.youtube.com/watch?v=MvTqi2Mb_PM
32 url:https://www.meritshot.com/introduction-to-flask-for-machine-learning/
35url:https://www.researchgate.net/publication/387701314_Real-
time_Intrusion_Detection _System_on_Windows_Implementation_of_a_Network-
Based_Security_Information_System
10url:https://www.upguard.com/blog/top-free-network-based-intrusion-detection-
systems-ids-for-the-enterprise
55 url:https://flask-monitoringdashboard.readthedocs.io/
34 url:https://www.analyticsvidhya.com/blog/2021/10/flask-python/
39 url:https://cloud.google.com/chronicle/docs/investigation/alerts-iocs
40 url:https://apps.dtic.mil/sti/tr/pdf/ADA428197.pdf
1 url:https://kuey.net/index.php/kuey/article/download/3936/2619/9001
26 url:https://arxiv.org/pdf/2101.05067
12 url:https://test-jicce.inforang.com/journal/download_pdf.php?
spage=7&volume=7&number=1
22 url:https://www.mdpi.com/2076-3417/13/16/9363
21 url:https://www.kaggle.com/code/maryamanwer/ddos-attack-detection-using-ml
24 url:https://www.researchgate.net/figure/Confusion-matrix-for-detecting-DoS-attack-
types_fig2 _327065277
31 url:https://telnyx.com/learn-ai/calculating-f1-score
28url:https://www.deepchecks.com/how-to-apply-and-calculate-the-f1-score-in-machine-
learning/
51 url:https://magai.co/weighted-metrics-for-multi-class-models-explained/
52 url:https://www.evidentlyai.com/classification-metrics/multi-class-metrics
11 url:https://www.clearnetwork.com/top-intrusion-detection-and-prevention-systems/
10 url:https://www.upguard.com/blog/top-free-network-based-intrusion-detection-
systems-ids-for -the-enterprise
47url:https://medium.com/voice-tech-global/machine-learning-confidence-scores-all-you-
need-to- know-as-a-conversation-designer-8babd39caae7
30 url:https://www.mindee.com/blog/how-use-confidence-scores-ml-models
56 url:
https://www.digitalocean.com/community/tutorials/how-to-use-web-forms-in-a-flask-
application
19 url:https://www.researchgate.net/figure/Features-of-UNSW-NB15-
dataset_tbl1_324601933
1 url:https://kuey.net/index.php/kuey/article/download/3936/2619/9001
20url:https://www.researchgate.net/publication/332265020_UNSW-
NB15_dataset_feature_
selection_and_network_intrusion_detection_using_deep_learning
57 url:https://www.mdpi.com/2076-3417/13/19/11067
24 url:https://www.researchgate.net/figure/Confusion-matrix-for-detecting-DoS-attack-
types_fig2 _327065277
21 url:https://www.kaggle.com/code/maryamanwer/ddos-attack-detection-using-ml
29 url:https://www.grammarly.com/blog/ai/what-is-f1-score/
31 url:https://telnyx.com/learn-ai/calculating-f1-score
53url:https://wandb.ai/wandb/plots/reports/Confusion-Matrix-Usage-and-Examples--
VmlldzozMDg1NTM
49url:https://www.analyticsvidhya.com/blog/2021/06/confusion-matrix-for-multi-class-
classification/
13 url:https://www.researchgate.net/publication/383322932_Predicting_F1-
Scores_of_Classifiers _in_Network_Intrusion_Detection_Systems
48 url:https://github.com/makt96/treebased-ids
30 url:https://www.mindee.com/blog/how-use-confidence-scores-ml-models
23 url:https://www.mdpi.com/2079-9292/13/2/293
36url:https://www.numberanalytics.com/blog/dashboard-best-practices-data-
warehousing-business -intelligence
37 url:https://www.aufaitux.com/blog/cybersecurity-dashboard-ui-ux-design/
45 url:
https://docs.dynatrace.com/docs/analyze-explore-automate/dashboards-classic/charts-and-
tiles/visualization-graph
58 url:https://cloud.google.com/chronicle/docs/reports/dashboards-overview
41 url:https://moldstud.com/articles/p-flask-security-best-practices-protect-your-web-
applications-from-common-vulnerabilities
42 url:https://flask.palletsprojects.com/en/stable/web-security/
21 url:https://www.kaggle.com/code/maryamanwer/ddos-attack-detection-using-ml
50 url:https://manujayanth07.medium.com/network-intrusion-detection-system-using-
confusion-matrix-f7db52896fb5
59 url:https://blog.roboflow.com/f1-score/
16 url:https://www.researchgate.net/figure/Analysis-of-probe-attack-
detection_fig9_367481862
53url:https://wandb.ai/wandb/plots/reports/Confusion-Matrix-Usage-and-Examples--
VmlldzozMDg1NTM
54 url:https://www.kaggle.com/code/mittalvasu95/multi-class-classification-c101
60 url:https://www.simplekpi.com/KPI-Dashboard-Examples/Security-Firm-Dashboard-
Examples
61 url:https://www.qlik.com/us/dashboard-examples
38 url:https://www.striim.com/blog/real-time-data-visualization-data-exploration/
46 url:https://blog.eclecticiq.com/real-time-threat-visibility-custom-dashboards-for-
cybersecurity-excellence
Appendices
A. Source Code Snippets
(Placeholder for relevant code snippets, e.g., 1DCNN model definition, GA feature
selection implementation, Flask API endpoint, data pre-processing scripts.)
B. Sample Input Data
(Placeholder for examples of pre-processed network flow data, both normal and attack,
that would be fed into the model.)
C. Screenshots of Web Application
(Placeholder for visual representations of the Flask dashboard, including prediction
results, confidence scores, and interactive elements.)
D. API Documentation
(Placeholder for detailed documentation of the Flask API endpoints, including
request/response formats and authentication requirements.) Works cited
1. Advancing Cybersecurity: Leveraging UNSW_NB15 Dataset for Enhanced
Detection and Prediction of Diverse Cyber Threats, accessed July 10, 2025,
https://kuey.net/index.php/kuey/article/download/3936/2619/9001
2. Analysis of Intrusion Detection Systems in UNSW-NB15 and NSL-KDD Datasets
with Machine Learning Algorithms - ResearchGate, accessed July 10, 2025,
https://www.researchgate.net/publication/371900034_Analysis_of_Intrusion_Detec
tion_Systems_in_UNSW-NB15_&_NSL-KDD_Datasets_with_ Machine_Learning
_Algorithms
3. Aiming to Catch'Em All: Our Research on Data for Network Intrusion Detection -
KInIT, accessed July 10, 2025, https://kinit.sk/aiming-to-catchem-all-data-for-
network-intrusion-detection/
4. List of attacks UNSW NB-15 dataset | Download Scientific Diagram -
ResearchGate, accessed July 10, 2025, https://www.researchgate.net/figure/List-of-
attacks-UNSW-NB-15-dataset_tbl1_325656462
5. Attack Types in CICIDS 2017 dataset | Download Scientific Diagram -
ResearchGate, accessed July 10, 2025, https://www.researchgate.net/figure/Attack-
Types-in-CICIDS-2017-dataset_tbl1_339527544
6. Network Intrusion Detection with 1D Convolutional Neural Networks, accessed
July 10, 2025,
https://pdfs.semanticscholar.org/726a/0aedfb044d255f9dbad024dcf6160
eb4769a.pdf
7. Detection of Probe Attacks Using Machine Learning ... - ARC Journals, accessed
July 10, 2025, https://www.arcjournals.org/pdfs/ijrscse/v2-i3/7.pdf
8. Enhancing Multi-Class Attack Detection in Graph Neural Network through Feature
Rearrangement - MDPI, accessed July 10, 2025, https://www.mdpi.com/2079-
9292/13/12/2404
9. CICIDS-2017 Dataset Feature Analysis With Information ... - SciSpace, accessed
July 10, 2025, https://scispace.com/pdf/cicids-2017-dataset-feature-analysis-with-
information-gain-5gdh5inehd.pdf
10. Top 6 Free Network Intrusion Detection Systems (NIDS) Software | UpGuard,
accessed July 10, 2025, https://www.upguard.com/blog/top-free-network-based-
intrusion-detection-systems-ids-for-the-enterprise
11. Top 10 Intrusion Detection and Prevention Systems - ClearNetwork, Inc, accessed
July 10, 2025, https://www.clearnetwork.com/top-intrusion-detection-and-
prevention-systems/
12. International Journal of KIMICS, Vol. 7, No. 1, March 2009 - Design of Hybrid
Network Probe Intrusion Detector using FCM, accessed July 10, 2025, https://test-
jicce.inforang.com/journal/download_pdf.php?spage=7&volume=7&number=1
13. Predicting F1-Scores of Classifiers in Network Intrusion Detection Systems -
ResearchGate, accessed July 10, 2025,
https://www.researchgate.net/publication/383322932_Predicting_F1-
Scores_of_Classifiers_in_Network_Intrusion_Detection_Systems
14. An Efficient Real Time DDoS Detection Model Using Machine Learning
Algorithms - arXiv, accessed July 10, 2025, https://arxiv.org/html/2501.14311v1
15. Comparative Evaluation of Machine Learning Efficacy in DoS Attack Detection |
Infonomics Society, accessed July 10, 2025, https://infonomics-society.org/wp-
content/uploads/Comparative-Evaluation-of-Machine-Learning-Efficacy-in-DoS-
Attack-Detection.pdf
16. Analysis of probe attack detection. | Download Scientific Diagram - ResearchGate,
accessed July 10, 2025, https://www.researchgate.net/figure/Analysis-of-probe-
attack-detection_fig9_367481862
17. Performance Study on the Use of Genetic Algorithm for Reducing ..., accessed July
10, 2025, https://www.mdpi.com/2079-8954/12/7/243
18. Important Features of CICIDS-2017 Dataset For Anomaly Detection in High
Dimension and Imbalanced Class Dataset - Sriwijaya University Repository,
accessed July 10, 2025, https://repository.unsri.ac.id/59517/1/Article-IJEEI-
Important%20Features%20of%20CICIDS-2017%20Dataset.pdf
19. Features of UNSW-NB15 dataset | Download Table - ResearchGate, accessed July
10, 2025, https://www.researchgate.net/figure/Features-of-UNSW-NB15-
dataset_tbl1_324601933
20. UNSW-NB15 dataset feature selection and network intrusion detection using deep
learning, accessed July 10, 2025, https://www.researchgate.net/publication
/332265020_UNSW-NB15_dataset_feature_selection_and_network_intrusion_
detection_using_deep_learning
21. DDoS attack detection using ML - Kaggle, accessed July 10, 2025,
https://www.kaggle.com/code/maryamanwer/ddos-attack-detection-using-ml
22. Feature Engineering and Model Optimization Based Classification Method for
Network Intrusion Detection - MDPI, accessed July 10, 2025,
https://www.mdpi.com/2076-3417/13/16/9363
23. Automated Network Incident Identification through Genetic Algorithm ..., accessed
July 10, 2025, https://www.mdpi.com/2079-9292/13/2/293
24. Confusion matrix for detecting DoS attack types. | Download ..., accessed July 10,
2025, https://www.researchgate.net/figure/Confusion-matrix-for-detecting-DoS-
attack-types_fig2_327065277
25. Using Machine Learning Multiclass Classification Technique to ..., accessed July
10, 2025, https://www.mdpi.com/1424-8220/24/14/4516
26. UNSW-NB15 Computer Security Dataset: Analysis through Visualization - arXiv,
accessed July 10, 2025, https://arxiv.org/pdf/2101.05067
27. arxiv.org, accessed July 10, 2025, https://arxiv.org/html/2502.06688v1
28. How to Apply and Calculate the F1 Score in Machine Learning - Deepchecks,
accessed July 10, 2025, https://www.deepchecks.com/how-to-apply-and-calculate-
the-f1-score-in-machine-learning/
29. How to Use F1 Score in Machine Learning - Grammarly, accessed July 10, 2025,
https://www.grammarly.com/blog/ai/what-is-f1-score/
30. Understanding Confidence Scores in ML: A Practical Guide - Mindee, accessed
July10,2025,https://www.mindee.com/blog/how-use-confidence-scores-ml-models
31. Understanding and calculating the F1 score in ML - Telnyx, accessed July 10,
2025, https://telnyx.com/learn-ai/calculating-f1-score
32. Introduction to Flask for Machine Learning - Meritshot, accessed July 10, 2025,
https://www.meritshot.com/introduction-to-flask-for-machine-learning/
33. Deploy ML Models as APIs with Flask in Python | Step-by-Step Guide - YouTube,
accessed July 10, 2025, https://www.youtube.com/watch?v=MvTqi2Mb_PM
34. Understanding Flask Framework: Installation & features - Analytics Vidhya,
accessed July10,2025,https://www.analyticsvidhya.com/blog/2021/10/flask-python/
35. Real-time Intrusion Detection System on Windows Implementation of a Network-
Based Security Information System - ResearchGate, accessed July 10, 2025,
https://www.researchgate.net/publication/387701314_Real-
time_Intrusion_Detection_System_on_Windows_Implementation_of_a_Network-
Based_Security_Information_System
36. Dashboard Best Practices - Number Analytics, accessed July 10, 2025,
https://www.numberanalytics.com/blog/dashboard-best-practices-data-
warehousing-business-intelligence
37. The Ultimate Guide to Cybersecurity Dashboard UI/UX Design, accessed July 10,
2025, https://www.aufaitux.com/blog/cybersecurity-dashboard-ui-ux-design/
38. Real-Time Data Visualization and Data Exploration - Striim, accessed July 10,
2025, https://www.striim.com/blog/real-time-data-visualization-data-exploration/
39. View Alerts and IOCs | Google Security Operations, accessed July 10, 2025,
https://cloud.google.com/chronicle/docs/investigation/alerts-iocs
40. Visualization Techniques for Intrusion Detection - DTIC, accessed July 10, 2025,
https://apps.dtic.mil/sti/tr/pdf/ADA428197.pdf
41. Flask Security Best Practices - Protect Your Web Applications from Common
Vulnerabilities, accessed July 10, 2025, https://moldstud.com/articles/p-flask-
security-best-practices-protect-your-web-applications-from-common-vulnerabilities
42. Security Considerations — Flask Documentation (3.1.x), accessed July 10, 2025,
https://flask.palletsprojects.com/en/stable/web-security/
43. What Is Attack Path Visualization (APV) In Cybersecurity? - Picus Security,
accessed July 10, 2025, https://www.picussecurity.com/resource/glossary/what-is-
attack-path-visualization
44. Visualizing Cyber Threats: An Introduction to Attack Graphs - PuppyGraph,
accessed July 10, 2025, https://www.puppygraph.com/blog/attack-graph
45. Configure and use a graph visualization in Dynatrace, accessed July 10, 2025,
https://docs.dynatrace.com/docs/analyze-explore-automate/dashboards-classic/
charts-and-tiles/visualization-graph
46. Real-time threat visibility: Custom Dashboards for cybersecurity excellence -
EclecticIQ Blog, accessed July 10, 2025, https://blog.eclecticiq.com/real-time-
threat-visibility-custom-dashboards-for-cybersecurity-excellence
47. Machine Learning Confidence Scores — All You Need to Know as a Conversation
Designer | by Guy TONYE | Voice Tech Global | Medium, accessed July 10, 2025,
https://medium.com/voice-tech-global/machine-learning-confidence-scores-all-
you-need-to-know-as-a-conversation-designer-8babd39caae7
48. makt96/treebased-ids: This project is a live network ... - GitHub, accessed July 10,
2025, https://github.com/makt96/treebased-ids
49. Confusion Matrix for Multi-Class Classification 2025 - Analytics Vidhya, accessed
July 10, 2025, https://www.analyticsvidhya.com/blog/2021/06/confusion-matrix-
for-multi-class-classification/
50. Network intrusion detection system using Confusion Matrix | by Manujayanth -
Medium, accessed July 10, 2025, https://manujayanth07.medium.com/network-
intrusion-detection-system-using-confusion-matrix-f7db52896fb5
51. Weighted Metrics for Multi-Class Models Explained - Magai, accessed July 10,
2025, https://magai.co/weighted-metrics-for-multi-class-models-explained/
52. Accuracy, precision, and recall in multi-class classification - Evidently AI,
accessed July10,2025,https://www.evidentlyai.com/classification-metrics/multi-
class-metrics
53. Confusion Matrix: Usage and Examples | plots – Weights & Biases - Wandb,
accessed July 10, 2025, https://wandb.ai/wandb/plots/reports/Confusion-Matrix-
Usage-and-Examples--VmlldzozMDg1NTM
54. Multi class Classification: C101 - Kaggle, accessed July 10, 2025,
https://www.kaggle.com/code/mittalvasu95/multi-class-classification-c101
55. What is Flask-MonitoringDashboard? — Flask-MonitoringDashboard 3.2.2
documentation, accessed July 10, 2025, https://flask-
monitoringdashboard.readthedocs.io/
56. How To Use Web Forms in a Flask Application - DigitalOcean, accessed July 10,
2025, https://www.digitalocean.com/community/tutorials/how-to-use-web-forms-
in-a-flask-application
57. A Novel Feature Selection Approach to Classify Intrusion Attacks in Network
Communications - MDPI, accessed July 10, 2025, https://www.mdpi.com/2076-
3417/13/19/11067
58. Dashboards overview | Google Security Operations, accessed July 10, 2025,
https://cloud.google.com/chronicle/docs/reports/dashboards-overview
59. What is F1 Score? A Computer Vision Guide. - Roboflow Blog, accessed July 10,
2025, https://blog.roboflow.com/f1-score/
60. Security Firm KPI Dashboard Template | Track Incidents & Compliance -
SimpleKPI.com, accessed July 10, 2025, https://www.simplekpi.com/KPI-
Dashboard-Examples/Security-Firm-Dashboard-Examples
61. Best Dashboard Examples: Over 100 by Industry & Role - Qlik, accessed July 10,
2025, https://www.qlik.com/us/dashboard-examples