A Highly Accurate Forest Fire Prediction Model Based On An
A Highly Accurate Forest Fire Prediction Model Based On An
sciences
Article
A Highly Accurate Forest Fire Prediction Model Based on an
Improved Dynamic Convolutional Neural Network
Shaoxiong Zheng 1 , Peng Gao 1 , Weixing Wang 1,2, * and Xiangjun Zou 3,4, *
1 College of Electronic Engineering, South China Agricultural University, Guangzhou 510642, China;
[email protected] (S.Z.); [email protected] (P.G.)
2 Guangdong Engineering Research Center for Monitoring Agricultural Information, Guangzhou 510642, China
3 Guangdong Laboratory for Lingnan Modern Agriculture, College of Engineering, South China Agricultural
University, Guangzhou 510642, China
4 Foshan-Zhongke Innovation Research Institute of Intelligent Agriculture and Robotics, Foshan 528000, China
* Correspondence: [email protected] (W.W.); [email protected] (X.Z.)
Abstract: In this work, an improved dynamic convolutional neural network (DCNN) model to
accurately identify the risk of a forest fire was established based on the traditional DCNN model.
First, the DCNN network model was trained in combination with transfer learning, and multiple
pre-trained DCNN models were used to extract features from forest fire images. Second, principal
component analysis (PCA) reconstruction technology was used in the appropriate subspace. The
constructed 15-layer forest fire risk identification DCNN model named “DCN_Fire” could accurately
identify core fire insurance areas. Moreover, the original and enhanced image data sets were used to
evaluate the impact of data enhancement on the model’s accuracy. The traditional DCNN model was
improved and the recognition speed and accuracy were compared and analyzed with the other three
DCNN model algorithms with different architectures. The difficulty of using DCNN to monitor forest
fire risk was solved, and the model’s detection accuracy was further improved. The true positive
Citation: Zheng, S.; Gao, P.; Wang, rate was 7.41% and the false positive rate was 4.8%. When verifying the impact of different batch
W.; Zou, X. A Highly Accurate Forest sizes and loss rates on verification accuracy, the loss rate of the DCN_Fire model of 0.5 and the batch
Fire Prediction Model Based on an size of 50 provided the optimal value for verification accuracy (0.983). The analysis results showed
Improved Dynamic Convolutional that the improved DCNN model had excellent recognition speed and accuracy and could accurately
Neural Network. Appl. Sci. 2022, 12, recognize and classify the risk of a forest fire under natural light conditions, thereby providing a
6721. https://doi.org/10.3390/ technical reference for preventing and tackling forest fires.
app12136721
Academic Editor: Antonella Keywords: forest fire; deep learning; transfer learning; convolutional neural network
D’Alessandro
that could achieve good recognition results. Ciprián-Sánchez et al. [11] extracted and
combined information from various imaging modes for detecting and extracting features,
and predicting the spread of wildfires via a state-of-the-art (SOTA) model based on deep
learning (DL); the architecture, loss function, and different types of image combinations
of the SOTA model were evaluated to identify the parameters related to improving the
segmentation results and comparing them with traditional fire segmentation technology,
thereby improving the ability to identify the risk of a forest fire. Moayedi H et al. [12] used
a hybrid evolutionary algorithm to build a forest fire risk prediction model; a forest fire
sensitivity map of fire-prone areas in Iran was generated based on combining an adaptive
neuro-fuzzy inference system, a genetic algorithm, particle swarm optimization, and differ-
ential evolution algorithm with reliable accuracy. Vikram R et al. [13] used a support vector
machine (SVM) for a semi-supervised classification model to divide the forest area into
different subareas: high activity (HA), medium activity (MA), and low activity (LA). Due
to energy limitations, a sensor node that only monitored one parameter could predict the
fire risk in different areas with 90% accuracy. Moreover, greedy forwarding technology was
used to continuously and periodically send data packets from HA and MA areas to the base
station, respectively, but prevented them being sent from LA areas. This data forwarding
technology improved the network life and reduced congestion during data transmission
from forest areas to the base station. Achu A L et al. [14] proposed an automatic fire
monitoring system to identify forest fires in their early stages that could predict the size of
the flame areas. A deep CNN was used for forest fire risk image classification to extract
descriptors from images and was then applied to a logistic regression classifier. After using
882 images to form an ordered data set and related image metadata (such as flame, smoke,
fog, cloud, and human elements) during testing, the classification accuracy of 695 daytime
and 187 night-time scene images was 94.1% and 94.8%, respectively, with good accuracy in
estimating the flame area. Pham B T et al. [15] compared image frames via a segmentation
method based on distance measurements in a color channel histogram and a fast frame
comparison and extraction algorithm based on time. The module uses key advances in
video frame technology and processes the frames through a deep CNN that was trained
via normalization prior to segmentation to assist in detecting fire and smoke. The fire risk
recognition accuracy was 90%. Michael Y et al. [16] used two pre-trained DL CNN models
(vgg16 and resnet50) for fire detection and enhanced the depth of network learning by
fine-tuning the network on the basis of a fully connected layer. Although this improved the
detection accuracy, it also increased the network training time. However, the deep CNN
performed well on complex data sets.
At present, research on forest fire risk detection based on ML mainly involves flame
detection and smoke detection [17–21]. The forest fire risk monitoring methods proposed
by most scholars are for specific scenes, with several problems that need to be solved.
How to recognize the color, form, and texture of fire in different chaotic backgrounds is
the key to forest fire risk recognition [22–26]. The purpose of this study was to effectively
extract the complex upper layer features of fire risk images and improve the robustness of
input conversion by constructing an unmanned aerial vehicle (UAV) image-based forest
fire risk prediction model based on a deep learning back-propagation neural network. The
problems to be solved in the process of fire risk identification are feature extraction from
training image data sets and training with a large number of fire risk image data sets [27,28].
In this study, we attempted to solve the above problems as follows. First, the dynamic
CNN (DCNN) network was trained in combination with transfer learning, and multiple
pre-trained DCNNs were used to extract features from forest fire images. The traditional
DCNN model was improved and the recognition speed and accuracy were compared and
analyzed with the other three DCNN model algorithms with different architectures. The
difficulty of using DCNN to monitor forest fire risk was solved and the model’s detection
accuracy was further improved. Therefore, the forest fire risk model based on the improved
DCNN could accurately identify and classify the risk of a forest fire under natural light
conditions. The main research contents of this paper were as follows:
Appl. Sci. 2022, 12, 6721 3 of 15
(1) The DCNN network was trained in combination with migration learning, and multiple
pre-trained DCNN models were used to extract the features of forest fire images.
(2) Principal component analysis (PCA) reconstruction technology was used to enhance
image category differentiation.
(3) A 15-layer DCNN model called “DCN_Fire”, as an improvement on the traditional
DCNN model, was constructed and analyzed.
(4) The recognition speed and accuracy of the improved DCNN model were compared
with three other DCNN models with different architectures.
(5) The DCNN model could accurately recognize the risk of a forest fire risk in natural
light; thus it was suitable as an early warning system for forest fires.
In Section 2, we discuss the extraction and enhancement of pertinent image features
using the DCNN. Section 3 presents the improved DCNN model and its construction
process. Section 4 provides an analysis of the model verification and experimental results.
Finally, conclusions based on the study findings are presented in Section 5.
Figure1.1.The
Figure Thearchitecture
architectureof
ofthe
theDCN_Fire
DCN_Firemodel.
model.
exp( ai )
So f tmax = (1)
∑ j exp a j
Pixels of size 128 × 128 were input into the DCNN, after which the error between the
actual output value and the required value was calculated. The stochastic gradient descent
method was used on the training-depth CNN. After passing the back-propagation error in
the previous layer, the method calculated the derivative on the optimization coefficient and
weight. The learning process of the model was iteratively optimized until the training error
was small enough to be ignored.
combined classification technology with various features [38] to improve the accuracy of
recognizing forest fire risk from images. These features could be expressed as
n om
Fi = f ik (2)
k =1
F 0 i = Ωi Fi (3)
where m
1 1
Ωi = /∑ (4)
EFi i =1
EFi
and Fi is the original feature of f ( x ) = max (0, x ). The value of the series is calculated
by using
T
FFV = F 0 1 , · · · · · · , F 0 m (5)
The quantity product of the characteristics is calculated as follows:
The influence of the features is eliminated and the richness of the fusion feature sub-
space is enhanced in the above feature fusion methods [38]. We used reconstruction tech-
nology based on PCA to enhance the distinction between specific categories. The variables
x and y are combined as ( x1 , y1 ), ( x2 , y2 ), · · · , ( xk , yk ), where xi ∈ <n ,yi ∈ {1, 2, · · · , M } is
the real class label of xi estimated according to its density pi ( x ).
ŷ = argmaxpi ( x ),
(10)
i = 1, · · · , M
where pi ( x ) = ∑ kj=1 y j λij C x, x j + bi when x is positive (the greater the value, the higher
the confidence level), where λ is the coefficient term, C represents the penalty function, and
b is the deviation parameter.
We adopted a linear SVM based on scikit-learn in Python. The bottleneck feature of
DCNN is usually linearly separable, and so the features of the fused DCNN model could
be processed well by using a linear SVM.
3.2. Significance
SignificanceSegmentation
Segmentationand
andData
DataEnhancement
Enhancement
We used
We used significance
significance detection
detection to
to segment
segment the
the flame
flame image
image to
to standard
standard sizes
sizes and
and train
train
the model.
model.The
Themodel
model was used
was to process
used images
to process withwith
images flames, smoke,
flames, or flames
smoke, and smoke.
or flames and
Figure 2Figure
smoke. shows2the processing
shows results. results.
the processing
(a)
(b)
(c)
(d)
Figure2.2. Image
Figure Image segmentation
segmentation based
based on
on saliency
saliency detection.
detection. (a)
(a) Original
Original images;
images;(b)
(b)significance
significancetest
test
results;(c)
results; (c)ROI
ROIselection
selectionresults;
results;and
and(d)
(d)standard
standardsegmentation
segmentationresults.
results.
Figure 2a shows the original forest fire images while Figure 2b exhibits the results of
significance detection. As shown in Figure 2c, the proposed significance method effec-
tively located the core flame area in the aerial images and even very small ignition areas.
In addition, Figure 2d shows that the original image was segmented into independent
(d)
Appl. Sci. 2022, 12, 6721 7 of 15
Figure 2. Image segmentation based on saliency detection. (a) Original images; (b) significance test
results; (c) ROI selection results; and (d) standard segmentation results.
Figure2a
Figure 2ashows
showsthetheoriginal
originalforest
forestfire
fireimages
imageswhile
whileFigure
Figure2b2bexhibits
exhibitsthe
theresults
resultsof
of
significancedetection.
significance detection.AsAsshown
shown in in Figure
Figure 2c, 2c,
the the proposed
proposed significance
significance methodmethod effec-
effectively
tively located
located theflame
the core core flame
area inarea
thein the aerial
aerial images images and even
and even veryvery
smallsmall ignition
ignition areas.
areas. In
In addition,
addition, Figure
Figure 2d shows
2d shows thatoriginal
that the the original
imageimage was segmented
was segmented into independent
into independent flame
flame images
images containing
containing complete complete flame using
flame features featurestheusing the proposed
proposed method.
method. Data Data en-
enhancement
hancement
could could
effectively effectively
solve solve the
the over-fitting over-fitting
problem causedproblem
by usingcaused
a smallby using data
training a small
set.
Figure 3 shows
training examples
data set. Figure 3ofshows
the test results. of the test results.
examples
(a)
(b)
(c)
(d)
(e)
(f)
Figure 3.
Figure 3. Examples
Examples of
of fire
fire location
location and
and segmentation
segmentationscenes.
scenes. (a–f)
(a–f) Examples
Examples1–6.
1–6.
We tested the performance of the fire risk location and segmentation algorithm with
550 images of flames selected from the original data set. The test results are reported in
Table 2. From the test results, the true positive rate was 7.41% and the false positive rate
was 4.8%. Since the images were small, the characteristics of flames and smoke covered
Appl. Sci. 2022, 12, 6721 8 of 15
We tested the performance of the fire risk location and segmentation algorithm with
550 images of flames selected from the original data set. The test results are reported in
Table 2. From the test results, the true positive rate was 7.41% and the false positive rate
was 4.8%. Since the images were small, the characteristics of flames and smoke covered
most areas. At this time, segmentation was not necessary. To enhance the recognition of
fire risk images, the data set was enhanced. The new enhanced image database named
“DB_Fire” contained 3845 images. Table 3 shows the parameters of the original image
database and enhanced image database.
Image
Total ROI Detected from 240 Aerial Fire Images 492 Total ROI Detected from 310 Normal-View Images 683
Type
TP FN FP TN TPR FPR TP FN FP TN TPR FPR
Performance
329 19 8 136 96.1% 4.8% 463 47 16 157 92.2% 7.41%
TP, true positive; TN, true negative; FP, false positive; FN, false negative; TPR, the true positive rate; FPR, the false
positive rate.
Table 4. The influence of batch size and loss rate on the verification accuracy of DCN_Fire.
4. Model Validation
4.1. Model Performance Verification
To verify the model’s accuracy, we explored several forest fire risk monitoring DCNN
models with various hierarchical structures and calculated the image processing time and
verification accuracy of each one using the original and enhanced image data sets. The
models involved in the comparison included the DCN_Fire model and 13-layer, 15-layer,
and 20-layer DCNN models. Figure 4 shows the comparison of flame image verification
accuracy. Figure 5 shows the comparison of flame image processing times by the models.
To verify the model’s accuracy, we explored several forest fire risk monitoring
DCNN models with various hierarchical structures and calculated the image processing
time and verification accuracy of each one using the original and enhanced image data
sets. The models involved in the comparison included the DCN_Fire model and
13-layer, 15-layer, and 20-layer DCNN models. Figure 4 shows the comparison of flame
Appl. Sci. 2022, 12, 6721 9 of 15
image verification accuracy. Figure 5 shows the comparison of flame image processing
times by the models.
Figure
Figure 5.5.Flame
Flameimage
image processing
processing time
time comparison
comparison ofDCN_Fire
of the the DCN_Fire
modelmodel and 13-layer,
and 13-layer, 15-layer, 15-layer
and 20-layerDCNN
and 20-layer DCNN models.
models.
Although
Althoughthe 20-layer
the DCNN
20-layer DCNN model achieved
model a verification
achieved accuracyaccuracy
a verification similar to similar
that of to that
the DCN_Fire model (0.983), its processing time was 1.67 longer due to
of the DCN_Fire model (0.983), its processing time was 1.67 longer due to the larger the larger number
of layers, increasing the calculation time and cost. Therefore, the DCN_Fire model was
number of layers, increasing the calculation time and cost. Therefore, the DCN_Fire
superior in terms of processing time and verification accuracy. Meanwhile, the original
model was superior
and enhanced image datain terms
sets wereof used
processing timethe
to evaluate and verification
impact accuracy. Meanwhile
of data enhancement on
the original and enhanced image data sets were used to evaluate
model accuracy. The data revealed that the verification accuracy of the same model the impact of data en-
using
hancement
the enhancedon model
image dataaccuracy. Thethan
set was better data revealed
that using thethat the verification
original image data set.accuracy of the
sameAfter
model using the enhanced image data set was better than that
dividing the image samples into training and verification sets, the DCN_Fire using the original
model and
image datathe 13-layer and 15-layer DCNN models were trained and their loss values were
set.
calculated. Figure
After dividing 6 shows the training
the image and verification
samples into traininglossand
curves of the foursets,
verification models.
the DCN_Fire
The loss curves is showen in Figure 6, it mark the training
model and the 13-layer and 15-layer DCNN models were trained and their loss once every 100 iterations
loss values
and the validity loss once every 500 iterations.
were calculated. Figure 6 shows the training and verification loss curves of the four
The data from Figures 4–6 show that the DCN_Fire model had better recognition
models.
accuracy than the other models with minimal training and verification losses.
hancement on model accuracy. The data revealed that the verification accuracy of the
same model using the enhanced image data set was better than that using the original
image data set.
After dividing the image samples into training and verification sets, the DCN_Fire
model and the 13-layer and 15-layer DCNN models were trained and their loss values
Appl. Sci. 2022, 12, 6721 10 of 15
were calculated. Figure 6 shows the training and verification loss curves of the four
models.
(a)
(b)
(c)
Figure 6.
Figure 6. Training and validation
Training and validation loss
loss curve
curve of
of (a) the DCN_Fire
(a) the DCN_Fire model,
model, (b)
(b) the 13-layer DCNN
the 13-layer DCNN
model, and (c) the 13-layer DCNN model.
model, and (c) the 13-layer DCNN model.
The lossofcurves
4.2. Analysis is showenby
Fire Recognition inthe
Figure 6, it mark
DCN_Fire Modelthe training loss once every 100 itera-
tionsFigure
and the validity loss once every 500 iterations.
7 shows the visual characteristics after the input images had passed through
The data
the Conv 2 andfrom
ConvFigures 4–6 Itshow
3 layers. couldthat
be the DCN_Fire
observed model
that the lowerhadlayer
better recognition
captured low-
accuracy than the other models with minimal training and verification losses.
The DCN_Fire
The DCN_Fire model
modelwaswasusedusedtotoidentify
identifyforest firefire
forest images
imagesandand
verify the effective-
verify the effective-
ness and rationality of the classification of the forest fire risk from aerial
ness and rationality of the classification of the forest fire risk from aerial images images (Figure 8). 8).
(Figure
Since the
Since thesmoke
smokefeatures
featuresininthe
thesmoke
smoke(Figure
(Figure8c)8c)
andandmistmist (Figure
(Figure 8d)8d) images
images were
were highly
highly similar, DCN_Fire mistakenly classified forest mist images as fire
similar, DCN_Fire mistakenly classified forest mist images as fire risk images. This was not risk images.
This was not surprising as humans can also mistakenly identify mist as smoke.
surprising as humans can also mistakenly identify mist as smoke.
The test set consists of 1200 random images with and without flames from the database.
From the results in Table 5, it could be seen see that the false negative rate with the DCN_Fire
model (0.13%) was very low.
Table 5. The performance of the DCN_Fire model on the test data set.
The same training and test data sets were used to verify and compare the recognition
accuracy of the DCN_Fire model and other mainstream ML models (Kim’s CNN, AlexNet,
8-layer CNN + Fisher vector, HOG + (b)SVM, and deep belief net + neural net) for fire risk
recognition. The training results showed that DCN_Fire was more accurate than the other
models and performed well (Figure 9, Table 6).
(d)
(a) The original input image, (b) the Conv 2 feature map, and (c) the Conv 3 feature map.
The DCN_Fire model was used to identify forest fire images and verify the effective-
ness and rationality of the classification of the forest fire risk from aerial images (Figure 8).
Appl. Sci. 2022, 12, 6721
Since the smoke features in the smoke (Figure 8c) and mist (Figure 8d) images12 were of 15
highly similar, DCN_Fire mistakenly classified forest mist images as fire risk images.
This was not surprising as humans can also mistakenly identify mist as smoke.
(a)
The test set consists of 1200 random images with and without flames from the da-
tabase. From the results in Table 5, it could be seen see that the false negative rate with
the DCN_Fire model (0.13%) was very low.
Table 5. The performance of the DCN_Fire model on the test data set.
(c)
TP FN TN FP TPR TNR FNR General Accuracy
592 8 588 12 98.7% 98.0% 0.13% 98.3%
TP, true positive; TN, true negative; FP, false positive; FN, false negative; TPR, the true positive
rate; TNR, the true negative rate; FNR, The false negative rate.
The same training and test data sets were used to verify and compare the recogni-
tion accuracy of the DCN_Fire model and other mainstream ML models (Kim’s CNN,
(d)
AlexNet, 8-layer CNN + Fisher vector, HOG + SVM, and deep belief net + neural net) for
fire risk8.recognition.
Figure Theclassification
Fire risk image training results
usingshowed that DCN_Fire
the DCN_Fire model. (a)was more
Images accurate
without fire than
character-
the other models and performed well (Figure 9, Table 6).
istics; (b) images with flames; (c) images with smoke; and (d) misclassified images with mist.
Figure
Figure 9.
9. Accuracy
Accuracycomparison
comparisonofof
the DCN_Fire
the model
DCN_Fire andand
model other mainstream
other ML ML
mainstream models.
models.
Appl. Sci. 2022, 12, 6721 13 of 15
5. Conclusions
The detection and recognition of forest fire risk images is of great significance for the
detection and prevention of forest fires. Combined with migration learning, the work in this
paper trained a DCNN network, used multiple pre-trained DCNN models to extract the
features of forest fire images, used PCA reconstruction technology to convert features into
a shared feature subspace, and established the forest fire risk recognition model. According
to the experimental results, the main research conclusions of this paper were as follows:
(1) A DCNN network combined with migration learning was constructed and trained
using multiple pre-trained DCNN models to extract features from forest fire images
and used PCA reconstruction technology to convert features into a shared feature
subspace to establish a forest fire risk recognition model. We used 550 flame images
to test the fire risk location performance and the segmentation algorithm. The true
positive rate was 7.41% and the false positive rate was 4.8%. When verifying the
impact of different batch sizes and loss rates on verification accuracy, the loss rate of
the DCN_Fire model of 0.5 and the batch size of 50 provided the optimal value for
verification accuracy (0.983).
(2) Comparing the test results showed that the performance of the improved DCNN
model was comparable with that of other models. In terms of processing time and
verification accuracy, the DCN_Fire model was considered a better DCNN architecture
in fire risk identification. When calculating the training and verification losses of
several models, those of the DCN_Fire model were minimized.
Although the improved DCNN model achieved good results in the detection of forest
fire risk, the detection accuracy still warrants improvement. In future research, we will
further optimize the network model structure and improve the network performance of
the forest fire risk identification model.
Author Contributions: Conceptualization, S.Z. and W.W.; methodology, S.Z., P.G., W.W. and X.Z.;
software, S.Z. and X.Z.; validation, S.Z., W.W. and X.Z.; formal analysis, S.Z., P.G. and X.Z.; investiga-
tion, S.Z., W.W. and X.Z.; resources, S.Z., W.W. and X.Z.; data curation, X.Z. and P.G.; writing—original
draft preparation, S.Z.; writing—review and editing, W.W. and X.Z.; visualization, S.Z. and X.Z.;
supervision, W.W., X.Z. and S.Z.; project administration, W.W. and X.Z. All authors have read and
agreed to the published version of the manuscript.
Funding: This work was supported by the Key Field R&D Program Project of Guangdong Province,
China (Grant No. 2019B020223003), Guangzhou Science and Technology Plan Project Innovation
Platform Construction and Sharing (Grant No. 201605030013), Guangdong Laboratory of Lingnan
Modern Agriculture Project under Grant NT2021009 and the No. 03 Special Project and the 5G Project
of Jiangxi Province under Grant 20212ABC03A27.
Acknowledgments: We would like to thank our anonymous reviewers for their critical comments
and suggestions for how to improve our manuscript.
Conflicts of Interest: The authors declare that there are no conflict of interest.
References
1. Tariq, A.; Shu, H.; Siddiqui, S.; Mousa, B.G.; Munir, I.; Nasri, A.; Waqas, H.; Lu, L.; Baqa, M.F. Forest fire monitoring using
spatial-statistical and Geo-spatial analysis of factors determining forest fire in Margalla Hills, Islamabad, Pakistan. Geomat. Nat.
Hazards Risk 2021, 12, 1212–1233. [CrossRef]
2. Vigna, I.; Besana, A.; Comino, E.; Pezzoli, A. Application of the socio-ecological system framework to forest fire risk management:
A systematic literature review. Sustainability 2021, 13, 2121. [CrossRef]
3. Naderpour, M.; Rizeei, H.M.; Khakzad, N.; Pezzoli, A. Forest fire induced Natech risk assessment: A survey of geospatial
technologies. Reliab. Eng. Syst. Saf. 2019, 191, 106558. [CrossRef]
4. Çolak, E.; Sunar, F. Evaluation of forest fire risk in the Mediterranean Turkish forests: A case study of Menderes region, Izmir.
Int. J. Disaster Risk Reduct. 2020, 45, 101479. [CrossRef]
5. Van Hoang, T.; Chou, T.Y.; Fang, Y.M.; Nguyen, N.T.; Nguyen, Q.H.; Xuan Canh, P.; Ngo Bao Toan, D.; Nguyen, X.L.; Meadows,
M.E. Mapping forest fire risk and development of early warning system for NW Vietnam using AHP and MCA/GIS methods.
Appl. Sci. 2020, 10, 4348. [CrossRef]
Appl. Sci. 2022, 12, 6721 14 of 15
6. Janiec, P.; Gadal, S. A comparison of two machine learning classification methods for remote sensing predictive modeling of the
forest fire in the North-Eastern Siberia. Remote Sens. 2020, 12, 4157. [CrossRef]
7. Gigovic, L.; Jakovljevic, G.; Sekulović, D.; Regodić, M. GIS multi-criteria analysis for identifying and mapping forest fire hazard:
Nevesinje, Bosnia and Herzegovina. Teh. Vjesn. 2018, 25, 891–897.
8. Mohajane, M.; Costache, R.; Karimi, F.; Pham, Q.B.; Essahlaoui, A.; Nguyen, H.; Laneve, G.; Oudija, F. Application of remote
sensing and machine learning algorithms for forest fire mapping in a Mediterranean area. Ecol. Indic. 2021, 129, 107869. [CrossRef]
9. Kalantar, B.; Ueda, N.; Idrees, M.; Janizadeh, S.; Ahmadi, K.; Shabani, F. Forest fire susceptibility prediction based on machine
learning models with resampling algorithms on remote sensing data. Remote Sens. 2020, 12, 3682. [CrossRef]
10. Stula, M.; Krstinic, D.; Seric, L. Intelligent forest fire monitoring system. Inf. Syst. Front. 2012, 14, 725–739. [CrossRef]
11. Ciprián-Sánchez, J.F.; Ochoa-Ruiz, G.; Rossi, L.; Morandini, F. Assessing the impact of the loss function, architecture and image
type for Deep Learning-based wildfire segmentation. Appl. Sci. 2021, 11, 7046. [CrossRef]
12. Moayedi, H.; Mehrabi, M.; Bui, D.T.; Pradhan, B.; Foong, L.K. Fuzzy-metaheuristic ensembles for spatial assessment of forest fire
susceptibility. J. Environ. Manag. 2020, 260, 109867. [CrossRef] [PubMed]
13. Vikram, R.; Sinha, D.; De, D.; Das, A.K. EEFFL: Energy efficient data forwarding for forest fire detection using localization
technique in wireless sensor network. Wirel. Netw. 2020, 26, 5177–5205. [CrossRef]
14. Achu, A.L.; Thomas, J.; Aju, C.D.; Gopinath, G.; Kumar, S.; Reghunath, R. Machine-learning modelling of fire susceptibility in a
forest-agriculture mosaic landscape of southern India. Ecol. Inform. 2021, 64, 101348. [CrossRef]
15. Pham, B.T.; Jaafari, A.; Avand, M.; Al-Ansari, N.; Dinh Du, T.; Yen, H.P.H.; Phong, T.V.; Nguyen, D.H.; Le, H.V.; Mafi-Gholami,
D.; et al. Performance evaluation of machine learning methods for forest fire modeling and prediction. Symmetry 2020, 12, 1022.
[CrossRef]
16. Michael, Y.; Helman, D.; Glickman, O.; Gabay, D.; Brenner, S.; Lensky, I.M. Forecasting fire risk with machine learning and
dynamic information derived from satellite vegetation index time-series. Sci. Total Environ. 2021, 764, 142844. [CrossRef]
17. Naderpour, M.; Rizeei, H.M.; Ramezani, F. Forest fire risk prediction: A spatial deep neural network-based framework. Remote
Sens. 2021, 13, 2513. [CrossRef]
18. Wang, C.; Luo, T.; Zhao, L.; Tang, Y.; Zou, X. Window zooming–based localization algorithm of fruit and vegetable for harvesting
robot. IEEE Access 2019, 7, 103639–103649. [CrossRef]
19. Resco de Dios, V.; Nolan, R.H. Some challenges for forest fire risk predictions in the 21st century. Forests 2021, 12, 469. [CrossRef]
20. Bowman, D.M.; Williamson, G.J. River flows are a reliable index of forest fire risk in the temperate Tasmanian Wilderness World
Heritage Area, Australia. Fire 2021, 4, 22. [CrossRef]
21. Razavi-Termeh, S.V.; Sadeghi-Niaraki, A.; Choi, S.M. Ubiquitous GIS-based forest fire susceptibility mapping using artificial
intelligence methods. Remote Sens. 2020, 12, 1689. [CrossRef]
22. Salazar, L.G.F.; Romão, X.; Paupério, E. Review of vulnerability indicators for fire risk assessment in cultural heritage.
Int. J. Disaster Risk Reduct. 2021, 60, 102286. [CrossRef]
23. Son, B.H.; Kang, K.H.; Ryu, J.R.; Roh, S.J. Analysis of Spatial Characteristics of Old Building Districts to Evaluate Fire Risk Factors.
J. Korea Inst. Build. Constr. 2022, 22, 69–80.
24. Wang, H.; Dong, L.; Zhou, H.; Luo, L.; Lin, G.; Wu, J.; Tang, Y. YOLOv3-Litchi detection method of densely distributed litchi in
large vision scenes. Math. Probl. Eng. 2021, 2021, 8883015. [CrossRef]
25. Anderson-Bell, J.; Schillaci, C.; Lipani, A. Predicting non-residential building fire risk using geospatial information and convolu-
tional neural networks. Remote Sens. Appl. Soc. Environ. 2021, 21, 100470. [CrossRef]
26. Maffei, C.; Lindenbergh, R.; Menenti, M. Combining multi-spectral and thermal remote sensing to predict forest fire characteristics.
ISPRS J. Photogramm. Remote Sens. 2021, 181, 400–412. [CrossRef]
27. Hansen, R. The Flame Characteristics of a Tyre Fire on a Mining Vehicle. Min. Metall. Explor. 2022, 39, 317–334. [CrossRef]
28. Tomar, J.S.; Kranjčić, N.; Ðurin, B.; Kanga, S.; Singh, S.K. Forest fire hazards vulnerability and risk assessment in Sirmaur district
forest of Himachal Pradesh (India): A geospatial approach. ISPRS Int. J. Geo-Inf. 2021, 10, 447. [CrossRef]
29. Ozenen Kavlak, M.; Cabuk, S.N.; Cetin, M. Development of forest fire risk map using geographical information systems and
remote sensing capabilities: Ören case. Environ. Sci. Pollut. Res. 2021, 28, 33265–33291. [CrossRef]
30. Lin, G.; Zhu, L.; Li, J.; Zou, X.; Tang, Y. Collision-free path planning for a guava-harvesting robot based on recurrent deep
reinforcement learning. Comput. Electron. Agric. 2021, 188, 106350. [CrossRef]
31. Park, M.; Tran, D.Q.; Lee, S.; Park, S. Multilabel Image Classification with Deep Transfer Learning for Decision Support on
Wildfire Response. Remote Sens. 2021, 13, 3985. [CrossRef]
32. Šerić, L.; Pinjušić, T.; Topić, K.; Blažević, T. Lost person search area prediction based on regression and transfer learning models.
ISPRS Int. J. Geo-Inf. 2021, 10, 80. [CrossRef]
33. Lawal, M.O. Tomato detection based on modified YOLOv3 framework. Sci. Rep. 2021, 11, 1447. [CrossRef] [PubMed]
34. Tang, Y.; Zhu, M.; Chen, Z.; Wu, C.; Chen, B.; Li, C.; Li, L. Seismic performance evaluation of recycled aggregate concrete-filled
steel tubular columns with field strain detected via a novel mark-free vision method. In Structures; Elsevier: Amsterdam,
The Netherlands, 2022; Volume 37, pp. 426–441.
35. Arif, M.; Alghamdi, K.K.; Sahel, S.A.; Alosaimi, S.O.; Alsahaft, M.E.; Alharthi, M.A.; Arif, M. Role of machine learning algorithms
in forest fire management: A literature review. J. Robot. Autom. 2021, 5, 212–226.
Appl. Sci. 2022, 12, 6721 15 of 15
36. Quintero, N.; Viedma, O.; Urbieta, I.R.; Moreno, J.M. Assessing landscape fire hazard by multitemporal automatic classification
of landsat time series using the Google Earth Engine in West-Central Spain. Forests 2019, 10, 518. [CrossRef]
37. Akilan, T.; Wu, Q.J.; Zhang, H. Effect of fusing features from multiple DCNN architectures in image classification. IET Image
Processing 2018, 12, 1102–1110. [CrossRef]
38. Roy, A.M.; Bhaduri, J. Real-time growth stage detection model for high degree of occultation using DenseNet-fused YOLOv4.
Comput. Electron. Agric. 2022, 193, 106694. [CrossRef]