Minerals 14 01021 v3
Minerals 14 01021 v3
Review
A Review of Mineral Prospectivity Mapping Using
Deep Learning
Kang Sun 1,2 , Yansi Chen 1,2, * , Guoshuai Geng 1,2 , Zongyue Lu 1,2 , Wei Zhang 1,2 , Zhihong Song 1,2 ,
Jiyun Guan 3 , Yang Zhao 4 and Zhaonian Zhang 5
1 Center for Geophysical Survey, China Geological Survey, Langfang 065000, China;
sunkang@[Link] (K.S.); hnsmxggs@[Link] (G.G.); luzongyue@[Link] (Z.L.);
zhangweicugb@[Link] (W.Z.); songzhihong2018@[Link] (Z.S.)
2 Technology Innovation Center for Earth Near Surface Detection, China Geological Survey,
Langfang 065000, China
3 Kunming Natural Resources Comprehensive Survey Center, China Geological Survey,
Kunming 650100, China; guanjiyun@[Link]
4 Langfang Comprehensive Survey Center of Natural Resources, China Geological Survey,
Langfang 065000, China; zhaoyang02@[Link]
5 School of the Earth Sciences and Resources, China University of Geosciences, Beijing 100083, China;
zhangzhaonian1996@[Link]
* Correspondence: chenyansi@[Link]
Abstract: Mineral resources are of great significance in the development of the national economy.
Prospecting and forecasting are the key to ensure the security of mineral resources supply, promote
economic development, and maintain social stability. The methods for prospecting prediction
have evolved from qualitative to quantitative prediction, from empirical research to mathematical
analysis. In recent years, deep learning algorithms have gradually entered the attention of geologists
due to their robust learning and simulation ability in the application of prospecting prediction.
Deep learning algorithms can effectively analyze and predict data, which have great significance
in improving the efficiency and accuracy of mineral exploration. However, there are not many
specific examples of their application in mineral exploration prediction, and researchers have not
yet conducted a comprehensive discussion on the advantages, disadvantages, and accuracy of deep
learning algorithms in mineral prospectivity mapping applications. This paper reviews and discusses
Citation: Sun, K.; Chen, Y.; Geng, G.;
the application of deep learning in prospecting prediction, highlighting the challenges faced by deep
Lu, Z.; Zhang, W.; Song, Z.; Guan, J.;
learning in the application of prospecting prediction in data preprocessing, data enhancement, system
Zhao, Y.; Zhang, Z. A Review of
parameter adjustment, and accuracy evaluation, and puts forward specific suggestions for research in
Mineral Prospectivity Mapping Using
Deep Learning. Minerals 2024, 14,
these aspects. The purpose of this paper is to provide a reference for the application of deep learning
1021. [Link] to researchers and practitioners in the field of prospecting prediction.
min14101021
Keywords: mineral prospectivity mapping; deep learning; data enhancement method
Academic Editor: Behnam Sadeghi
because of its ability to deal with complex relationships and nonlinear representation of
high-dimensional data.
summarized the advantages and possible limitations of deep learning compared to other
methods in mineral prospectivity mapping practice. By reviewing the current research
results, methods, and cases, the application prospects and existing problems of deep
learning technology in this field are discussed, helping researchers to understand the
current research hotspots and future development direction, and providing references for
research and application in related fields. The article’s structure is as follows. The first
section mainly introduces the data types and data formats used by deep learning in the
application of prospecting; the second section mainly introduces the main types of deep
learning algorithm models; the third section introduces several application examples of
deep learning in geological prospecting; the fourth section systematically summarizes the
outstanding problems and future development trends of deep learning in the application
of prospecting; and the fifth section concludes.
2. Data Foundation
2.1. Data Types
The types of data required for mineral prediction mapping mainly include geological
data, geophysical data, geochemical data, remote sensing data, and so on. These data types
play a crucial role in mineral prediction mapping and can help scientists and geological
experts better understand and predict the location and type of mineral deposits. Geological
data are the basis for mineral prospecting prediction, including information on petrology,
mineralogy, ore deposit geology, and ore-forming structures. These data help determine
geological structures and rock types, extract key geological information related to min-
eralization, delineate geological units as a basis for regional analogies, and thereby infer
potential mineral deposit locations. Geophysical data are obtained by measuring changes
in the Earth’s physical fields, such as gravity, magnetic force, and seismic wave data. These
data can reveal differences in the density and elasticity of underground materials, pro-
viding a comprehensive assessment of the Earth’s physical field within a selected area,
and thus helping to locate mineral deposits. Geochemical data are obtained by analyzing
the chemical compositions of soil, water, and rocks. Abnormal concentrations of specific
elements and compounds may indicate the presence of mineral deposits, and the anomaly
zones are delineated through the integration of various types of geochemical data within
the study area. Remote sensing data use satellite and aerial photography technology to
obtain information about surface features, helping to identify surface cover and vegetation
patterns, as well as alteration information related to mineralization, and delineate areas
with key information for mineralization. MPM involves combining all the data anomalies
to make mineral prospecting predictions within the study area.
The quality of the data we use has a significant impact on the results of our mineral
prospecting predictions. Therefore, when selecting data, we should pay attention to the
following issues:
(1) Accuracy of data. The accuracy of data directly affects the credibility of data analysis
and decision-making. If there are errors or biases in the data, it may lead to incorrect
decision-making and analysis results.
(2) Integrity of data. Collecting more complete data can lead to more accurate mineral
deposit prediction, such as specific information about known mineral deposits.
(3) Accuracy of data. In geological work, geological mapping, geophysical prospecting,
geochemical prospecting, and remote sensing are all carried out at a certain scale. In
practical work, the higher the precision and resolution, the more beneficial it is for
predicting results.
(4) The timeliness of data. As the times change and people’s perceptions evolve, some
data may become outdated, so we need to regularly filter the data we collect.
Minerals 2024, 14, 1021 5 of 35
regional faults such as the Jiaojia Fault and Zhaoping Fault, which are formed in the
extensional tectonic background [49].
The structure plays an important role in the formation of the deposit. With the increase
in the distance from the tectonic plane, the fluidity of the fluid decreases, the heat loss
increases, and the formation of the deposit becomes difficult. Therefore, the structure
plays an important role in controlling and guiding the mineralization process, and the
study of the structural characteristics is of great significance for prospecting and prediction.
Structural information typically includes mineralizing structures, host structures, structures
controlled by faults, folds, deep-seated faults, and so on. Like geological data, the defect of
structural data is that structural surfaces have dips, and predicting in plain view is affected
by the distribution of structures on the ground.
Minerals 2024, 14, 1021 introduces the algorithms which are widely used in prospecting prediction and which
9 of 35
have made some research progress.
(a) (b)
(c) (d)
Figure 1. 1.
Figure High-frequency
High-frequencykeyword
keywordco-occurrence
co-occurrenceand
and timeline
timeline chart of main
chart of mainkeywords
keywordsininCNKI
CNKIand
and WOS (made by Citespace): (a) High-frequency keywords co-occurrence in CNKI; (b) The time-
WOS (made by Citespace): (a) High-frequency keywords co-occurrence in CNKI; (b) The timeline
line chart of main keywords in CNKI; (c) High-frequency keywords co-occurrence in WOS; (d) The
chart of main keywords in CNKI; (c) High-frequency keywords co-occurrence in WOS; (d) The
timeline chart of main keywords in WOS.
timeline chart of main keywords in WOS.
[Link].
Deep
DeepAutoencoder
Autoencoder
AnAn autoencoder
autoencoder (AE)
(AE)is is
ananunsupervised
unsupervised feature
featurelearning
learningnetwork
network that utilizes
that a a
utilizes
backpropagation algorithm so that the target output value equals the input
backpropagation algorithm so that the target output value equals the input value. The value. The
autoencoder
autoencoder consists of an
consists ofinput layer,layer,
an input a hidden layer,layer,
a hidden and anandoutput layer. After
an output [Link]
inputthe
layer
inputis input,
layer isthe datathe
input, aredata
compressed and encoded
are compressed in the hidden
and encoded layer to
in the hidden enhance
layer the
to enhance
characteristics of the data,
the characteristics of theand then
data, anddecoded [30,76]; the
then decoded encoding
[30,76]; Formula (1)
the encoding and decod-
Formula (1) and
ingdecoding
FormulaFormula
(2), respectively, are
(2), respectively, are
𝑍 = φ Weθx (1)
Z = φ(Weθx) (1)
X=
X =W
WDZZ==WWφ Weθx
D φ(Weθx)
(2)(2)
where
where x represents
x represents thethe
training
trainingdata (including
data (including bias), WeWe
bias), represents
represents thethe
encoder,
encoder, φ repre-
φ repre-
sents the activation function, and Z represents the output value. W
sents the activation function, and Z represents the output value. WD represents thethe D represents de-
decoder,
coder, which translates the output value to the input value. Since the
which translates the output value to the input value. Since the input data and output input data and out-
putvalues
values of of
thethe autoencoder
autoencoder model
model unitunit follow
follow the basic
the basic logiclogic of a single
of a single real neuron
real neuron [77], the
[77], the decoding
decoding operation
operation processprocess
function function is an identity
is an identity functionfunction
to obtain tothe
obtain
output thevalue
output and
value
the and the encoded
encoded input value.
input value.
TheTheDeep
Deep Belief Network
Belief Network (DBN)
(DBN) was
wasproposed
proposed byby
Hinton
Hinton in in
2006.
[Link] is
It aisprobabilistic
a probabilistic
generation model composed of stacked constrained Boltzmann
generation model composed of stacked constrained Boltzmann sets. By training sets. By training the
the weights
weights
between between neurons,
neurons, the whole
the whole neural neural
networknetwork can generate
can generate training
training datadata according
according to the
to the maximum
maximum probability
probability [30,78,79].
[30,78,79]. TheThedeepdeep autoencoder
autoencoder (DAE)
(DAE) is developed
is developed on deep
on deep belief
belief networks.
networks. First,First, a model
a model is trained
is trained to learntothe
learn the pattern
pattern distribution
distribution of the the
of the sample, sample,
sample
is reconstructed by minimizing the difference between the output value and the input value,
and then the sample with a high reconstruction error is identified as an anomaly.
Minerals 2024, 14, 1021 10 of 35
The autoencoder can train samples without difference, reduce the extraction of irrele-
vant feature information, and it has a good effect on data dimensionality reduction [4,80].
Since the autoencoder has a high reconstruction error for abnormal samples and a low
reconstruction error for background samples, it has good applicability in processing geo-
physical, geochemical, and remote sensing data and can quickly identify data anomalies.
According to the metallogenic law and prospecting practice, the formation and discov-
ery of the deposit are small probability events. Anomalies in the data associated with
mineralization often belong to low-probability samples, which have little contribution to
the autoencoder, so their coding and reconstruction will be poor and have relatively high
reconstruction errors [29,30,81]. We can use these data with large reconstruction errors to
identify outliers in ore prospecting and prediction.
network for prospecting prediction, they often used its super ability to generate sample
sets for data enhancement operations, and the subsequent prospecting prediction modeling
process used other deep learning algorithms. Therefore, when predecessors used the
generative adversarial network to make prospecting predictions, they often used its super-
strong sample generation ability to enhance the data, and the generated data had better
consistency with the original data. On this basis, the researchers combined it with other
deep learning methods to make prospecting predictions.
Figure 2.
Figure 2. AA basic
basic LSTM
LSTM block
block [98].
[98]. ft,
ft, it,
it, and
and Ot
Ot are
are outputs
outputs of
of the
the forget
forgetgate,
gate,the
theinput
inputgate,
gate,and
andthe
output gate, respectively, at time step t. C − 1 and Ct correspond to the cell state
the output gate, respectively, at time step t.t Ct − 1 and Ct correspond to the cell state of the time of the time t and
t
∼
and
timetime
t − 1. Ct represents
t −C1.t represents thethe fresh
fresh information
information brought
brought intime
in at at time
stepstep t. σL
t. σL andand
σRσRareare used
used as as
the
the activation
activation function.
function. + represents
+ represents addition.
addition. × represents
× represents multiply.
multiply.
Recurrent
Recurrent neural
neural networks
networks focus
focus on
on learning
learning the
the interaction
interaction between
between variables
variables that
that
interact
interact with
with each
each other
other andand have
have the
the potential
potential to
to integrate
integrate highly
highly correlated
correlated geological
geological
features
features for
for prospecting
prospecting prediction.
prediction. LSTM
LSTM isis capable
capable of
of capturing
capturing contextual
contextual information
information
within
within sequences,
sequences, which
which complements
complements the the limitations
limitations ofof the
the CNN
CNN inin context
context sequence
sequence
modeling
modeling and is the first choice for interpreting geochemical elements related
is the first choice for interpreting geochemical elements related to genesis to genesis
[99].
[99].
This This advantage
advantage of recurrent
of the the recurrent neural
neural network
network plays
plays a very
a very important
important role
role in in
thethe use
use of
text-based
of text-based geological
geologicaldata,
data,which
whichcan canintegrate
integratethe
theoriginal
original data
data and extract important
important
information related
information related toto mineralization.
mineralization.
4. Application
4. Applicationof ofDeep
DeepLearning
LearningininMineral
Mineral Prospectivity
Prospectivity Mapping
Mapping
4.1. Application of DAE
4.1. Application of DAE
The deep autoencoder encodes the input data, and the decoder reconstructs the output
The deep autoencoder encodes the input data, and the decoder reconstructs the out-
from the encoding. Because the small probability sample has little contribution to the
put from the encoding. Because the small probability sample has little contribution to the
autoencoder and a high reconstruction error, it has a good advantage in the identification
autoencoder and a high reconstruction error, it has a good advantage in the identification
of geochemical anomalies.
of geochemical anomalies.
Xiong et al. [30] used the deep autoencoder algorithm to identify geochemical anomaly
Xiong et al. [30] used the deep autoencoder algorithm to identify geochemical anom-
data. Firstly, the deep autoencoder network was constructed, the improved unsupervised
aly data. Firstly, the deep autoencoder network was constructed, the improved unsuper-
building module (Continuous Restricted Boltzmann Machine) was selected as the basis
vised building module (Continuous Restricted Boltzmann Machine) was selected as the
of the deep autoencoder, and the initial weights were pre-trained. Attempts were made
basis of the deep autoencoder, and the initial weights were pre-trained. Attempts were
to find and enhance correlations between visible and hidden cell values, creating neural
made to find
networks. Theand enhancedescent
gradient correlations between
technique was visible
used toand hidden
adjust cell values,
the network in creating
order to
neural networks. The gradient descent technique was used to adjust the network
minimize reconstruction errors. Secondly, data preprocessing was carried out, the drainagein order
to minimize
sediment reconstruction
geochemical errors.
data in Secondly,
the study data selected
area were preprocessing was data,
as the basic carried
theout, the
closure
problem of the basic data was processed by using isometric logarithmic transformation,
and the measured values were normalized to the range [0, 1]. Then the model was trained
and the parameters were adjusted. Finally, the identification of geochemical data anomalies
was carried out. Xiong et al. [30] believe that when applying the deep autoencoder, it is
necessary to focus on the setting of learning rate and iteration number, as well as the size
drainage sediment geochemical data in the study area were selected as the basic data, the
closure problem of the basic data was processed by using isometric logarithmic transfor-
mation, and the measured values were normalized to the range [0,1]. Then the model was
trained and the parameters were adjusted. Finally, the identification of geochemical data
Minerals 2024, 14, 1021 anomalies was carried out. Xiong et al. [30] believe that when applying the deep autoen- 13 of 35
coder, it is necessary to focus on the setting of learning rate and iteration number, as well
as the size of the hidden layer of the geochemical data. When the reconstruction error is
minimal
of and stable,
the hidden thethe
layer of autoencoder
geochemical network and itsthe
data. When corresponding
reconstructionparameters are opti-
error is minimal
malstable,
and for thethe
modeling of geochemical
autoencoder network and samples. The results show
its corresponding that theare
parameters geochemical
optimal forhigh
the
modeling
anomaly areaof geochemical
accounts forsamples.
2.4% of theThe results
total area show that the
and 31.5% geochemical
of the total known high
ironanomaly
deposits
area accounts
(Figure 3). Theformedium
2.4% of the total area
anomaly and
area 31.5% offor
accounts the34.1%
total known iron deposits
of the total area and(Figure
68.4% 3).
of
The medium anomaly area accounts for 34.1% of the total area and 68.4% of the
the known iron reserves. Compared with the recognition results of the constrained Boltz- known iron
reserves. Compared
mann machine, with the
the results recognition
of the results of have
deep autoencoder the constrained Boltzmann
similar spatial machine,
distribution char-
the results ofwhich
acteristics, the deep autoencoder
proves that the have
deep similar spatialhas
autoencoder distribution characteristics,
a good ability which
in geochemical
proves
anomaly that the deep autoencoder has a good ability in geochemical anomaly recognition.
recognition.
Thedeep
The deepautoencoder
autoencoderplays
playsananimportant
importantrole roleininidentifying
identifyinganomalies,
anomalies,but butthe
thetype
type
of
ofanomaly
anomalydata dataassociated
associatedwith
withprospecting
prospecting prediction
prediction requires researchers
requires researchers to analyze the
to analyze
geological evolution process, that is, the judgment of expert experience
the geological evolution process, that is, the judgment of expert experience is needed as is needed as the
basis for data
the basis adoption.
for data When
adoption. When thethe
autoencoder
autoencoder is is
used
usedforformetallogenic
metallogenicprediction,
prediction,the the
abnormal
abnormalregions
regionsare
areoften
oftenextracted
extractedincorrectly
incorrectlyduedueto tothe
thehigh
high noise
noise of
of the
the original
original data,
data,
so
sosome
somemethods
methodsneedneedtotobe
be used
used toto suppress
suppress the the noise
noise andand reconstruct
reconstruct the
the error.
error. Research
Research
showed
showedthatthatincreasing
increasingthe
thenumber
numberof ofhidden
hiddenunits
unitsin inthe
thedeep
deepautoencoder
autoencodernetwork
networkcan can
effectively reduce the error between the input data and the generated output
effectively reduce the error between the input data and the generated output value, but value, but this
method will greatly
this method increase
will greatly the risk
increase theofrisk
overfitting of the of
of overfitting model [81], so[81],
the model we need
so wetoneed
adjustto
the parameters repeatedly when using
adjust the parameters repeatedly when using [Link].
4.2.
4.2. Application
Application of
of CNN
CNN
The convolutional neural network is the most widely used deep learning algorithm in
The convolutional neural network is the most widely used deep learning algorithm
mineral prospectivity mapping at present, including LeNet, AlexNet, VggNet, GoogleNet,
in mineral prospectivity mapping at present, including LeNet, AlexNet, VggNet, Goog-
ResNet, U-Net, and other structural types. The main difference between types lies in the
leNet, ResNet, U-Net, and other structural types. The main difference between types lies
structure and depth of the network (Figure 4). The LeNet structure is relatively simple, and
in the structure and depth of the network (Figure 4). The LeNet structure is relatively
people usually choose the Sigmond function or Tanh function as the activation function.
simple, and people usually choose the Sigmond function or Tanh function as the activation
Liu et al. [33] applied the convolutional neural network based on LeNet to mine the cou-
function. Liu et al. [33] applied the convolutional neural network based on LeNet to mine
pling correlation between the distribution characteristics of elements and the underground
placement space of the ore body, and the accuracy rate reached 93%. Xu et al. [100] used
the CNN of the LeNet structure to effectively learn the relationship between geological,
geophysical, and geochemical data and deposit distribution locations, and the results
showed that 90% of known gold deposits were distributed in prospective areas, accounting
for only 15% of the studied area. During training, temporarily dropping certain units in
the neural network from the network according to a certain probability is Dropout. The
AlexNet structure is deeper than the LeNet structure, uses the ReLu function as an acti-
[100] used the CNN of the LeNet structure to effectively learn the relationship between
geological, geophysical, and geochemical data and deposit distribution locations, and the
results showed that 90% of known gold deposits were distributed in prospective areas,
accounting for only 15% of the studied area. During training, temporarily dropping cer-
Minerals 2024, 14, 1021 tain units in the neural network from the network according to a certain probability is
14 of 35
Dropout. The AlexNet structure is deeper than the LeNet structure, uses the ReLu func-
tion as an activation function to reduce gradient disappearance [101], and performs data
enhancement
vation function during training,
to reduce adding
gradient Dropout to suppress
disappearance [101], andoverfitting.
performs data Li et enhancement
al. [102] used
the
during training, adding Dropout to suppress overfitting. Li et al. [102] used thegeochemical
CNN of the AlexNet structure to learn the internal relationship between CNN of the
data,
AlexNet sedimentary
structure tostrata,
learn structure,
the internalwater system,between
relationship and other geologicaldata,
geochemical information
sedimentary and
the location of ore occurrence, and delineated the metallogenic prospect
strata, structure, water system, and other geological information and the location of ore area, and the ver-
ification
occurrence, accuracy rate was 86.21%.
and delineated Li et al. [103]
the metallogenic predicted
prospect sedimentary
area, and manganese
the verification accuracyore
prospecting
rate was 86.21%. in theLiSongtao–Huayuan
et al. [103] predicted areasedimentary
based on themanganese
AlexNet network, and obtained
ore prospecting in thea
classification model based on the CNN, with an accuracy of 88.89%.
Songtao–Huayuan area based on the AlexNet network, and obtained a classification model VggNet [9,104] has a
simple structure and reduces the number of weights by stacking
based on the CNN, with an accuracy of 88.89%. VggNet [9,104] has a simple structure multiple 3 × 3 convolu-
tion cores instead
and reduces of largeofconvolution
the number weights by cores.
stacking It also uses ReLU
multiple as the activation
3 × 3 convolution coresfunction
instead
after the fully connected layer to suppress overfitting, which
of large convolution cores. It also uses ReLU as the activation function after the fullyis highly applicable. The
GoogleNet
connected layer structure, designed
to suppress by the Google
overfitting, whichteam in 2014,
is highly can extract
applicable. Thefeatures
GoogleNet from dif-
struc-
ferent convolution
ture, designed by thekernels
Googleofteamfeature images
in 2014, canin parallel,
extract enriching
features the information
from different convolutioncon-
tained
kernelsinofmulti-scale
feature images featurein maps. Multi-source
parallel, enriching the geological data provide
information contained theinbasis for fea-
multi-scale
tures to integrate feature information with the GoogleNet structure,
feature maps. Multi-source geological data provide the basis for features to integrate making this method
more
feature suitable
informationfor prospecting and prediction
with the GoogleNet [87,105].
structure, ResNet
making thiswas
methodproposed
more by He et for
suitable al.
[106] in 2016,
prospecting andaiming to introduce
prediction [87,105].a ResNet
deep residual learning
was proposed byframework to solve
He et al. [106] in 2016,theaiming
prob-
lem of training
to introduce a deepaccuracy
residual degradation by adding
learning framework to residual
solve the structure.
problem ofResNet
trainingartificially
accuracy
makes certainbylayers
degradation adding of the neural
residual networkResNet
structure. skip theartificially
connections of the
makes next layer
certain layersofofneu-the
neuralconnecting
rons, network skip the the connections
layers separately,ofweakening
the next layer the of neurons,
strong connecting
connections the layers
between each
separately,
layer. weakeninginput
By combining the strong connections
and output, ResNet between eachalleviates
effectively layer. By thecombining input and
loss of geo-infor-
output, ResNet effectively alleviates the loss of geo-information
mation caused by convolution operations, and also plays a positive role in solving the caused by convolution
operations,
problem of and also plays
gradient a positive role
disappearance in solvinginthe
or explosion problem
deep of gradient
networks disappearance
[107,108]. U-Net is a
or explosionupdated
completely in deep convolutional
networks [107,108].
neural U-Net
network is awith
completely updated convolutional
an encoder–decoder structure.
neural
The network
encoder with an
extracts encoder–decoder
features structure. The
through convolution, encoder
pooling, extracts
etc., features through
and gradually reduces
convolution, pooling, etc., and gradually reduces the input
the input dimension. According to the data provided by the encoder, the decoder can dimension. According to the
re-
data provided by the encoder, the decoder can repair the detailed
pair the detailed features to improve the accuracy. Its advantage are that the required features to improve the
[Link]
training Its advantage
are smaller,are thethat the is
result required trainingand
more accurate, dataitare
hassmaller, the result
an advantage is more
in the pro-
accurate, and it has an advantage
cessing of small-scale data [109–111]. in the processing of small-scale data [109–111].
Figure 4. CCN structures for mineral prospectivity modeling (modified from [9,87,107]): (a) Different
CCN structures for mineral prospectivity modeling: a) LeNet, b) AlexNet, c) VggNet, d) GoogleNet;
e) ResNet-50; (b) A basic structure for convolution; (c) the detailed process for “Inception” in
GoogleNet; (d) the detailed process for “Identity Block” in ResNet-50.
The Nanling area is one of the most important non-ferrous metal metallogenic belts in
China, and tungsten tin is the dominant mineral. There are several world-class tungsten tin
deposits in this area, such as the Xihuashan tungsten deposit, Dachang Tin deposit, and
ent CCN structures for mineral prospectivity modeling: a) LeNet, b) AlexNet, c) VggNet, d) Goog-
leNet; e) ResNet-50; (b) A basic structure for convolution; (c) the detailed process for “Inception” in
GoogleNet; (d) the detailed process for “Identity Block” in ResNet-50.
Minerals 2024, 14, 1021 The Nanling area is one of the most important non-ferrous metal metallogenic 15 belts
of 35
in China, and tungsten tin is the dominant mineral. There are several world-class tungsten
tin deposits in this area, such as the Xihuashan tungsten deposit, Dachang Tin deposit,
and Shizhuyuan
Shizhuyuan tungsten
tungsten deposit
deposit [112].[112].
Li et Li
al. et [Link]
[36] [36] made
use ofuse of multivariate
multivariate data indata
theinarea
the
area and adopted the convolutional neural network model to
and adopted the convolutional neural network model to make prospecting predictions. make prospecting predic-
tions.
The The selected
selected dataincluded
data layers layers included the following:
the following: granite and granite
fault;and
localfault; localanomaly
gravity gravity
anomaly closely related to W-Sn ore mineralization; distribution of
closely related to W-Sn ore mineralization; distribution of W, Sn, Bi, Be, Pb, Ag, Mo, and W, Sn, Bi, Be, Pb, Ag,
Mo,geochemical
Zn and Zn geochemical
elements. elements.
The bufferThe zone buffer zone was according
was delimited delimited to according to the
the distance of dis-
the
tance ofand
granite thefault,
granite and fault,and
respectively, respectively,
the buffer zoneand the
mapbuffer zone mapThe
was generated. wasgeophysical
generated. and The
geophysical data
geochemical and geochemical
were mappeddata were mapped
according to inverse according
distance to inverseDue
weights. distance weights.
to insufficient
Due to insufficient training samples of geological data, the researchers
training samples of geological data, the researchers used sliding windows and random used sliding win-
dows and random zero noise
zero noise for data enhancement. for data enhancement.
Sincethere
Since thereare
aremany
many channels
channels in the
in the geological
geological datadata
layer,layer, the traditional
the traditional convolu-
convolutional
tional neural
neural network network is complicated
is complicated to calculate
to calculate and itand it iseasy
is not not to
easy to automatically
automatically extract
extract key
key channel
channel information.
information. A feature
A feature of the
of the human
human visual
visual systemisistotoselectively
system selectivelyfocus
focus onon
highlighted parts of the
highlighted the entire
entire scene
scenetotobetter
bettercapture
capturethe thevisual
visualstructure.
structure. Therefore,
Therefore, re-
searchers add
researchers addananattention
attention mechanism
mechanism based
basedon the characteristics
on the characteristicsof theof human
the human eye after
eye
after the convolutional
the convolutional layer layer to obtain
to obtain the channel
the channel attention
attention graph,graph,
whichwhich
enhances enhances the
the repre-
representation
sentation of key of features
key features
and and weakens
weakens the performance
the performance of channels
of channels withwith
low low weight
weight val-
values.
ues. Its Its network
network structure
structure is shown
is shown in in Figure
Figure 5. 5.
Figure5.5. ATT-CNN
Figure ATT-CNNstructure
structurefor
formineral
mineralprospectivity
prospectivitymodeling
modeling(modified
(modifiedfrom [36]).
from[36]).
rate will affect the convergence rate of the model. If the learning rate is too small, the
convergence speed will be fast, and if the learning rate is too large, it is difficult to reach
the extreme point [9]. In practical application, it is necessary to consider the density and
cartographic scale of different types of data samples, and carry out continuous debugging
Minerals 2024, 14, 1021 16 of 36
to select the appropriate mesh size, convolutional kernel size, neural network layer number,
learning rate, and other parameters.
Figure 6.
Figure 6. MPM
MPM of
of W-Sn ore obtained
W-Sn ore obtained by
by using
using sliding
sliding window
window ATT-CNN
ATT-CNN (modified
(modified from
from[36]).
[36]).
The following
4.3. Application of RNN problems should also be paid attention to when using the convolu-
tionalThe
neural network
recurrent for prospecting
neural network hasprediction
the ability modeling:
to integratethe highly
size of the grid, thegeologi-
correlated size of
the features,
cal convolutional
is very kernel,
friendly theto
number of layers,
geological data inthegeological
learning rate, and the and
prospecting number of iter-
prediction,
ations.
and hasThe smaller
a good grid can make
application betterinuse
prospect of the existing
research areas with dataabundant
and makegeological
the calculation
data.
more accurate, but it will lead to excessive calculation. If
Wang et al. [96] made use of geological, geochemical, and geophysical data in the the grid is too large, it cannot
study
effectively
area use
to carry out allweight
the data, resulting
function in reduced
analysis based on accuracy [102]. Zuo
the singularity [58] studied
of deposit the size
location, and
of theaoutput
built long and unitshort
and proved
term memorythat different
network grid sizesfor
model havethecertain
evidenceeffects on the
layer. Theconcen-
results
tration distribution
showed that all known of geochemical
deposits fell in anomalies and the texture
the high prospect area, and structure
10% of theof different
prospect geo-
area
chemical types.
accounted for more Complex
than 90% neural networks
of the iron oredo not necessarily
deposits. improve
By combining the accuracy
LSTM with the of the
CNN,
[Link]
Wang et al.
al. [99] [22] found
focused on the that in artificial
inherent featureneural networks,ofthe
representation mean squared
adjacent samples error
and theof
the model does not decrease significantly with the increase in the
contextual association information learning of geochemical variables, and captured 96% of number of neurons. The
learning
the graniterate
in will
15%affectof thethe convergence
study rate ofathe
area, providing new model. If the
idea for thelearning rate isoftoo
identification small,
regional
geological features
the convergence and will
speed contributing
be fast, andto the mineral
if the prospectivity
learning rate is toomapping
large, it of areas with
is difficult to
insufficient mapping.
reach the extreme point [9]. In practical application, it is necessary to consider the density
The Baguio District
and cartographic scale ofofdifferent
the Philippines
types ofisdata
one samples,
of the most andimportant
carry outgold depositsde-
continuous in
the world,
bugging to with
selectproven gold reserves
the appropriate mesh ofsize,
moreconvolutional
than 800 tons,kerneland the gold
size, deposits
neural are
network
mainly volcanic
layer number, epithermal
learning deposits.
rate, and Yin et al. [97] used four types of geological data
other parameters.
within the region, such as NE faults, NW faults, Agno magmatite margins, and porphyry
intrusive contact
4.3. Application of zones,
RNN as data layers to build models, built buffers and converted grids
fromThethe four types
recurrent neural of geological
network data,
has and used nonlinear
the ability to integrate control
highly functions
correlatedof geological
geological
characteristics to assign artificial values to grids in all buffers.
features, is very friendly to geological data in geological prospecting and prediction, andThe data enhancement
method
has a goodwasapplication
used to generateprospect sufficient samples.
in research areas with abundant geological data. Wang et
al. [96] made use of geological, geochemical, and geophysical data in the study area to
carry out weight function analysis based on the singularity of deposit location, and built
a long and short term memory network model for the evidence layer. The results showed
that all known deposits fell in the high prospect area, and 10% of the prospect area ac-
counted for more than 90% of the iron ore deposits. By combining LSTM with the CNN,
Wang et al. [99] focused on the inherent feature representation of adjacent samples and
the world, with proven gold reserves of more than 800 tons, and the gold deposits are
mainly volcanic epithermal deposits. Yin et al. [97] used four types of geological data
within the region, such as NE faults, NW faults, Agno magmatite margins, and porphyry
intrusive contact zones, as data layers to build models, built buffers and converted grids
Minerals 2024, 14, 1021 from the four types of geological data, and used nonlinear control functions of geological 17 of 35
characteristics to assign artificial values to grids in all buffers. The data enhancement
method was used to generate sufficient samples.
The GRU
The GRU model
model inin the
the recurrent
recurrent neural
neural network
networkwaswasselected
selectedtotobuild
buildthe
themodel,
model,
which includes a cyclic layer, a dense layer, and an activation layer. Its network structureis
which includes a cyclic layer, a dense layer, and an activation layer. Its network structure
is shown
shown in in Figure
Figure 7. 7.
Figure [Link]
Figure Datarepresentation
representationfor
forthe
the GRU
GRU model
model (modified
(modified fromfrom [97]
[97]): (a)):Basic
(a) Basic process
process of theofmodel;
the
model; (b) Inner structure of the hidden layer; (c) Inner structure of the GRU. The black arrows
(b) Inner structure of the hidden layer; (c) Inner structure of the GRU. The black arrows represent
represent normal flow and the red arrows represent recurrent connection.
normal flow and the red arrows represent recurrent connection.
When using
When using recurrent
recurrent neural
neural network
networkalgorithms,
algorithms,the theordering
orderingofofthe
theevidence
evidencelayers
layers
is critical to the accuracy and predictive power of the final model,
is critical to the accuracy and predictive power of the final model, and sinceand since the GRU can
the GRU
process sequence data through special cyclic states, 24 different input orders
can process sequence data through special cyclic states, 24 different input orders can can be se-be
lected for the four geological evidence layers to model. By analyzing the accuracy,
selected for the four geological evidence layers to model. By analyzing the accuracy, AUC AUC
value, Kappa
value, Kappa value,
value, Matthews
Matthews correlation
correlationcoefficient
coefficient(MCC),
(MCC),and
andrecall
recallrate
rateofof2424different
different
order models, the researchers determined the optimal evidence layer [Link]
order models, the researchers determined the optimal evidence layer ranking. Theresults
results
(Figure 8)
(Figure 8) show
show that
that the
the sequence
sequence beginning
beginningwith
withthe
theporphyry
porphyryintrusion
intrusioncontact
contactzone zonehas
has
better accuracy, while the sequence ending with the porphyry intrusion contact
better accuracy, while the sequence ending with the porphyry intrusion contact zone has zone has
Minerals 2024, 14, 1021 the worst
the worstaccuracy.
[Link] Inthe
themineral
mineralpotential
potentialmap,
map,thethe extremely
extremely high
high anomaly
anomaly areaarea con-
18 of 36
contains
tains 18 deposits, accounting for only 19.23% of the surface, showing
18 deposits, accounting for only 19.23% of the surface, showing good accuracy. good accuracy.
Figure 8.
Figure 8. Mineral
Mineral potential
potential analysis
analysis in
in Baguio
Baguioregion
regionbased
basedon
onGRU
GRUmodel
model(modified
(modifiedfrom
from[97]).
[97]).
experimental results
The experimental results show
show that
that the
theknown
knowngoldgolddeposits
depositsare
aremainly
mainlydistributed
distributedin
in the
the extremely
extremely high
high andand high
high occurrence
occurrence areas
areas after
after the the classification
classification of the
of the prospect
prospect map
map generated by the GRU by the quantile discontinuous method. The results
generated by the GRU by the quantile discontinuous method. The results of the success of the suc-
cess curve
rate rate curve evaluation
evaluation showshow
that that the GRU
the GRU method
method has has better
better prediction
prediction ability.
ability.
Figure 9.
Figure 9. Geochemical
Geochemical anomalies
anomalies associated
associated with
with polymetallic
polymetallicmineralization
mineralizationclassified
classifiedbybythe
the
SMOTified GAN (modified from [115]).
SMOTified GAN (modified from [115]).
4.5. Application
4.5. Application of Mixed Algorithm
Algorithm
Deep learning
Deep learning has a strong ability
ability in
in data
data mining,
mining,and
anddifferent
differentdeep
deeplearning
learningalgo-
algo-
rithms have
rithms have slightly
slightly different
different capabilities
capabilities and
and focuses
focuses in
in practical
practicalapplications.
[Link]
Withthe
the
continuousdevelopment
continuous developmentofoftraditional
traditionalmachine
machinelearning
learningandand deep
deep learning
learning algorithms,
algorithms, the
the combination of different algorithms has become the main way for researchers to avoid
their own defects.
The deep autoencoder can be directly used for prospecting prediction. However,
since the noise of its own data is generally large, and the prediction results are easily af-
fected by noise, combining it with the convolutional neural network can help eliminate
the influence of noise. Zhang et al. [116] use the Convolutional Autoencoder (CAE) to learn
meaningful abstract representations of multi-source geographic information at different
Minerals 2024, 14, 1021 19 of 35
combination of different algorithms has become the main way for researchers to avoid their
own defects.
The deep autoencoder can be directly used for prospecting prediction. However,
since the noise of its own data is generally large, and the prediction results are easily
affected by noise, combining it with the convolutional neural network can help eliminate
the influence of noise. Zhang et al. [116] use the Convolutional Autoencoder (CAE) to learn
meaningful abstract representations of multi-source geographic information at different
scales, and determine the degree of influence of reconstruction errors of each data variable
by reducing the removal of evidence variables. The types of geological data that have a
greater influence on the reconstruction error are analyzed and the factors that have a greater
correlation with mineralization are determined. The core of the CAE is to combine the
convolutional operation with the autoencoder structure to form the convolutional encoder
and the convolutional decoder, so as to optimize the training results. The success rate curve
shows that the predicted potential mineralization area accounts for 23.8% of the study area
and contains 77.8% of known gold deposits.
Xie et al. [80] used the combination of the autoencoder and the GAN to conduct
metallogenic prediction in the Lhasa area, which not only ensured the stability of the
training process, but also ensured the clarity of the training results. The AUC value of the
model reached 0.95. The Convolutional neural network mixed the channel information
in the process of feature extraction. The combination of the CAE and CNN [29,116] can
remove the data types in the evidence layer that have little influence on the reconstruction
error, and determine the factors with greater correlation with mineralization. The prediction
accuracy AUC value can reach 0.863 and 0.908 in the two regions, respectively.
In order to solve the problem of the lack of labeled data in deep learning algorithms,
Li et al. [114] proposed a data enhancement method based on generative adversarial net-
works, used the enhanced data to carry out research on the convolutional neural network
Minerals 2024, 14, 1021 algorithm, and conducted prospecting prediction analysis on rare earth deposits in 20 south-
of 36
ern Jiangxi Province, China. The results (Figure 10) show that the GAN can learn the
internal structural features of real samples, and can synthesize enhanced data from random
inputs. The model training
inputs. training accuracy
accuracy reaches
reaches 99.7%,
99.7%,and
andthe
thevalidation
validationaccuracy
accuracyreaches
reaches
98.9%. The research also shows that
98.9%. that although
although“black
“blackbox”
box”isisaaproblem
problemthat
thatdeep
deeplearning
learning
cannot avoid,
cannot avoid, the analysis of
of prediction
predictionresults
resultscombined
combinedwith
withgeological
geologicaland
andgeochemical
geochemical
data can
data can show
show the reliability and accuracy
accuracy of
of deep
deep learning
learningalgorithm
algorithmapplication.
application.
Figure 10.
Figure 10. Regolith-hosted
Regolith-hostedREE
REEdeposits
depositsprospectivity
prospectivitymap
mapderived
derivedby
bythe
theGAN
GANand
andCNN
CNN(modified
(modi-
fied from [114]).
from [114]).
The combination of different algorithms will provide more ideas and methods for
mineral prospectivity mapping.
5. Discussion
5.1. Preprocessing of Geological Data
Minerals 2024, 14, 1021 20 of 35
The combination of different algorithms will provide more ideas and methods for
mineral prospectivity mapping.
5. Discussion
5.1. Preprocessing of Geological Data
Geological data are generally text data, geophysical and geochemical data are often
point data, and remote sensing data are image data. Different data types need to be
preprocessed in different ways when prospecting predictions are made. The uses of various
data types are shown in Table 2.
Geological data generally do not include digital information, and it is difficult for
machine learning algorithms to read these text data; geological data use generally two
types of methods. One is to select ore-controlling strata and magmatic rocks closely related
to mineralization as geological data sources, define different lithologies as different values,
and simply classify according to lithology; strata and magmatic rock data mostly adopt
this method. Zheng et al. [35] classify different strata according to 1, 2, 3. . . The unknown
geologic body is coded as 0, and the lithology value directly defined during grid training is
taken as the grid value. The other type is interpolated according to the distance between
strata, structures, and known deposits, the buffer distance is delimited according to the
distance size, and the distance value is used directly as the grid value. Researchers often
measure the impact of geological formations using the distance from the target location to
the structure, converting the data into a distance buffer map that retains distance, position,
and orientation information in an image that the machine can recognize. The types of buffer
distance graph mainly include the discrete buffer distance graph and continuous buffer
distance graph. In the discrete buffer distance graph, the same distance value is in the
same “step size” [117,121,122], which simplifies the distance information and makes the
calculation more convenient, but this method will reduce the accuracy of the construction
information expression. The continuous buffer distance graph uses the continuous value to
represent the distance between the target point and the structure, and the result is more
accurate and complete [9,22]. Xiong et al. [121], Sun et al. [22], and Yang et al. [9] used
the continuous buffer distance to measure the influence range of the geological structure,
which better preserved the detailed information of the geological structure in the image,
and improved the ability of deposit prediction. In addition, Sun et al. [22] also used the
distance between the magma intrusion contact site and the target location and the density
of regional faults as the evidence layer of geological characteristics to participate in the
modeling. Li et al. [102] used convolutional neural networks to explore the coupling
relationship between geological data such as strata, faults, and water systems in the study
area and the mining area. In addition, using a natural language model to process earth
science text has certain potential in mineral prospect modeling, and has been used in some
meaningful applications [123].
Minerals 2024, 14, 1021 21 of 35
It is worth noting that in the process of geological data application, most researchers
did not consider the comprehensive use of the relevant characteristics of the whole process
of deposit formation, such as source, transport, trap, sedimentation and enrichment, weath-
ering and denudation, etc., and the data types selected were relatively simple, which may
have a certain impact on the predicted results. Therefore, it is also necessary to strengthen
the exploration of the content, method, and form of geological data types used in deep
learning prospecting.
The interpolation of point data sets into raster plots is a common method in geochemi-
cal mapping, and the most common method is inverse distance weighted interpolation. The
inverse distance weighted interpolation method takes the distance between the interpola-
tion point and the sample point as the weighted average, and the closer to the interpolation
point, the higher the weight [27]. The data of geochemical discrete points can be interpo-
lated into the geochemical continuous concentration map, which has a good smoothing
effect on the original data. In the past, various univariate or multivariate geochemical sta-
tistical analysis methods were used to describe geochemical data and anomalies, requiring
normal or lognormal distribution of geochemical data; otherwise, geochemical anomalies
could not be correctly identified [81]. In fact, the formation of mineral deposits is a rare
event, and the basic requirement of normal or lognormal distribution of geochemical data
cannot be guaranteed. Traditional multivariate statistical methods often lose their function
when dealing with geochemical data under complex conditions, but deep learning algo-
rithms can learn the deep-level evidence information in the data without data processing.
Sometimes geochemical data are the composition of a certain substance, for example, H is a
part of H2 O, the content of H is correlated with H2 O, and the contribution of H to statistical
analysis is very likely to represent the contribution of H2 O to statistical analysis, which will
lead to closure problems in multivariate statistical analysis [124]. The researchers used the
method of isometric logarithmic ratio transformation to process the raw data.
In addition, geochemical data of rocks or mineral deposits, such as major trace ele-
ments, hydrogen and oxygen isotopes, Re-Os isotopes, U-Th-Pb isotopes, Sr-Nd isotopes,
etc., are not stable and representative of in situ data due to their sampling locations and
sampling principles different from grid sampling of ordinary geophysical and geochemical
data. Therefore, this type of data is generally not used. However, when part of the data
has in situ data characteristics and has a certain correlation with mineralization, it can be
selected as a feature of deep learning to participate in the calculation.
The geochemical data often come from or are close to the earth’s surface and may
not necessarily include key ore-controlling geological factors, which can lead to a poor
MPM result [125]. Li et al. [102] made use of the method of coupling geochemical data with
geological data, projected the information of geological data onto geochemical data, and
used deep learning to make better use of geochemical data for prospecting and prediction,
achieving good results and providing us with new ideas for using geological data.
Remote sensing data itself has the characteristics of multi-dimensional numbers, un-
even noise, and so on, which brings many problems to data processing. However, deep
learning can extract useful features from massive remote sensing data, and has great poten-
tial in mineral deposit classification and prospecting prospect prediction. The recognition of
remote sensing images related to mineralization mainly includes the recognition of a linear
ring structure and the recognition of surrounding rock alteration. Zidan et al. [68] used the
deep learning algorithm of the CNN to extract hydrothermal alteration regions related to
porphyry copper deposits based on ASTER images, and delineated the scope for further
determining the potential areas for prospecting. Sun et al. [22] used the iron oxide alteration
and mud alteration obtained from Landsat ETM+ images as evidence layers and other
geological data to conduct deep learning modeling and delineate the prospecting target
area, providing an important basis for further prospecting and exploration. Fu et al. [27]
used the CNN algorithm based on GF-5 remote sensing images, ASTER images, and geo-
chemical data to predict the mineral prospect area in the Duolong region of Tibet, and
used alteration minerals such as dolomite and kaolin, identified by the short-wave infrared
Minerals 2024, 14, 1021 22 of 35
band, as the evidence layer to construct the CNN model for prospecting prediction. The
results show that the alteration zone extracted from remote sensing images has good indi-
cation significance for prospecting prediction. Therefore, the processing of remote sensing
data mainly uses a variety of image processing techniques to extract geological structure
information and alteration information as data layers for deep learning modeling.
By summarizing the previous studies, we find that geochemical data, geological data,
and remote sensing data are widely used in prospecting prediction, the research methods
are increasing, and the varieties are becoming more and more abundant. As an impor-
tant means of prospecting and exploration, geophysical methods have more and more
diversified physical property types and data forms. As geophysical methods play an
increasingly important role in deep prospecting, geophysical data as the evidence layer
of deep learning algorithms can help improve the accuracy and generalization ability of
the model. Sun et al. [22] used gravity data, aeromagnetic exploration data, and resistivity
data in the Tongling area of Anhui Province, China to generate a magnetic anomaly pole
map, lithology density map, −800m resistivity interpretation map, and other geological
data as data layers. Based on the source, transport, and storage processes of mineralization,
the support vector machine, random forest, and machine learning algorithms were used to
model and analyze the ore prospecting prediction, the ore prospecting prediction in the
study area was evaluated, and the favorable metallogenic prediction area was delineated.
Wang et al. [96] used aeromagnetic data combined with regional granite mass, regional
major faults, and geochemical data to predict the iron ore prospecting potential in Fujian
Province, China, and the results showed the feasibility and reliability of prospecting predic-
tion by using aeromagnetic data. However, due to the multi-solution of geophysical data,
there are still some limitations in geological prospecting prediction, and the application
of geophysical data in deep learning prospecting is not extensive enough. The next step
is to combine geophysical data with geological data to improve the coupling ability of
geophysical data and ore deposits.
Sliding window is a commonly used data enhancement method. Core areas are
selected at each evidence layer in turn, a window smaller than the core area is used to
slide in the area for resampling, and then resampling results of each evidence layer are
combined. The generated window remains in the core area until the required number
of samples is generated. The advantage of sliding windows is that new data can carry
all the information in the original evidence layers and resampling. Wu et al. [52] use
sliding window technology to enhance data samples and solve the problem of scarcity and
imbalance of label data in deep learning environments.
Adding random zero noise is another processing method. By adding zero noise points
to the original evidence layer and combining the original evidence layer to achieve the
purpose of increasing data, this method can maintain the original geographical location,
correlation, and other relative information of the data, and can maintain the integrity of
the data and highlight the overall characteristics of the data. However, it is necessary to
consider the ratio of random zero noise when using it. Adding different ratios of zero noise
will make the generated sample data carry different characteristics. The purpose of adding
zero noise is to generate more samples to expand the data, and from the experience of
previous people, the random noise ratio is set between 5% and 10% for the best effect.
In addition, Brandmeier et al. [127] increased the number of training samples by
disturbing the location of the deposit and adding data points around the real deposit.
Yang et al. [9,87] used a clipping and repairing method to generate additional training
samples. This method not only ensured that the expanded sample data and the test
data had similar deposit indication characteristics and deposit location, but also ensured
that the expanded sample data and the test data had different combinations of multi-
source geographic information in different spatial locations, which has a good reference
significance. Zhang et al. [126] used the pixel-to-feature method to reassemble the pixels of
the sample to generate enough training samples.
With the development of deep learning algorithms, researchers use the GAN and other
algorithms based on sample data to generate more usable sample sets, which can retain
deeper features in the original samples. Wu et al. [52] add random noise to the original
data, combine corresponding label information from real samples to generate basic data,
and use the generated adversarial network for data enhancement, which is essentially a
comprehensive application of adding random noise and the GAN. Li et al. [114] also used
the GAN to enhance the geochemical data, which achieved good results.
which was better than that of random forest and the support vector machine (0.973 and
0.959, respectively). Wang et al. [96] conducted prospecting prediction in iron ore deposits
in Fujian, China. According to the success rate curve, the recurrent neural network and
logistic regression model accounted for more than 90% and 80% of iron ore deposits,
respectively, in 10% prospective areas, and the linear regression results were significantly
better. According to previous studies, deep learning algorithms have been applied to
various types of minerals such as Au, Fe, Cu, W-Sn, Pb-Zn, REE, etc., with good effect, and
their future application prospects are very wide.
In the field of mineral prospectivity mapping, the application of deep learning algo-
rithms is gradually increasing, which significantly improves the efficiency and accuracy
of mineral resource exploration. Different algorithms are suitable for different application
scenarios because of their unique advantages. For example, GANs are mainly used in
data enhancement, DAEs are used in data enhancement and denoising, RNNs are widely
used in text information extraction and time series data analysis, and CNNs are widely
used in geological image analysis and processing to identify mineralization features and
geological structures due to their excellent performance in image processing, providing an
important basis for prospecting prediction. Table 4 lists the application examples of deep
learning algorithms in prospecting prediction. In terms of data volume, compared with
traditional statistical analysis methods, deep learning has a very strong data processing
ability, and geological data processing of mineral deposits, mining areas, and regions can
be completed. The number of iterations, learning rate, and batch size of deep learning are
greatly affected by the quality of the original data. Due to the imbalance of positive and
negative samples, data noise, and other problems in the original data, it is often necessary
to conduct several iterations to stabilize the model. From the perspective of model accuracy
of deep learning, the AUC value is generally greater than 0.8, and in convolutional neural
networks, the AUC value can even reach more than 0.98, which is far more accurate than
traditional algorithms. It can be seen that deep learning algorithms have the ability to mine
deep features in complex data and hidden associations between data.
However, we should also note that deep learning algorithms have certain limitations,
specifically including the following questions:
Minerals 2024, 14, 1021 25 of 35
(1) The support for the algorithm by textual geological data. Deep learning algorithms
need to convert some text information into numerical information for computation. In
the application of geological information, the data can only be displayed, but a lot of
geological information will be discarded. For example, when applying fault data, the
distance between different positions and faults will be reflected, but different types of
faults will not be reflected, which is unreasonable in geological understanding. Natural
language processing (NLP) is a branch of computer science and artificial intelligence
that studies how computers understand, generate, and translate human language.
RNNs can process sequence data and are particularly effective for natural language
processing tasks, so the application of RNNs in geological data can be increased in
future research.
(2) The research on geological data enhancement methods. Data enhancement can enrich
data, but the large-scale use of data enhancement methods will increase the complex-
ity of training, sometimes resulting in some categories being “over-enhanced”, and
others being ignored, resulting in an imbalance between categories. Nowadays, many
scholars try to use the combination of transfer learning and deep learning for rock and
ore analysis and mineral prospecting prediction [108,129–131]. The robustness and
generalization ability of transfer learning are proved, the convergence speed of deep
learning is accelerated, and the learning efficiency of the prospecting prediction model
is enhanced. So, we believe that the use of transfer learning and the adversarial gener-
ation network in the future will greatly improve the reliability and logic of enhanced
data, which will greatly improve the accuracy of prospecting prediction.
(3) Avoidance and treatment of model overfitting. Overfitting often occurs when testing
with a small number of training sets, and will greatly affect the test results and the
performance of the model. The most direct way to improve the performance of
neural networks is to increase the number of parameters such as depth and network,
but this will make it easier for the updated network to reach the overfitting state,
especially when the number of positive samples, such as in prospecting prediction,
is limited. At present, in other fields, the methods to solve the overfitting of deep
learning models mainly include data preprocessing, simplifying the model structure,
adding regularization terms, adding Dropout layers, and adjusting model parameters.
In mineral prospectivity mapping, the application of Dropout layers has achieved
good results. Krizhevsky et al. [101] solve the overfitting problem by increasing the
number and size of layers while using the method of Dropout. Li et al. [117] believe
that increasing the number of layers and parameters will increase the computational
amount of learning, and adding Dropout can reduce the training time during the
training process with a large amount of data. Moreover, from the application point of
view [90], after using Dropout, the hidden neurons no longer depend on the existence
of other hidden neurons, and the co-adaptability with other neurons decreases, which
has a good effect on solving the overfitting problem. In the next research process, the
diversity of data can be increased in the data preprocessing process, and the learning
rate, batch size, and other parameters in the model can be adjusted in time.
(4) Adjustment of parameters of the model. In some research examples, we can see that
researchers often need to constantly try to modify various parameters, such as the
learning rate, the epoch, the batch size, and so on, according to the size of data or the
quality of data to obtain better prediction results. This is because the deep learning
model needs to optimize the performance of the model by adjusting parameters
during the training process, so that it can better fit the training data and generalize to
the unseen data. Therefore, we may need to continue to strengthen the self-adaptive
research of the model, so that the model can automatically adjust parameters according
to the original data and the target.
(5) The choice of backbone architecture and the effectiveness of different deep learning
algorithms. Convolutional neural network is the most widely used deep learning
algorithm in mineral prospectivity mapping at present, including LeNet, AlexNet,
Minerals 2024, 14, 1021 26 of 35
VggNet, GoogleNet, ResNet, U-Net and other structural types. The most common
methods of recurrent neural networks include Long Short-Term Memory (LSTM) and
Gated Recurrent Unit (GRU). We need to choose the backbone architecture and deep
learning algorithms according to the actual situation.
(6) The explanation of the model “black box” mechanism. The “black box” mechanism
of deep learning cannot know about the drivers of underlying phenomena and pro-
cesses [125], so the interpretability of model outputs, the relationship between data
input and output, and the internal operation mechanism need further study. At
present, the Google DeepMind team is exploring the “black box” mechanism in deep
learning from the perspective of cognitive psychology [132], but there is still a long
way to go. To enhance the explainability of deep learning, Fu et al. [27] used the
SHAP library in Python to explain the individual output results of deep learning. The
results showed that the different types of data selected for distinguishing between
the presence and absence of minerals in the study area are of significantly different
importance. Many scholars have found that the predictive results of deep learning are
not as good as those of SVM or random forest algorithms in some cases, and a major
reason for this is the type of data used [4,22].
TP
R= (5)
TP + FN
TP
P= (6)
TP + FP
2PR
F1 = (7)
P+R
R represents the recall rate, P represents the precision, TP represents the number of
actual true predictions that are true, FN represents the number of actual true predictions
that are false, FP represents the number of actual false predictions that are true, and F1
represents the harmonic mean of the recall rate and precision.
Accuracy function and loss function can also be used to evaluate the accuracy of deep
learning. A better effect needs a smaller loss function and a higher accuracy [35]. Generally,
the receiver operating characteristic curve (ROC) and Area Under the Curve (AUC) are
used for accuracy evaluation.
TP
TPR = (8)
TP + FN
FP
FPR = (9)
FP + TN
Minerals 2024, 14, 1021 27 of 35
TPR represents the true rate, FPR represents the false positive rate, and TN represents
the number of predictions that are false and actually false. It can be seen from the formula
that TPR is the same as the recall rate.
The vertical coordinate of the ROC curve is the true rate, and the horizontal coordinate
is the false positive rate. The TPR is the probability of being judged to be true in the true
sample, and the FPR is the probability of being misjudged to be true in the false sample.
The advantage of the ROC curve is that in the evaluation process, it is not necessary to add
a number as the judgment value of the evaluation, and the accuracy level of the prediction
results can be analyzed only according to the below-line area.
The Kappa coefficient is the proportion of correct classification after the removal of
accidental coincidence probability, and the calculation formula of the Kappa coefficient is
Po − Pe
kappa = (10)
1 − Pe
Po is the accuracy of prediction, that is, the consistency of model prediction results and
actual classification results. Pe is accidental consistency, that is, the consistency between
two variables (in this case, the classification result and the validation sample); even if the
two variables are completely independent, it will not be 0, and there are still cases where
chance causes the two variables to be consistent. Using the Kappa coefficient to evaluate
the reliability of prospecting prediction models can often increase the difference between
different models and make the advantages of models more prominent.
The predicted area curve composed of the prediction rate and area rate is another
method of accuracy evaluation. The intersection point of the two curves of deposit pro-
portion and occupied area proportion is taken as the evaluation standard. When the
intersection point of the two curves is closer to the top position, it means that more deposits
are contained in fewer areas, indicating that the prediction effect is more ideal [4,87,133,134].
In mineral prediction, researchers often use the success rate curve to evaluate the
accuracy of the prospecting prediction model with the success rate of known deposit
prediction, in which different slopes represent different ore-forming potentials, and the
steeper the slope, the more deposits are captured in smaller areas [4,135]. The higher the
success rate and the more deposits captured in a smaller area, the higher the accuracy of
the model and the better the prediction effect.
The accuracy rate is the proportion of correctly classified samples in the total samples,
it is easy to understand and explain, and in the case of a relatively balanced number
of positive and negative samples, accuracy can better reflect the overall performance of
the model. However, when the number of positive and negative samples is unbalanced,
the accuracy will be biased towards the majority class, which may lead to insufficient
recognition ability for the minority class. The accuracy rate cannot reflect the ability of the
model to recognize positive and negative samples alone, so it is necessary to evaluate the
performance of the model in combination with other indicators. The recall rate measures
how many truly positive samples the model is able to identify and is very sensitive to
unbalanced data sets. However, the recall rate may ignore the ability to identify negative
samples, and in the binary classification problem, different thresholds will lead to different
recall rates, so it is necessary to choose the appropriate threshold according to the specific
situation. The PR curve (precision rate–recall curve) is used to evaluate the performance
of the model in a specific category, especially a few categories [96]. It has recall as the
horizontal axis and precision as the vertical axis. The recall rate represents the proportion
of all positive samples that are correctly judged to be positive; the accuracy rate represents
the percentage of all samples judged to be positive that are actually positive. The PR
curve has higher sensitivity when dealing with unbalanced data sets, so it has a good
application prospect for accuracy evaluation in mineral prospectivity mapping. The ROC
curve is mainly used to evaluate the performance of classification models, especially in
binary classification problems, which can provide for the performance of models under
different thresholds. The Kappa coefficient is mainly used to evaluate the model accuracy
Minerals 2024, 14, 1021 28 of 35
At the same time, the confusion matrix, mean squared error series, and Brier score
are often applied to deep learning tasks. The confusion matrix is a specific table used to
visualize the performance of an algorithm, showing how the actual class compares to the
predicted class. The overall performance of the model can be assessed by size; the larger the
value on the diagonal, the better. But the confusion matrix may not work well for extremely
unbalanced data sets, as the performance of a few classes may be overshadowed by the
performance of the many classes. Therefore, it is necessary to use this method carefully
when making prospecting predictions. Mean squared error (MSE), Root mean squared error
(RMSE), and Mean absolute error (MAE) are used to measure the difference between the
predicted value and the actual value. The smaller the MSE and RMSE, the more accurate
the model prediction, while MAE directly reflects the average level of prediction error.
MSE is very sensitive to large deviations between predicted and actual values. Since the
errors are squared, larger errors will be penalized more severely, which helps the model
better capture and reduce these large errors. MSE, RMSE, and MAE are very sensitive to
small probability metallogenic events, but they are widely used in regression problems,
reflecting the overall error level of numerical prediction, so these evaluation methods can
be used carefully in future studies about mineral prospectivity mapping. The Brier score
is the mean square error of the difference between the forecast probability and the actual
result, and is often used to evaluate the degree of calibration and accuracy of probabilistic
models; the closer to 0 the prediction is, the better the prediction is. The Brier score is
sensitive to the change in predicted value near 0 or 1, which has a good application space
in deep learning prospecting prediction. Prediction accuracy evaluation methods are
varied. In practical application, in order to avoid various curves that may exist in a single
accuracy evaluation method, several different accuracy evaluation methods can be used to
comprehensively judge the reliability of the model. In the future, we also need to continue
to try to apply various other accuracy assessment methods to deep learning in mineral
prospectivity mapping.
Minerals 2024, 14, 1021 29 of 35
6. Conclusions
As a new data-driven model, deep learning algorithms can directly extract key features
from the original data by building multi-layer neural networks, which can capture nonlinear
relationships and demonstrate powerful data fitting capabilities. At present, they are
widely used in geological prospecting and prediction, and applied to a certain extent in
various ore deposits, with good results and optimistic prospects for future application.
Compared with traditional prospecting prediction, deep learning algorithms have obvious
advantages, and different types of deep learning algorithms, including the DAE, CNN,
RNN, GAN, etc., have enhanced the development of prospecting prediction methods to
varying degrees, and are of great significance to mineral development. Integrated methods
currently used in exploration projects, such as integrated geophysical inversion methods
and joint optimization methods, can also be integrated with deep learning in mineral
prospectivity mapping in the future.
Multivariate geological data, data enhancement, and deep learning algorithms are
three important components of deep learning-based mineral prospectivity mapping. How-
ever, at present, there are some challenges in mineral prospectivity mapping using deep
learning, mainly including the poor quality of geological basic data, the applicability of data
Minerals 2024, 14, 1021 30 of 35
enhancement methods, the interpretability of deep learning algorithms, and the selection
of prediction accuracy evaluation methods.
The application of deep learning in prospecting prediction is very effective, the
prospect is objective, and it can provide new ideas for future prospecting. The main re-
search directions of future deep learning in mineral prospectivity mapping are the category
extension and processing method of geological data, the further study of data enhancement
methods, and the comprehensive application of deep learning algorithms to give full play
to their respective advantages. Solving these problems will help deep learning make greater
progress in the field of mineral prospectivity mapping.
Author Contributions: Conceptualization, Y.C., G.G. and K.S.; formal analysis, K.S., Y.Z. and
Z.L.; writing—original draft preparation, K.S.; writing—review and editing, K.S., Y.C. and W.Z.;
supervision, J.G., Z.S., Z.L. and Z.Z. All authors have read and agreed to the published version of
the manuscript.
Funding: This research was funded by the Program of China Geological Survey (project numbers
DD20243184, DD20230591, DD20243187).
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. Zhang, S. Multi-Geoinformation Integration for Mineral Prospectivity Mapping in the Hezuo-Meiwu District, Gansu Province.
Ph.D. Thesis, China University of Geoscience, Beijing, China, 2022.
2. Zhao, P. Quantitative mineral prediction and deep mineral. Earth Sci. Front. 2007, 14, 001–010.
3. Xiao, K.; Ding, J.; Liu, R. The Discussion of Three-part Form of Non-fuel Mineral Resource Assessment. Geol. Rev. 2006, 52,
793–798. [CrossRef]
4. Yang, N.; Zhang, Z.; Yang, J.; Hong, Z. Mineral Prospectivity Prediction by Integration of Convolutional Autoencoder Network
and Random Forest. Nat. Resour. Res. 2022, 31, 1103–1119. [CrossRef]
5. Carranza, E.J.M.; Laborte, A.G. Data-driven predictive mapping of gold prospectivity, Baguio district, Philippines: Application of
Random Forests algorithm. Ore Geol. Rev. 2015, 71, 777–787. [CrossRef]
6. Carranza, E.J.M. Geocomputation of mineral exploration targets. Comput. Geosci. 2011, 37, 1907–1916. [CrossRef]
7. Porwal, A.; Carranza, E.J.M. Introduction to the Special Issue: GIS-based mineral potential modelling and geological data analyses
for mineral exploration. Ore Geol. Rev. 2015, 71, 477–483. [CrossRef]
8. Zhang, D. Spatially Weighted Technology for Logistic Regression and Its Application in Mineral Prospective Mapping.
Ph.D. Thesis, China University of Geoscience, Wuhan, China, 2015.
9. Yang, N.; Zhang, Z.; Yang, J.; Hong, Z. Applications of data augmentation in mineral prospectivity prediction based on
convolutional neural networks. Comput. Geosci. 2022, 161, 105075. [CrossRef]
10. Singer, D.A. Basic concepts in three-part quantitative assessments of undiscovered mineral resources. Nonrenew. Resour. 1993, 2,
69–81. [CrossRef]
11. Zhao, P.; Hu, J.; Li, Z. The theory and practices of statistical prediction for mineral deposits. Earth Sci.- J. Wuhan Coll. Geol. 1983, 4,
107–121.
12. Cheng, Q. Ideas and methods for mineral resources integrated prediction in covered areas. Earth Sci.- J. Wuhan Coll. Geol. 2012, 37,
1109–1125.
13. Agterberg, F.P.; Cheng, Q. Conditional Independence Test for Weights-of-Evidence Modeling. Nat. Resour. Res. 2002, 11, 249–255.
[CrossRef]
14. Cheng, Q. BoostWofE: A New Sequential Weights of Evidence Model Reducing the Effect of Conditional Dependency. Math.
Geosci. 2015, 47, 591–621. [CrossRef]
15. Chen, S. Research of Multiple Geoscience Information Prospecting Prediction in Xikuangshan Antimony Ore Field. Ph.D. Thesis,
China University of Geoscience, Beijing, China, 2012.
16. Li, X.; Yuan, F.; Zhang, M.; Jia, C.; Jowitt, S.; Ord, A.; Zheng, T.; Hu, X.; Li, Y. Three-dimensional mineral prospectivity modeling
for targeting of concealed mineralization within the Zhonggu iron orefield, Ningwu Basin, China. Ore Geol. Rev. 2015, 71, 633–654.
[CrossRef]
17. Porwal, A.; González-Álvarez, I.; Markwitz, V.; McCuaig, T.C.; Mamuse, A. Weights-of-evidence and logistic regression modeling
of magmatic nickel sulfide prospectivity in the Yilgarn Craton, Western Australia. Ore Geol. Rev. 2010, 38, 184–196. [CrossRef]
18. Zuo, R.; Carranza, E.J.M. Support vector machine: A tool for mapping mineral prospectivity. Comput. Geosci. 2011, 37, 1967–1975.
[CrossRef]
19. Shabankareh, M.; Hezarkhani, A. Application of support vector machines for copper potential mapping in Kerman region, Iran.
J. Afr. Earth Sci. 2017, 128, 116–126. [CrossRef]
Minerals 2024, 14, 1021 31 of 35
20. Chen, Y.; Wu, W. Application of one-class support vector machine to quickly identify multivariate anomalies from geochemical
exploration data. Geochem. Explor. Environ. Anal. 2017, 17, 231–238. [CrossRef]
21. Chen, Y.; Wu, W. Mapping mineral prospectivity by using one-class support vector machine to identify multivariate geological
anomalies from digital geological survey data. Aust. J. Earth Sci. 2017, 64, 639–651. [CrossRef]
22. Sun, T.; Chen, F.; Zhong, L.; Liu, W.; Wang, Y. GIS-based mineral prospectivity mapping using machine learning methods: A case
study from Tongling ore district, eastern China. Ore Geol. Rev. 2019, 109, 26–49. [CrossRef]
23. Carranza, E.J.M.; Laborte, A.G. Random forest predictive modeling of mineral prospectivity with small number of prospects and
data with missing values in Abra (Philippines). Comput. Geosci. 2015, 74, 60–70. [CrossRef]
24. Wang, Z.; Zuo, R.; Dong, Y. Mapping geochemical anomalies through integrating random forest and metric learning methods.
Nat. Resour. Res. 2019, 28, 1285–1298. [CrossRef]
25. Gao, Y.; Zhang, Z.; Xiong, Y.; Zuo, R. Mapping mineral prospectivity for Cu polymetallic mineralization in southwest Fujian
Province, China. Ore Geol. Rev. 2016, 75, 16–28. [CrossRef]
26. Zhang, F.M.A. Research on Deep Learning Extraction Method in Open Mining Area Based on Multi-Source Remote Sensing Images; Anhui
University: Hefei, China, 2020.
27. Fu, Y.; Cheng, Q.; Jing, L.; Ye, B.; Fu, H. Mineral Prospectivity Mapping of Porphyry Copper Deposits Based on Remote Sensing
Imagery and Geochemical Data in the Duolong Ore District, Tibet. Remote Sens. 2023, 15, 439. [CrossRef]
28. Feng, J.; Zhang, Q.; Luo, J. Deeply mining the intrinsic value of geodata to improve the accuracy of predicting by quantitatively
optimizing method for prospecting target areas. Earth Sci. Front. 2022, 29, 403–411. [CrossRef]
29. Zhang, C.; Zuo, R. Recognition of multivariate geochemical anomalies associated with mineralization using an improved
generative adversarial network. Ore Geol. Rev. 2021, 136, 104264. [CrossRef]
30. Xiong, Y.; Zuo, R. Recognition of geochemical anomalies using a deep autoencoder network. Comput. Geosci. 2016, 86, 75–82.
[CrossRef]
31. Zuo, R.; Xiong, Y. Big Data Analytics of Identifying Geochemical Anomalies Supported by Machine Learning Methods. Nat.
Resour. Res. 2018, 27, 5–13. [CrossRef]
32. Shi, L.; Jianping, C.; Jie, X. Prospecting Information Extraction by Text Mining Based on Convolutional Neural Networks–A Case
Study of the Lala Copper Deposit, China. IEEE Access 2018, 6, 52286–52297. [CrossRef]
33. Liu, Y.; Zhu, L.; Zhou, Y. Application of Convolutional Neural Network in prospecting prediction of ore deposits: Taking the
Zhaojikou Pb-Zn ore deposit in Anhui Province as a case. Acta Petrol. Sin. 2018, 34, 3217–3224.
34. Liu, Y. Experimental research on big data mining and intelligent prediction of prospecting target area—Application of convolu-
tional neural network model. Geotecton. Metallog. 2020, 44, 192–202.
35. Zheng, X.; Zhang, M.; Ren, W. Application of convolution neural networks in gold exploration and prediction in Shandong
Province. Geophys. Geochem. Explor. 2023, 47, 1433–1440.
36. Li, Q.; Chen, G.; Luo, L. Mineral prospectivity mapping using attention-based convolutional neural network. Ore Geol. Rev. 2023,
156, 105381. [CrossRef]
37. Li, Y.-S.; Chai, S.-L. Soil geochemical prospecting prediction method based on deep convolutional neural networks-Taking Daqiao
Gold Deposit in Gansu Province, China as an example. China Geol. 2022, 5, 71–83.
38. Du, Y.; Cao, Y.; Huo, D.; Li, D.; Gao, Z. Petrology and geochemistry of Silurian-Triassic sedimentary rocks in the Tongling area;
Constraints on the genesis of stratabound skarn deposits. Earth Sci. Front. 2014, 21, 228–239.
39. Ling, Q.; LIU, C. REE behavior during formation of st ra ta-bound skarn and related deposit: A case study of Dongguashan ska
rn deposit in Anh ui province, China. Acta Petrol. Sin. 2003, 19, 192–200.
40. Deng, Y.; Song, X.; Jie, W.; Yuan, F.; Zhao, Z.; Wei, S.; Zhu, J.; Kang, J.; Wang, K.; Liang, Q.; et al. Determination of sedimentary
ages of strata in the Huangshan-Jingerquan mineralization belt and its geological significance. Acta Geol. Sin. 2021, 95, 362–376.
[CrossRef]
41. Xue, S.; Wang, Q.; Tang, D.; Mao, Y.; Yao, Z. Contamination mechanism of magmatic Ni-Cu sulfide deposits in orogenic belts:
Examples from Permian Ni-Cu deposits in Tianshan-Beishan. Miner. Depos. 2022, 41, 1–20.
42. Grobler, D.F.; Brits, J.A.N.; Maier, W.D.; Crossingham, A. Litho- and chemostratigraphy of the Flatreef PGE deposit, northern
Bushveld Complex. Miner. Depos. 2019, 54, 3–28. [CrossRef]
43. Li, L.-J.; Li, D.-X.; Mao, X.-C.; Liu, Z.-K.; Lai, J.-Q.; Su, Z.; Ai, Q.-X.; Wang, Y.-Q. Evolution of magmatic sulfide of the giant
Jinchuan Ni-Cu deposit, NW China: Insights from chalcophile elements in base metal sulfide minerals. Ore Geol. Rev. 2023,
158, 105497. [CrossRef]
44. Mukasa, S.B.; Vidal, C.C.E.; Injoque-Espinoza, J. Lead isotope-bearing on the metallogenesis of sulfide ore deposits in central and
southern Peru. Econ. Geol. 1990, 85, 1438–1446. [CrossRef]
45. Elongo, V.; Lecumberri-Sanchez, P.; Legros, H.; Falck, H.; Adlakha, E.E.; Roy-Garand, A. Paragenetic constraints on the Cantung,
Mactung and Lened tungsten skarn deposits, Canada: Implications for grade distribution. Ore Geol. Rev. 2020, 125, 103677.
[CrossRef]
46. Ord, A.; Hobbs, B.E.; Lester, D.R. The mechanics of hydrothermal systems: I. Ore systems as chemical reactors. Ore Geol. Rev.
2012, 49, 1–44. [CrossRef]
Minerals 2024, 14, 1021 32 of 35
47. Yang, L.; Yang, W.; Zhang, L.; Gao, X.; Shen, S.; Wang, S.; Xu, H.; Jia, X.; Deng, J. Developing structural control models for
hydrothermal metallogenic systems: Theoretical and methodological principles and applications. Earth Sci. Front. 2024, 31,
239–266.
48. Liu, Z. Ore-controlling structure of Withnell gold deposit, Pilbara Craton, Australia. J. Geol. 2021, 45, 154–160. [CrossRef]
49. Song, M.; Wang, B.; Song, Y.; Li, J.; Zheng, J.; Li, S.; Fan, J.; Yang, Z.; He, C.; Gao, M.; et al. Spatial coupling relationship between
faults and gold deposits in the Jiaodong ore concentration area and the effect of thermal doming-extension on mineralisation. Ore
Geol. Rev. 2023, 153, 105277. [CrossRef]
50. Lv, Q.; Dong, S.; Tang, J.; Shi, D.; Chang, Y. Multi-scale and integrated geophysical data revealing mineral systems and exploring
for mineral deposits at depth: A synthesis from SinoProbe-03. Chin. J. Geophys. 2015, 58, 4319–4343. [CrossRef]
51. Zhang, S.; Wang, Q.; Luo, Y. An overview on the development of the electrical prospecting method in China. Acta Geophys. Sin.
1994, 37, 408–424.
52. Wu, Y.; Liu, B.; Gao, Y.; Li, C.; Tang, R.; Kong, Y.; Xie, M.; Li, K.; Dan, S.; Qi, K.; et al. Mineral prospecting mapping with
conditional generative adversarial network augmented data. Ore Geol. Rev. 2023, 163, 105787. [CrossRef]
53. Shebl, A.; Abdellatif, M.; Elkhateeb, S.O.; Csámer, Á. Multisource Data Analysis for Gold Potentiality Mapping of Atalla Area
and Its Environs, Central Eastern Desert, Egypt. Minerals 2021, 11, 641. [CrossRef]
54. Shebl, A.; Abdellatif, M.; Hissen, M.; Ibrahim Abdelaziz, M.; Csámer, Á. Lithological mapping enhancement by integrating
Sentinel 2 and gamma-ray data utilizing support vector machine: A case study from Egypt. Int. J. Appl. Earth Obs. Geoinf. 2021,
105, 102619. [CrossRef]
55. Spicer, B. Geophysical signature of the Victoria property, vectoring toward deep mineralization in the Sudbury Basin. Interpret. A
J. Subsurf. Charact. 2016, 4, T281–T290. [CrossRef]
56. Wang, R.; Hao, X.; Hu, L.; Chen, H.; Liu, H.; Chen, F.; Yu, L.; Liu, W.; Fang, L.; Kang, Y. Discovery of skarn iron-rich deposit based
on gravity and magnetic data in the Qihe-Yucheng, Shandong Province: Enlightenment to prospecting of the superdeep coverage
area. Geol. China 2023, 50, 331–346.
57. Wang, S. The new development of theory and method of synthetic information mineral resources prognosis. Geol. Bull. China
2010, 29, 1399–1403.
58. Zuo, R. Exploring the effects of cell size in geochemical mapping. J. Geochem. Explor. 2012, 112, 357–367. [CrossRef]
59. Luo, X.; Yang, X. Study and prospecting prediction of hidden deposits by geoelectrochemical survey. Geol. Explor. 1989, 54, 43–51.
60. Liu, Y.; Luo, X.; Liu, P.; Zheng, C.; Liu, G.; Song, B.; Song, G. Application of geo-electrochemical integration technology to search
for concealed Pb-Zn ore in the Geluqiduishan mining area and its periphery, Inner Mongolia. Geol. Explor. 2018, 54, 1001–1012.
[CrossRef]
61. Yue, D.; Liu, P.; Liao, X.; Chen, J.; Li, J.; Yang, Q.; Yang, X. Research and application of AMT and geo-electrochemical measurements
in deep prospecting of the Murong lithium deposit in Yajiang, western Sichuan Province. Geol. Explor. 2023, 59, 760–773.
62. Yang, Q.; Luo, X.; Yue, D.; Liu, P.; Gao, W.; Wen, M.; Liao, X.; Li, J.; Liang, M.; Liu, Y. Intelligent prospecting method based on
probabilistic neural network: Taking the Murong lithium deposit in Yajiang County of Sichuan Province as an example. Geol.
Explor. 2023, 59, 985–999.
63. Henry, C.D.; John, D.A.; Leonardson, R.W.; McIntosh, W.C.; Heizler, M.T.; Colgan, J.P.; Watts, K.E. Timing of Rhyolite Intrusion
and Carlin-Type Gold Mineralization at the Cortez Hills Carlin-Type Deposit, Nevada, USA. Econ. Geol. 2023, 118, 57–91.
[CrossRef]
64. Yang, B.; Li, C.; Bo, H.; Hou, X.; Su, P.; Fan, S. Geochemical characteristics and genesis of Changhangou crystalline graphite
deposit in Hadamengou area, Inner Mongolia. Miner. Depos. 2023, 42, 444–462. [CrossRef]
65. Min, X.; Pengbo, Q.; Fengwei, Z. Research and application of logging lithology identification for igneous reservoirs based on
deep learning. J. Appl. Geophys. 2020, 173, 103929. [CrossRef]
66. Liu, J.-J.; Liu, J.-C. Integrating deep learning and logging data analytics for lithofacies classification and 3D modeling of tight
sandstone reservoirs. Geosci. Front. 2022, 13, 101311. [CrossRef]
67. Shirmard, H.; Farahbakhsh, E.; Beiranvand Pour, A.; Muslim, A.M.; Müller, R.D.; Chandra, R. Integration of Selective Dimension-
ality Reduction Techniques for Mineral Exploration Using ASTER Satellite Data. Remote Sens. 2020, 12, 1261. [CrossRef]
68. Zidan, U.; Desouky, H.A.E.; Gaber, M.M.; Abdelsamea, M.M. From Pixels to Deposits: Porphyry Mineralization With Multispectral
Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 9474–9486. [CrossRef]
69. Hunt, G.R.; Ashley, R.P. Spectra of altered rocks in the visible and near infrared. Econ. Geol. 1979, 74, 1613–1629. [CrossRef]
70. Shirmard, H.; Farahbakhsh, E.; Müller, R.D.; Chandra, R. A review of machine learning in processing remote sensing data for
mineral exploration. Remote Sens. Environ. 2022, 268, 112750. [CrossRef]
71. Shebl, A.; Abdellatif, M.; Badawi, M.; Dawoud, M.; Fahil, A.S.; Csámer, Á. Towards better delineation of hydrothermal alterations
via multi-sensor remote sensing and airborne geophysical data. Sci. Rep. 2023, 13, 7406. [CrossRef]
72. Haynes, D. The Olympic Dam ore deposit discovery—A personal view. SEG Newsl. 2006, 66, 1–15. [CrossRef]
73. Guo, J.; Zhu, G.; Zou, L.; Wang, R.; Han, Y.; Wang, W.; Xiang, A. Remote sensing geological survey of bauxite deposits in
Dazhuyuan-Longxing area of north Guizhou. Miner. Resour. Geol. 2016, 30, 117–121.
74. Xiao, L. Lithologic and mineral information extraction for bauxite deposits exploration using ASTER data in the Wuchuan-
Zheng’an-Daozhen area, northern Guizhou province, China. J. Mines Met. Fuels 2018, 66, 280–286.
Minerals 2024, 14, 1021 33 of 35
75. van Gerven, M.; Bohte, S. Editorial: Artificial Neural Networks as Models of Neural Information Processing. Front. Comput.
Neurosci. 2017, 11, 114. [CrossRef] [PubMed]
76. Jiao, L.; Yang, S.; Liu, F.; Wang, S.; Feng, Z. Seventy years beyond neural networks: Retrospect and prospect. Chin. J. Comput.
2016, 39, 1697–1716.
77. Majumdar, A. Graph structured autoencoder. Neural Netw. 2018, 106, 271–280. [CrossRef]
78. Hinton, G.E.; Osindero, S.; Teh, Y.-W. A Fast Learning Algorithm for Deep Belief Nets. Neural Comput. 2006, 18, 1527–1554.
[CrossRef]
79. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Dep. Comput. Sci. Univ. Tor. 2006,
313, 504–507. [CrossRef] [PubMed]
80. Xie, M.; Liu, B.; Wang, L.; Li, C.; Kong, Y.; Tang, R. Auto encoder generative adversarial networks—Based mineral prospectivity
mapping in Lhasa area, Tibet. J. Geochem. Explor. 2023, 255, 107326. [CrossRef]
81. Zhang, S.; Xiao, K.; Carranza, E.J.M.; Yang, F.; Zhao, Z. Integration of auto-encoder network with density-based spatial clustering
for geochemical anomaly detection for mineral exploration. Comput. Geosci. 2019, 130, 43–56. [CrossRef]
82. Vincent Dumoulin, I.B.; Poole, B.; Lamb, A.; Arjovsky, M.; Mastropietro, O.; Courville, A. Adversarially Learned Inference. arXiv
2016, arXiv:1606.00704.
83. Chen, Y.; Wang, Y.; Kirschen, D.; Zhang, B. Model-Free Renewable Scenario Generation Using Generative Adversarial Networks.
IEEE Trans. Power Syst. 2018, 33, 3265–3275. [CrossRef]
84. Farahbakhsh, E.; Maughan, J.; Müller, R.D. Prospectivity modelling of critical mineral deposits using a generative adversarial
network with oversampling and positive-unlabelled bagging. Ore Geol. Rev. 2023, 162, 105665. [CrossRef]
85. Anders Boesen Lindbo, L.; Søren Kaae, S.; Larochelle, H.; Winther, O. Autoencoding beyond pixels using a learned similarity
metric. In Proceedings of the International Conference on Machine Learning, New York City, New York, USA, 19–24 June 2016;
pp. 1558–1566.
86. Liu, W.; Liang, X.; Qu, H. Learning performance of convolutional neural networks with different pooling models. J. Image Graph.
2016, 21, 1178–1190. [CrossRef]
87. Yang, N.; Zhang, Z.; Yang, J.; Hong, Z.; Shi, J. A Convolutional Neural Network of GoogLeNet Applied in Mineral Prospectivity
Prediction Based on Multi-source Geoinformation. Nat. Resour. Res. 2021, 30, 3905–3923. [CrossRef]
88. Li, Z.; Xue, L.; Ran, X.; Li, Y.; Dong, G.; Li, Y.; Dai, J. Intelligent prospect prediction method based on convolutional neural
network: A case study of copper deposits in Longshoushan Area, Gansu Province. J. Jilin Univ. (Earth Sci. Ed.) 2022, 52, 418–433.
89. Sankar, M.; Batri, K.; Parvathi, R. Earliest diabetic retinopathy classification using deep convolution neural networks. Int. J. Adv.
Eng. Technol. 2016, 10, M9.
90. Li, J.; Yuan, Z.; Li, Z.; Ren, A.; Ding, C.; Draper, J.; Nazarian, S.; Qiu, Q.; Yuan, B.; Wang, Y. Normalization and dropout for
stochastic computing-based deep convolutional neural networks. Integration 2019, 65, 395–403. [CrossRef]
91. Wu, H.; Zhao, J. Deep convolutional neural network model based chemical process fault diagnosis. Comput. Chem. Eng. 2018, 115,
185–197. [CrossRef]
92. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv
2014, arXiv:1412.3555.
93. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw.
1994, 5, 157–166. [CrossRef]
94. Van Houdt, G.; Mosquera, C.; Nápoles, G. A review on the long short-term memory model. Artif. Intell. Rev. 2020, 53, 5929–5955.
[CrossRef]
95. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder
approaches. arXiv 2014, arXiv:1409.1259.
96. Wang, Z.; Zuo, R. Mineral prospectivity mapping using a joint singularity-based weighting method and long short-term memory
network. Comput. Geosci. 2022, 158, 104974. [CrossRef]
97. Yin, B.; Zuo, R.; Xiong, Y. Mineral Prospectivity Mapping via Gated Recurrent Unit Model. Nat. Resour. Res. 2022, 31, 2065–2079.
[CrossRef]
98. Huijie Zhao, K.D.; Li, N.; Wang, Z.; Wei, W. Hierarchical Spatial-Spectral Feature Extraction with Long Short Term Memory
(LSTM) for Mineral Identification Using Hyperspectral Imagery. Sensors 2020, 20, 6854. [CrossRef]
99. Wang, Z.; Li, T.; Zuo, R. Leucogranite mapping via convolutional recurrent neural networks and geochemical survey data in the
Himalayan orogen. Geosci. Front. 2024, 15, 181–192. [CrossRef]
100. Xu, Y.; Li, Z.; Xie, Z.; Cai, H.; Niu, P.; Liu, H. Mineral prospectivity mapping by deep learning method in Yawan-Daqiao area,
Gansu. Ore Geol. Rev. 2021, 138, 104316. [CrossRef]
101. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017,
60, 84–90. [CrossRef]
102. Li, S.; Chen, J.; Xiang, J. Applications of deep convolutional neural networks in prospecting prediction based on two-dimensional
geological big data. Neural Comput. Appl. 2020, 32, 2037–2053. [CrossRef]
103. Li, S.; Chen, J.; Xiang, J.; Zhang, Z.; Zhang, Y. Two-dimensional prospecting prediction based on AlexNet network: A case study
of sedimentary Mn deposits in Songtao-Huayuan area. Geol. Bull. China 2019, 38, 2022–2032.
104. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556.
Minerals 2024, 14, 1021 34 of 35
105. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with
convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA,
USA, 7–12 June 2015; Volume 7, pp. 1–9. [CrossRef]
106. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
107. Gao, L.; Huang, Y.; Zhang, X.; Liu, Q.; Chen, Z. Prediction of Prospecting Target Based on ResNet Convolutional Neural Network.
Appl. Sci. 2022, 12, 11433. [CrossRef]
108. Boiger, R.; Churakov, S.V.; Ballester Llagaria, I.; Kosakowski, G.; Wüst, R.; Prasianakis, N.I. Direct mineral content prediction
from drill core images via transfer learning. Swiss J. Geosci. 2024, 117, 8. [CrossRef]
109. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of
the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich,
Germany, 5–9 October 2015; Proceedings, Part III 18; pp. 234–241.
110. Liu, C.; Wang, W.; Tang, J.; Wang, Q.; Zheng, K.; Sun, Y.; Zhang, J.; Gan, F.; Cao, B. A deep-learning-based mineral prospectivity
modeling framework and workflow in prediction of porphyry–epithermal mineralization in the Duolong ore District, Tibet.
Ore Geol. Rev. 2023, 157, 105419. [CrossRef]
111. Mou, N.; Carranza, E.J.; Wang, G.; Sun, X. A Framework for Data-Driven Mineral Prospectivity Mapping with Interpretable
Machine Learning and Modulated Predictive Modeling. Nat. Resour. Res. 2023, 32, 2439–2462. [CrossRef]
112. Wang, D.; Huang, F.; Wang, Y.; He, H.; Li, X.; Liu, X.; Sheng, J.; Liang, T. Regional metallogeny of Tungsten-tin-polymetallic
deposits in Nanling region, South China. Ore Geol. Rev. 2020, 120, 103305. [CrossRef]
113. Jordão, H.; Azevedo, L.; Sousa, A.J.; Soares, A. Generative adversarial network applied to ore type modeling in complex geological
environments. Math. Geosci. 2022, 54, 1165–1182. [CrossRef]
114. Li, T.; Zuo, R.; Zhao, X.; Zhao, K. Mapping prospectivity for regolith-hosted REE deposits via convolutional neural network with
generative adversarial network augmented data. Ore Geol. Rev. 2022, 142, 104693. [CrossRef]
115. Guo, M.; Chen, Y. A SMOTified-GAN-augmented bagging ensemble model of extreme learning machines for detecting geochemi-
cal anomalies associated with mineralization. Geochemistry 2024, 126156. [CrossRef]
116. Zhang, S.; Carranza, E.J.M.; Wei, H.; Xiao, K.; Yang, F.; Xiang, J.; Zhang, S.; Xu, Y. Data-driven mineral prospectivity mapping by
joint application of unsupervised convolutional auto-encoder network and supervised convolutional neural network. Nat. Resour.
Res. 2021, 30, 1011–1031. [CrossRef]
117. Li, T.; Zuo, R.; Xiong, Y.; Peng, Y. Random-Drop Data Augmentation of Deep Convolutional Neural Network for Mineral
Prospectivity Mapping. Nat. Resour. Res. 2020, 30, 27–38. [CrossRef]
118. Li, S.; Chen, J.; Liu, C.; Wang, Y. Mineral Prospectivity Prediction via Convolutional Neural Networks Based on Geological Big
Data. J. Earth Sci. 2021, 32, 327–347. [CrossRef]
119. Chen, L.; Guan, Q.; Feng, B.; Yue, H.; Wang, J.; Zhang, F. A multi-convolutional autoencoder approach to multivariate geochemical
anomaly recognition. Minerals 2019, 9, 270. [CrossRef]
120. Luo, Z.; Zuo, R.; Xiong, Y.; Wang, X. Detection of geochemical anomalies related to mineralization using the GANomaly network.
Appl. Geochem. 2021, 131, 105043. [CrossRef]
121. Xiong, Y.; Zuo, R.; Carranza, E.J.M. Mapping mineral prospectivity through big data analytics and a deep learning algorithm.
Ore Geol. Rev. 2018, 102, 811–817. [CrossRef]
122. Xiong, Y.; Zuo, R. GIS-based rare events logistic regression for mineral prospectivity mapping. Comput. Geosci. 2018, 111, 18–25.
[CrossRef]
123. Atalay, F. Estimation of Fe Grade at an Ore Deposit Using Extreme Gradient Boosting Trees (XGBoost). Min. Metall. Explor. 2024,
41, 2119–2128. [CrossRef]
124. Filzmoser, P.; Garrett, R.G.; Reimann, C. Multivariate outlier detection in exploration geochemistry. Comput. Geosci. 2005, 31,
579–587. [CrossRef]
125. Hronsky, J.M.A.; Kreuzer, O.P. Applying spatial prospectivity mapping to exploration targeting: Fundamental practical issues
and suggested solutions for the future. Ore Geol. Rev. 2019, 107, 647–653. [CrossRef]
126. Zhang, C.; Zuo, R.; Xiong, Y. Detection of the multivariate geochemical anomalies associated with mineralization using a deep
convolutional neural network and a pixel-pair feature method. Appl. Geochem. 2021, 130, 104994. [CrossRef]
127. Brandmeier, M.; Cabrera Zamora, I.G.; Nykänen, V.; Middleton, M. Boosting for Mineral Prospectivity Modeling: A New GIS
Toolbox. Nat. Resour. Res. 2019, 29, 71–88. [CrossRef]
128. Chen, L.; Guan, Q.; Xiong, Y.; Liang, J.; Wang, Y.; Xu, Y. A Spatially Constrained Multi-Autoencoder approach for multivariate
geochemical anomaly recognition. Comput. Geosci. 2019, 125, 43–54. [CrossRef]
129. Li, S.; Liu, C.; Chen, J. Mineral Prospecting Prediction via Transfer Learning Based on Geological Big Data: A Case Study of
Huayuan, Hunan, China. Minerals 2023, 13, 504. [CrossRef]
130. Mantilla-Dulcey, A.; Goyes-Peñafiel, P.; Baez-Rodríguez, R.; Khurama, S. Porphyry-type mineral prospectivity mapping with
imbalanced data via prior geological transfer learning. Gondwana Res. 2024, 136, 236–250. [CrossRef]
131. Wu, B.; Li, X.; Yuan, F.; Li, H.; Zhang, M. Transfer learning and siamese neural network based identification of geochemical
anomalies for mineral exploration: A case study from the CuAu deposit in the NW Junggar area of northern Xinjiang Province,
China. J. Geochem. Explor. 2022, 232, 106904. [CrossRef]
Minerals 2024, 14, 1021 35 of 35
132. Ritter, S.; Barrett, D.G.T.; Santoro, A.; Botvinick, M.M. Cognitive psychology for deep neural networks: A shape bias case study.
Int. Conf. Mach. Learn. 2017, 2940–2949.
133. Yousefi, M.; Carranza, E. Prediction–area (P–A) plot and C–A fractal analysis to classify and evaluate evidential maps for mineral
prospectivity modeling. Comput. Geosci. 2015, 79, 69–81. [CrossRef]
134. Roshanravan, B.; Aghajani, H.; Yousefi, M.; Kreuzer, O. An Improved Prediction-Area Plot for Prospectivity Analysis of Mineral
Deposits. Nat. Resour. Res. 2019, 28, 1089–1105. [CrossRef]
135. Rodriguez-Galiano, V.; Sanchez-Castillo, M.; Chica-Olmo, M.; Chica-Rivas, M. Machine learning predictive models for mineral
prospectivity: An evaluation of neural networks, random forest, regression trees and support vector machines. Ore Geol. Rev.
2015, 71, 804–818. [CrossRef]
136. Nathwani, C.L.; Wilkinson, J.J.; Brownscombe, W.; John, C.M. Mineral Texture Classification Using Deep Convolutional Neural
Networks: An Application to Zircons From Porphyry Copper Deposits. J. Geophys. Res. Solid Earth 2023, 128, e2022JB025933.
[CrossRef]
137. Zuo, R.; Cheng, Q.; Agterberg, F.P. Application of a hybrid method combining multilevel fuzzy comprehensive evaluation with
asymmetric fuzzy relation analysis to mapping prospectivity. Ore Geol. Rev. 2009, 35, 101–108. [CrossRef]
138. Zuo, R.; Zhang, Z.; Zhang, D.; Carranza, E.J.M.; Wang, H. Evaluation of uncertainty in mineral prospectivity mapping due to
missing evidence: A case study with skarn-type Fe deposits in Southwestern Fujian Province, China. Ore Geol. Rev. 2015, 71,
502–515. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.