Soumare eal BloDota Mining
(2021) 1430
hps/dorg/10.1186/s13040-021-00256-7 BioData Mining
New neural network classification ®
method for individuals ancestry prediction
aa
from SNPs data
H, Soumare!?" ®, S. Rezgui’, N. Gmati* and A. Benkahla?
atts makina
BMC
Abstract
Atificial Neural Network (ANN) algorithms have been widely used ta analyse genomic
data, Single Nucleotide Polymorphisms(SNPs) represent the genetic vatiations, the
most common in the human genome, it has been shown that they are involved in
many genetic diseases, and can be used ta predict their development, Developing
AN to handle this type af data can be considered as a great success in the medical
‘world, However, the high dimensionality of genomic data and the availabilty of a
limited number of samples can make the learning task very complicated. In this work,
we propose 2 New Neural Network classification method based on input perturbation
The idea is frst to use SVD to reduce the dimensionality ofthe input data and to train a
Classification network, which prediction errors are then reduced by perturbing the SYD
projection matrix. The proposed methad has been evaluated on data fram individuals,
‘with different ancestral origins, the experimental results have shown the effectiveness,
of the proposed method. Achieving up to 96.23% of classification accuracy, this
‘approach surpasses previous Deep learning approaches evaluated on the same dataset.
Keywords: Artificial neural network, Dimensionality reduction, Input perturbation,
Single nucleotide polymorphism, Singular value decomposition
Introduction
‘The human genome contains three billion of base pairs, with only 0.1% difference between
individuals [1]. The most common type of genetic variations between individuals is called
Single Nucleotide Polymorphism (SNP) [2]. An SNP is a change from one base pair
to another, which occurs about once every 1000 bases. Most of these SNPs have no
impact on human health. However, many studies have shown that some of these genetic
variations have important biological effects and are involved in many human diseases
[3,4], SNPs are commonly used to detect genes associated with the development of adis-
case within families [5]. In addition, SNPs can also help to predict a persons response
to drugs or their susceptibility to develop one or more particular diseases, In genetics,
othe Autor 2021 Open Aces atc eeses una a Cate Comnans Attn
‘shih parm ie hata adaplaton debian andreeraton n ay medumcr format ng aso ove nro
{heat thea aha ink tothe Gestne Commons ence angie changer wee
"erat pea soy nparare eyo redio tan pemssen recy
Camnans Pubic Doman Dedkaten waver fio /ceatneconmonscva/pubsorah/aea// apps oe aSoumare etal BiaDataMining (2021) 1430 Page 20f14
Genome-Wide Association Studies (GWAS) are observational studies that use high-
throughput genotyping technologies to identify a set of genetic variants that are associ-
ated to a given trait or disease [6], by comparing variants in a group of cases with variants
in a group of controls. However, this approach is only optimal for populations from the
same ancestry group, as it is challenging to
disease from those that characterize the genetic of human populations. In this context,
‘numerous machine learning algorithms have been used to classify individuals accord-
ing to genetic differences that affect their population. Support Vector Machines (SVM)
‘ciate the variations associated with a
methods have been applied to infer recent genetic ancestry of a subgroup of communities
in the USA [7] or coarse ethnicity [8]. However, SVM methods are very sensitive to the
choice of kernel and its parameters [9]. Deep learning algorithms, such as Neural Net-
works have been widely used to analyse genomic data as well as gene expression data to
classify certain diseases (10~20]. But, the high dimensionality of genomic data (when the
number of input features is several times higher than the number of training examples)
makes the learning task very difficult. Indeed, when data is composed of a large number
of input features nr for a small number of samples 7 (n << m), the problem of overft-
ting becomes inevitable. In general, overfitting in machine learning occurs when a model
fits well with the training data, but not ft the unseen data. The model learns details and
noise in the training data, which negatively impact the performance of the model on new
data. One way to avoid the problem of overfitting is to reduce the complexity ofthe prob-
lem by removing features that do not contribute or decrease the accuracy of the model
[21. Different techniques are used to deal with the problem of overftting. The most well-
known onesare Land £2 regularizations [22]. The idea of these techniques is to penalize
the higher weights in the model by adding extra terms in the loss function. Another com
monly used regularization technique, called "Dropout’, introduced by Hinton et al. [23]
consists of dropping neurons at random (in hidden layers) in each learning round. How-
ever, with such difference between the number of features versus the number of samples,
it increases the problem of overfitting. To overcome this problem, dimensionality reduc-
tion techniques need to be combined with unsupervised learning methods or other data
preprocessing techniques
‘There are many ways to transform a high-dimensional data to low-dimensional
data, Singular Value Decomposition (SVD), Principal Component Analysis (PCA)) and
Autoencoder(AE) are the most common dimensional reduction techniques. SVD and
PGA are the most popular linear dimensionality reduction techniques. Both attempt to
find k orthogonal dimensions in an n-dimensional space, so that k < nt. They are related to
each other, but PCA uses the covariance matrix of the input data, while SVD is performed
on the input matrix itself. The Autoencoder is a Neural Network that tries to reconstruct
the input data from their compressed form. Indeed, the Autoencoder is used as a method
of non-linear dimensionality reduction, it works by mapping an n-dimensional input data
into a k-dimensional data (with k
)
with rank(X) = rand decomposes it into a product of three matrices [54]; two orthogonal
matrices U € R"™" and V € R™*" anda matrix E = diag(o1.02..+.40n) € R04 > 0
for 1 r+ 1, such that
Xx? usvt = Duss.
The first r columns of the orthogonal matrices Lf and V are, respectively, the right
and the let eigenvectors associated with the r nonzero eigenvalues of XTX. Uj, Vi and
B, are, respectively, the ith column of U, V and E. The diagonal elements of © are the
nonnegative square roots of the 1 eigenvalues of XTX,
The dimension of the input matrix X is then reduced by projecting it onto a space
spanned by {l/1,LL,.... Url, the top k (kr) singular veetors of X. Given a set of sam
ples vi,22,....2 of dimension m, the projection matrix LU whose columns are formed
by the & first singular vectors of X must minimize
w v
Ded — sil = Oss — 85
xu — x),
where Pis the projection defined by
Pi RM —> RE
xv aa
‘The input data in reduced dimension is denoted by X! = X U*Soumare etal BiaDataMining (2021) 1430 Page 6 of 4
Description of the model
Let's considera L hidden layers ofa Multi-Layer Perceptron(MLP),in which » input train-
ing samples X = (xi,x2,...,%e) are labeled, ie, for each input x), the corresponding
output by the model is known and denoted y, or (xi). ¥ is a vector that contains all the
labels, A MEP can be described as follows
(2). a
f= Sia? if aD ot, @)
7
‘ia
where Bf and al (a = ay, for an input x-= (xy 2 ...2)") are the fh hidden unit,
bias term and activation function of layer f respectively. wis the weight that links the th
unt of the (J— 1)th layer to the jth unit ofthe /th layer. w{ and al/— are, respectively, the
incoming weight vector to the jth neuron of layer / and the output vector of (L-1)th layer,
(is any activation function, Learning the model consists in finding all the parameters w)
and bj so that the output a! from the model approximates the true output vector ¥(x),
for all training inputs x.. For simplification, we consider that there are no bias terms Bf
or simply we consider it as an additional component of w/ and denote by W! the matrix
whose columns are the vectors w) (Fig. 2).
Due to the high dimension of the input data, the proposed approach consists to first
project the original data onto a lower dimensional space using SVD. Once the dimen-
sion of the input data is reduced, a multilayer perceptron (MLP) classification network is
constructed in three steps:
Stop 1 Learning the weight matrix W’: First, a classification network(see Fig, 3) is
trained to find W*, the optimal weight by solving:
W? = arg min Cw(X'Y), @)
w
Where Cw(X'¥) = Ildur(X') = YI} and ¥ = gur-(2). gar is the output
activation function for the weight matrix W. ¥ represents the true
classification labels
Input Hidden Hidden Ouput
layer ayer ayer layer
Fig.2 Classication networtdML)Suma eta. loDatatining
(2021) 1430 Page? of 4
ke
xp ie] [P]
Fig. 3 Clication network before input peturbtion tl?)
Step 2
Step 3
Input matrix perturbation X': Once the classification network is sufficiently
trained, its weight matrix Wis fixed and the training input matrix X’ is
perturbed to find X”* solution of the following problem
argnin Cue(Za¥s a
‘To perturb the input data, we use an iterative version of FGSM (see
Appendices: Fast gradient sign method) that adds a non random noise whose
direction is opposed to the gradient of the loss funetion.
Projection matrix perturbation U/: After finding the optimal perturbation
X,we look fora perturbed projection matrix U*” by solving the following
linear system
U® = arg min \|XV —X"||3. (5)
Where X is the original training matrix and V any matrix, with the same size
as UK, After the three construction steps , the output of the MLP, is
$ip(X"). Once U** is calculated, we project the original test set on the
latter to evaluate the performance of the classification network
Itis worth noting that, after recovery ofthe perturbed inputs, the classification network
(see Fig. 4) can be re-trained or tested with the fixed weight matrix W*(in Step 2). From
Step 1 and after having solved the system (4), the input matrix X ean be perturbed by
solving
ue
arg min gues XV) ~ Yl CO)
But the high dimensionality of input data makes the non-linear optimization problem.
difficult to solve and the results less accurate.
Results
In this section, the obtained results using the proposed method are reported and its per-
formance is compared to that ofthe once recommended in [33] (the Per class histogram,
see Appendices: Thin parameters for fat genomics, Table 2)
Proposed method
In the table below, we summarize the accuracy of the classification with respect to the
number of modes (principal components) k chosen between 20 and 1000.
aE) uae)" FF)
Fig. 4 Clssicaion networkatter inputs penubation MPI)Soumare etal BiaDataMining (2021) 1430 Page sof 4
1 Results obtained by the classification network, befor
3 after inputs perturbation
k MP MLPP MLPLPR
0 7646 Baas 3782
20 eas 5202 9229
30 9188 9623 9571
10 27s 9521 ass
200 sa 9565 9568
00 9595 9492 50s
000 05 9434 402
Table 1 represents in the second column (resp. third column) the results obtained by the
classification network before (resp. after) input perturbation. After input perturbation,
the training model can be evaluated using the fixed weight matrix (inthe third column)
as well as re-trained (in the last column). It is clear from the above results that input
perturbation has significantly reduced misclassification
‘Toillustrate the effectiveness of the proposed method, we display the confusion matrix
of our classification network to see the effect of input perturbation.
In Fig. 4a (before input perturbation), we observe high classification errors between
some population groups such as Chinese Dai in Xishuangbanna and the Kinh in Ho
Chi Minh Gity; Indian Telugu in the UK and Sri Lankan Tamil in the UK; or British
in England and Scotland and Utah Residents (CEPH) with Northern and Western
Ancestry. Figure 4b shows how our approach has reduced these miselassifiations,
particularly the classification error between the CDX and KHV classes from 0.95%
00.05%.
However, as the number of modes increases and the classification errors decrease, one
can notice throughout our experiences weak classification error between the British eth-
nic groups in England, Scotland and Utah Residents (CEPH) with Northern and Western
Ancestry, who appear to be genetically very similar (Figs. 5, 6,78 and 9).
‘SETSREERERS UR REECEEEE SEREGESTAESSMeEasaze=tare
si
abopa Re Evesfoone:
a
(2) Bofore input perturbation (O) After input perturbation
Fig.5 10 modeonfusion matrixSuma eta. loDatatining
(2021) 1430
‘RRSGEESSEREGREOREANCCEEE?
(a) Before input perturbation
Fig. 6 20 mode-confuson matix
7
EeapgaSst 220088
2 mae
‘SpaeREsaaEpauESSeeEEDEE=
(a) Before input perturbation
Fig.7 50.mase-confuson mare
SURESEERDgeeREEGEESEEP
(6) Before input perturbation
Fig. 8 100 mode confuson matic
8
ees
at
:
gazeebiRtyaa bed
a RESEERSQUENEGEESEEPE
{b) After input perturbation
‘MERLE ERE NA UCN SAEEETSE
(b) After inp perturbation
(b) Aer input perturbation
Page 9 of 4Soumare etal BiaDataMining (2021) 1430 Page 100f 4
S22 PARE EE RS
208s!
£
a
apn taie
gadbeteGExan G00 s099000
1a
‘ESESREEESERSUCEEEAEUCEEES SGROGESESERSENCMEEE OEE EE
(2) Before input perturbation (b) After input perturbation
Fig, 9200 mode-confusion matic
Perclass histogram
In Fig. 10, we present confusion matrices obtained by per histogram embedding methods
and Per class histogram embedding input perturbation, Perturbing per class embed-
ding input reduced misclassification errors and allowed the classifier to reach 94,49% of|
accuracy,
Discussion
Deep learning application to high-dimensional genomic data, such as SNPs is more chal-
lenging. In order to deal with problems of high dimensionality, many efforts have been
‘made. In [11] the authors proposed to learn the feature presentations using « Neural
Network followed by another classification network. Unsupervised clustering or Deep
Autoencoder is jointly trained with a classification network [13, 3 ]. However,
these methods are generally applied to datasets with relatively small features where, the
computational cost increases linearly with the number of features and they require more
training samples to converge. When Autoencoder network was trained jointly with the
classification network on the used dataset, the best accuracy obtained was 85.36%. In
5
aiid
e820)
i
ENG Pas ebaueegaoRe
ibe
El
Fr
(a) Before input perturbation
Fig. 10 Perclasshstogtam confusion matixSumare etal BloDetaMining —(2021)1430 Page 11 of 4
addition to the high dimensionality of the data, there is another challenge related to the
high genetic similarity between certain population groups. To mitigate these difficulties,
the proposed method reduces the dimension of the input data using SVD algorithm.
However the SVD algorithm extracts linear combinations of features from the input data
and fails to take into account the genetic similarity between some population groups
as shown in Figs. 4a-10a, To improve these results, the SVD projection matrix is mod:
ified to minimize the training loss function of the classification network using FGSM
algorithm. The FGSM algorithm allowed us to find the best representation of the input
features in SVD projection space. This new representation makes the elassifieation net
work more robust to small variations in the input and takes into account the genetic
similarity between different populations, as shown in the last two columns of Table 1 and
Figs. 4b-10b. We are not limited to the SVD algorithm, when Per class histogram is used
to reduce the dimension of the input data, the proposed perturbation has significantly
reduced classification errors.
‘The proposed method has achieved its best results when the input features were
reduced from 300M to 50, which means that the number of free parameters of the clas-
sification network has reduced by a factor of 6000. This method outperforms previous
work (see Appendices: Thin parameters for fat genomics) in term of accuracy and the
number of free parameters required by the model. For future work, we expect to improve
this method by using different targeted attacks algorithms with other dimensionality
reduction techniques.
Conclusion
In this work, we proposed a New Neural Network method for the prediction of indi
Vidual ancestry from SNPs data, To deal with the high dimensionality of the SNPs data,
our approach first uses SVD to reduce the dimensionality of its inputs, then train a
classification network and then reduce prediction errors by perturbing the input data set.
‘The obtained results showed how input perturbation reduced classification errors
despite genetic similarities between some ethnic groups. With such accuracy in the task
of predicting genetic ancestry, this method will make it possible to deal with more com:
plex problems in the healthcare field. We therefore, intend to apply our method to gene
expression profiles as well as SNPs data in order first to predict and then prevent the
development of patients genetic diseases.
‘Appendices
Fast gradient sign method
FGSM ((40)) : uses the gradient of the loss function to determine in which direction the
input data features should be changed to minimize the loss function :
¥ =x esign(V.CwOy)),
«isa tunable parameter. Iterative Fast Gradient Sign Method (JFGSM) consists in adding
the perturbation iteratively [47]. In our context, given any input training sample 2: (a row
of the training input matrix X) and its corresponding one-hot label ., we pertub it in the
direction ofthe input space which yields to the hightest decrease of te loss function Cw,
using the Targeted Iterative Fast Gradient Sign Method (IFGSM) given by the formulaSoumare etal BiaDataMining (2021) 1430 Page 120f14
Table 2 Dbtained ro
Model &Embedding =a MacnesEror. (8) =f ee param.
Basic sist
Rawwend2end 2k
Fondo Poon 101k
snPavec 101k
Fer dasshistorams 79K
Basic with reconstaction oM
Raw end2end with reconsiucton m13k
Random Proton with reconstuction 202k
sNP1Vec wth econstucton 202%
Perla toga wth econtucion 1a
{0 = 2f-D — sign (Ven: (2.x),
where m = 1,...,M, 2 = zj, M is the number of iterations and z;* = z;" the per-
turbed version of zi After perturbation, the rows of the matrix X" are composed of z/*
fori = 1,...,1% Where mis the number of training samples.
1000 genome project legends
Population ethnicity legend
ACB. African Caribbeans in Barbados; ASW: Americans of African Ancestry in SW USA:
BEB: Bengali from Bangladesh: CDX: Chinese Dai in Xishuangbanna; CEU: Utah Resi-
dents (CEPH) with Northern and Western Ancestry; CHB: Han Chinese in Bejing: CHS:
Southern Han Chinese; CLM: Colombians from Medellin; ESN: Esa ia; FIN:
Finnish in Finland; GBR: British in England and Scotland; GIH: Gujarati Indian from
Houston; GWD: Gambian in Western Divisions in the Gambia; IBS: Iberian Population
in Spain; ITU: Indian Telugu from the UK; JPT: Japanese in Tokyo; KHV: Kinh in Ho Chi
‘Minh City; LWK: Luhya in Webuye; MSL: Mende in Sierra Leone; MXL: Mexican Ances-
try from Los Angeles; PEL: Peruvians from Lima; PJL: Punjabi from Lahore; PUR: Puerto
Ricans; STU: Sri Lankan Tamil from the UK; TSE Toscani in Italia and YR: Yoruba in
Ibadan.
Geographical region legend
AFR: African; AMR; Ad Mixed American; EAS: East Asian; EUR: European and SAS:
South Asian.
‘Thin parameters for fat genomics
We represent in Table 2, different results from [93]
‘Authors’ contributions
Tre ahora and approves the fra manu
Funding
‘Thsproge was party funded by HSAIeNet,whichssuperted by the Natal Institutes of Heath Common Fun
Under grant number UA1HGO064) The conient of ths pueation soley the esponsbity ofthe authors an does
not recess represent he ofical ea fe Nationa linus of Hes
‘Avalabltyof dat and materials
‘The dataset ued in thes uerkis ret aval tp 1000genomesebiac uk Veo fplese/20 30502
suppating/hd_genetype-ch/ andthe opensource frais wsed be an feundhere itzsi/yntecsorhorcra/
uideheravovenvendSumare etal BloDetaMining —(2021)1430 Page 13 14
Declarations
Consent for publication
Not appicae This manuscript does nt conan data tom any ndidual peson
Competing interests
‘he authors cece that they havena comprting interests
‘Author details
"The Labora of Matnematial Madelng and Numesc n Engineering cence, Natlonal Engineeting Scot of Tun
RucBéchir Salem Bethe Compus urwertare, BP. 37,1012 Tons Bees, Uriversy of Tm Marat Tans Tans
Laboratory of lnfonratics boNathemales and batt 3 place Pat, BP. 74 1002 Tun aed neat
Pasteur de Tun Unneraty of Tun Mena, Tune, Tuna, “ADAGOS Le evédere cere, rue i raroum E
Merza, Tunis Tunisia *Cotege of sciences Eas and Appia Scent Reser Cente, mam Abdurahman in
Fal Univer, PO. Box 1982, 3148), Dama Kingdom of aud vb, Rirahmon Bn Fool iver
Dammam Saut abla
Received: 1 December 2020 Accepted: 29 arch 2021
Pubs anne: 28 June 2021
References
1 ALCS LoyEy, Salim Patan ¥, Cha KS. The asco ofhuran genet vrstion and ther useas esate
‘maners past resent and future. Hum Genet, 201055(}408 has/olara 0383 201055,
2 Coll FS S126, Chakra AA dna polymorphism dover rescue fr esarch on human gate
‘ato Geno Res. 196 12F1229-31, ipso og 101101978 12122,
3. Group Ker al Aman ofrunan genome sequence aratencontanig 12 milon single nuieatide
abmerphisms Nature. 2010968224928, .s//coLong/ 0 0383805748.
4. Moyer Lndenberg A Werberga OR Intermediate phenotypes and genic mechanisms of paychatic dors
Nat Rew Neuros. 20067003818 psd a0 1038/nI593
5. isch Searching fr genetic determinatsin the new milrium Nature 20040516783} 847-hps/oor/10
esasoiszi8
6 WieterD, MacArthur, Morales) Bucket, Hal? unkins, KemmA,FleekP, MeroloT, Hino etal The
‘at as alga cuted resource of spa assciaens Nude Ack Res 201340101 DION 6/30
(19/1053 mariah 29, Odor rivers ress.
7 Hoss, McCanyCA.PayseurBA Genetic acest inference using suppor vector machines athe active
femergsceof a unique american population, EG 2013.21) 388 ip /acuorg/I0.1028/ef92012258
8 Lee Mirco Nsen Infring ethnyram mtochontal dhe sequen. BM proceedings v5.
Spinger 2011 p.1-8.
9. Cowley GC, Tabet NIC On cre tng h model selection and subsequent selection Blasn performance
feluaton JMIR 201051:2079-107.hps//20.0/10.106) patcog 200612015.
10, Wien, Thibea-Sute Ose Neo M, SomperGonzslr Router ottaniS, Dormont D,DunemanS, Burgos
1 alot Oe al Convolutional neural ncwons for dasfcation of azherner’s dense Overview and eproducle
luster Mes nage Ara 202063101654
11. Zhou, Tung KH, ZhuX. Shen D-Eecove eae learing and fuson of mutmeealty cats ung stage wlse
‘ep neurd network fr demerts dagnoss Hum Ben Mapp, 201902) 1001=16,
12, MatlonadoC, MoreF, ConiersSoto R Ahmar, Chen FI do Aral ink AT, Scagim CA, Gevomewide
Dredeton of omple rats ino cucossing lant specs rcugh dep leming and besa equated
oust neon Pont Pant et 2025511734
13, Pumorad 5, Tesnehab MS, Zari N, Shot AAset-egenzing deep auto encoder aperosch fr clsfiction
of comple diseases sings genarics dita Ap oft Comput 202087 105718
14 Faleer®, LadhakR, Naz. Huber M Usng dep eaming to anhance cancer agnosis ardcasfcation.
Proceed often Cont on ML. New Yor ACN: 2013 Ftps/doverg/ 101 TO9/CScAN 0183641142.
15, Fergu®, MontanezCC, Abduaimms8, Lstoa ,ChalrersC Using des leaning and genome wide
_aiscation sts for oistataven pete th cssfcatoninalican-Aretan women EEIACM Tans
Comput el Boise 201817179668-78. hes /.or9/0 109TH 2078 7268267
16 FredmanS, Guthert, Fun, anks Lear ana deep medesformete acurt tenn cf =rp an inl
aria call Beanformtes 202036 712060-7
17, Do OU, Leek Choi, Le M The skin cancelation usin deep convolutional neva netwctk Mult Tool
‘App 2018778090824 eps. rg/10.21861 1936.
18, Mantesnos epezOA, Manteno: 1ope7IC, Singh, Lozano famiez Baén-Loper A, Motesnas oper A
‘isa J Amubiarat poisson deep earang model for genomic prescion of count dst. G3 Genes Gnomes
Genet, 202001241770,
19, Ganaee , Gracin, HendncDA, A deep leaming approachfor cancer detection andrlevant gene denficaton.
Insc Smpacum en bacomputne 27 Wad sant Tn Tuck unk Snaapor, So6224 Sncapors 201?
219-2.
2, Singh Lanenantn J, Robins Ol. Deeporvoe: deep esming fo predting gene expression fm stone
moaieatons Siarvormatis. 201632117) 39-48 Mtes,/ao.cr/ I. 1094/onFermates7.
21, Sen Lr Festus elton orciaerestion, ett Dts ns 1068) 131-36 hp 46 9/06/1088
a6G70008 5.
22, wen ABA YOO yr ofa an HdgeKqrsson, Contemp Maths 2074759724010
seoneya3/0855,Soumare etal BiaDataMining (2021) 1430 Page 14 of 14
2 Himion GE Smastnal Kihevly A Sutdever| Salakhutsnov FR. Improving neu nebo by praventing
‘o-aaplaton of ete detects, a ret arte 2070580 2012
24 Upp us 8, Gopaln PA deep arin approac to cect sp interactions. SW 2016 1) 0-965-75.
25. Zhou | Toyarshaya Oc Predicting les of noncoding variants wih deep leaning based sequence mod Nal
Methods. 2015 2108981-4
26 Pepin, Chang P<, Alsace, Schwan’ CohMUSLT, KUA, NewugED, Dymo gwen, Ar,
tal Aunienalsnpand saline! variant cater sing dep newra network Net otc tra, 201 B38 0)S83-7,
27, Heinen F, WatkeM, Das PP, Kamp M, Galas, ink, Sch A Ideiiaion af ego soos assoc
‘th viroar conve content of vel fabs based cn genyping by sequencing daa using deep earng
‘Genes, 220169818,
28 Lene, Hess M\ Brier H, Unsopenised deep eemng on ames data vith atzannmachines bio
2oresrans2
29, HeseM, LenzS, Piste, ulingerL, Bier H. Parton erring of deep botzmann machine fon dts
Benformatcs, 20173300)31 73-80,
an, Pepin NevburgerD, Bjomeo4, Nawjen, Loy, GrossS, McLean CY, DePistoMA. Creag 2 univers SNF
andsmalindelvaancaler wth deep evil pewarks, 2016 hips/oiarg 01101082850.
31, Balasngh Sk VipstaS Gondor AH, Panda, ihiS, Ranasubtareddy§ Aras thigh dimensional genomic
‘ita wing mapredce bed probable nel netwck Comput Methods Pog rned 2020105 108525
22, Khcardan’ Adem CoM Dannosis ond lfcation ofeancer sng hyd adel eed on vel anc
You might also like
(IJCST-V8I2P16) :P. Sravanthi, B.Sai Sudha, P.Sai Ganesh, M.Dinesh, T.Ravi Kumar, G.Rajasekharam
(IJCST-V8I2P16) :P. Sravanthi, B.Sai Sudha, P.Sai Ganesh, M.Dinesh, T.Ravi Kumar, G.Rajasekharam
12 pages