Deep Learning
Deep Learning
[email protected]
KP2WFEJ3RX
S U N I L KU M A R V U P PA L A
S U N I L .V U P PA L A @ G M A I L .C O M
W W W. L I N K E D I N .C O M / I N / S U N I LV U P PA L A /
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
Agenda
Session-1:
• DL - What, Where, Why, How? • Building DL models using Keras+Tensorflow
• Why deep learning now? • Convolutional Neural Network
• Applications of DL
• Machine learning vs Deep learning
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 3
Where it fits in?
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 4
Where to use DL?
• Finance domain (Categorical and Numerical data) :
• Identify the fraud detection in credit card transactions
Few more applications of DL are: Personalized recommendations, Prediction, Anomaly detection, Drug
discovery, Autonomous cars, Video analytics etc...
Convolution Neural Networks for Handwritten Recognition Google Brain Project on 16k
1958 Perceptron 1974 Backpropagation 1998 Cores
[email protected] 2012
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 6
Why second wave?
•More data from systems and sensors (IoT)
[email protected]
KP2WFEJ3RX
Important property:
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 7
Machine Learning vs Deep Learning
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 8
Supervised & unsupervised learning (Recap)
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 9
Artificial neural network
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 10
Tensorflow platground demo
A neural net model is composed a set of Layers.
Run multiple examples in the increasing order of
complexity
◦ Linear
◦ Comlicated circle
[email protected]
KP2WFEJ3RX
◦ Spiral
◦ Shallow learning
◦ Deep learning
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 11
[email protected]
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
Artificial neural network
Artificial neurons are elementary units in an
artificial neural network.
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 13
Back propagation
With the concept of gradient descent
◦ Forward propagation: Activation [input * weight
matrix]
◦ To get optimal value of each weight:
◦ Direction is opp. to the gradient
[email protected]
KP2WFEJ3RX ◦ Find weight with minimum error
◦ Derivative (slope of a tangent line - rate of change of a
function)
◦ Partial derivative (wrt one of the variables)
◦ Chain rule (derivatives of composite functions)
◦ calculate the error wrt each weight
◦ New weight = old weight - Derivative Rate * learning
rate
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 14
Feed forward nets
Information flow is unidirectional
• Data is presented to Input layer
• Passed on to Hidden Layer
[email protected]
KP2WFEJ3RX
• Passed on to Output layer
Information is distributed
Information processing is parallel
Backpropagation
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 15
Solving XOR with a neural net
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 16
Basic set of Layers
Dense Layer
Dropout Layer
Convolution1D
[email protected]
KP2WFEJ3RX
Convolution2D
MaxPooling1D
LSTM
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 17
Dense and Dropout layers
Dense Layer:
It creates a regular fully connected Neural net layer
Dense (output_dim , activation='linear')
◦ output_dim: (integer > 0 ) Specifies the size of the Layer ( Number of Neurons)
◦ activation: name of activation function
[email protected]
KP2WFEJ3RX
Dropout Layer:
Dropout: A Simple Way to Prevent Neural Networks from Over- fitting
Dropout ( p )
Applies Dropout to the input. Dropout consists in randomly setting a fraction p of the input units
to 0 at each update during the training phase, which helps prevent over-fitting.
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 18
Convolution1D/2D
Convolution operator for filtering neighborhoods of one-dimensional inputs.
1D convolution layer (e.g. temporal convolution)
2D convolution layer (e.g. spatial convolution over images)
Convolution1D ( nb_filter, filter_length, activation='linear', border_mode='valid',
[email protected]
KP2WFEJ3RX
subsample_length=1 )
◦ nb_filter: Number of convolution kernels to use (dimensionality of the output).
◦ filter_length: The extension (spatial or temporal) of each filter.
◦ activation: name of activation function to use
◦ border_mode: 'valid' or 'same'.
◦ subsample_length: factor by which to subsample output.
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 19
MaxPooling1D
Max pooling operation for temporal data.
The max-pooling layer would reduce the input Matrix into a down sampled size with max value for each block.
Please refer to the image shown below for an example.
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 20
Activation Functions
In Neural Networks, the activation function of a node defines the output of that node given an
input or set of inputs.
A standard computer chip circuit can be seen as a digital network of activation functions that
can be "ON" (1) or "OFF" (0), depending on input.
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 21
Properties of activation function
◦ Nonlinear:
◦ When the activation function is non-linear, then a two-layer neural network can be proven to be a universal function approximator.
◦ Continuously differentiable:
◦ This property is necessary for enabling gradient-based optimization methods.
◦ Range:
[email protected]◦ When the range of the activation function is finite, gradient-based training methods tend to be more stable.
KP2WFEJ3RX
◦ Smaller learning rates are typically necessary.
◦ Monotonic:
◦ When the activation function is monotonic, the error surface associated with a single-layer model is guaranteed to be convex.
◦ Smooth
◦ Functions with a Monotonic derivative have been shown to generalize better in some cases.
Approximates identity near the origin:
◦ The neural network will learn efficiently when its weights are initialized with small random values.
◦ When the activation function does not approximate identity near the origin, special care must be used when initializing the
weights.
◦ https://en.wikipedia.org/wiki/Activation_function
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 22
Activation functions
2
20
Logistic
2
18 1.5
16
14
1.5
1
Hyperbolic tangent 1
1
0.5
12
0.5
y=
0
10
exp( x) − exp(− x)
8 0 -0.5
y= 1 + exp(− x)
6 -0.5 -1
exp( x) + exp(− x)
-1 -1.5
2
-1.5 -2
0 -10 -8 -6 -4 -2 0 2 4 6 8 10
0 2 4 6 8 10 12 14 16 18 20
-2
[email protected] -10 -8 -6 -4 -2 0 2 4 6 8 10
KP2WFEJ3RX
Linear
y=x
Rectifier / ramp function
f(x) = max(0,x)
x is the input to a neuron.
smooth approximation to the rectifier
is softplus
f(x) = ln(1+ex)
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 23
Activation functions
Unit step (threshold):
The transfer function translates the input signals to
output signals. Four types of transfer functions are
commonly used, Unit step (threshold), sigmoid,
piecewise linear, and Gaussian.
[email protected]
KP2WFEJ3RX
Sigmoid:
The sigmoid function consists of 2 functions, logistic
and tangential. The values of logistic function range
from 0 and 1 and -1 to +1 for tangential function.
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 24
Deep Learning Algorithms
MLP – Multi Layer perceptron
◦ A multilayer perceptron (MLP) is a feed forward artificial neural network model that maps sets of input
data onto a set of appropriate outputs.
◦ An MLP consists of multiple layers of nodes in a directed graph, with each layer fully connected to the
next one.
◦ Had multiple hidden layers with logistic regression classifier transformation
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 25
Parameters to vary for tuning
•Number of layers •Batch size
•Number of neurons in each layer •Momentum parameter (weightage given to
earlier steps taken in the process of gradient
•Activation function in each layer
descent)
•Number of epochs
[email protected]
KP2WFEJ3RX •Kernels
•Error/loss functions
•Number of features
•Iteration (equivalent to when a weight update is
done) •Number of filters for images
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 26
Recap of evaluation measures
Accuracy: Overall, how often is the classifier correct?
(TP+TN)/total = (100+50)/165 = 0.91
False Positive Rate: When it's actually no, how often does it predict
yes?
Precision: When it predicts yes, how often FP/actual no = 10/60 = 0.17
is it correct?
TP/predicted yes = 100/110 = 0.91 Specificity: When it's actually no, how often does it predict no?
TN/actual no = 50/60 = 0.83
"Sensitivity" or "Recall": When it's actually equivalent to 1 minus False Positive Rate
yes, how often does it predict yes?
TP/actual yes = 100/105 = 0.95
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 27
Keras and Demos
Why we need Keras?
◦ Keras: Deep Learning library for Theano and
TensorFlow
◦ An API spec for building DL models across many
platforms
[email protected]
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 29
[email protected]
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
Convolutional Neural Network
CNN - Convolutional Neural Network
◦ Feed-forward artificial neural network
◦ Convolutional networks were inspired by
biological processes
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 31
Convolution operation
[email protected]
KP2WFEJ3RX
• Most deep networks use ReLU - max(0,x), since it trains much faster, is more expressive than logistic
function and prevents the gradient vanishing problem.
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 32
Convolution operation
Raw Image Pixel Filter or Kernel or Feature detector
[email protected]
KP2WFEJ3RX
Convolved Feature or
Activation Map or the
Feature Map.
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 33
Convolutional Neural Network
An image input constitutes a 3-dimensional structure called the Input Volume (255x255x3).
CNN’s use filters as kernels where the parameters or weights have to be learnt.A filter is a matrix of lower
size than the input to it.
The inputs are convolved with the filters and passed through the activation function.
[email protected]
KP2WFEJ3RX
The weights of the kernels are randomly initialized and are modified during training based on error-
minimization using backpropagation.
The real values of the kernel matrix change with each learning iteration over the training set, indicating that
the network is learning to identify which regions are of significance for extracting features from the data.
Stride: The shift of filter after each convolution.It can be increased from 1 to a larger value to decrease
overfitting.
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 34
ReLu and Max pooling
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 35
CNN in summary
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 36
[email protected]
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
Recurrent Networks
Feed forward networks:
◦ Information only flows one way
◦ One input pattern produces one output
◦ No sense of time (or memory of previous state)
Recurrency
[email protected]
KP2WFEJ3RX
◦ Nodes connect back to other nodes or themselves
◦ Information flow is multidirectional
◦ Sense of time and memory of previous state(s)
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 39
LSTM (Long Short Term Memory)
It create a layer of Long-Short Term Memory units .
LSTM (output_dim , activation='tanh', inner_activation='hard_sigmoid')
output_dim: dimension of the internal projections and the final output.
activation: name of activation function to use
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 40
Long short-term memory
LSTM - Long short-term memory
◦ Recurrent neural network (RNN)
◦ Take input not just the current input
example they see, but also what they
perceived one step back in time.
Feedback loop, ingesting their own
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 41
Text analysis
[email protected]
KP2WFEJ3RX
http://slideplayer.com/slide/9885340/
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
Type of chatbots
Usecases:
• Pizza Hut to help you order a pizza
• Uber to book a taxi
• CNN to keep you up-to-date with news
[email protected]
KP2WFEJ3RX content
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 43
Conversational interfaces
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 44
Pretrained models
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 45
Convnets and pretrained models
• LeNet (1990s)
• AlexNet (2012)
• ZF Net (2013)
[email protected]
KP2WFEJ3RX
• GoogLeNet (2014)
• VGGNet (2014)
• ResNets (2015)
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 46
Architecture of Alex Krizhevsky et al.
● 8 layers total
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
Impact on Computer Vision
ImageNet Challenge2012
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
Practical tips
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.
DL Project ideas
Image Speech
◦ Captioning ◦ Alexa / Home APIs
◦ extract embedded text ◦ Local languages
◦ emoji - extract sentiment
[email protected] Numerical / Categorical
KP2WFEJ3RX
◦ Volume, result prediction
Text
◦ Time series forecasting - weather / server
◦ Sarcasm
◦ Satillite data analysis - ISRO
◦ Chatbots - specific topic
◦ Govt data analysis
◦ Sentiment analysis
◦ http://data.gov.in
◦ Local languages
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 50
DL project ideas contd
Video Robot
◦ analytics - speed control ◦ Path planning / recommendations
◦ search ◦ reinforcement learning
◦ annotated data
Multimodal
[email protected]
KP2WFEJ3RX
Recommender systems
◦ Chatbots ◦ Specific product / item category
◦ Get information from multiple sources
◦ Generative
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 51
Practical challenges and tips while using
ML / DL
◦ Data availability
◦ Rules based + ML
◦ Application revisited: Gien ticket data of client, identify the potential candidates for
[email protected]
KP2WFEJ3RX
◦ Used ML for clustering the ticket data with preprocessing
◦ Tried ML algorithms including ensemble to reach
◦ Extended DL algorithms but the improvement is not more than 5%
◦ Discussion
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 52
Cheat sheets
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 53
Learning path
Building knowledge : Practise with assignments and projects:
K1. Refresh your fundamentals on statistics, P1. Kaggle challenges with available data
probability and linear algebra Cuisine prediction, lung cancer,
K2. Do Course era deep learning Choose text or image or
categorical/nurmerical probems
K3. Refer specific concepts in
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 54
Learning path references
K1. Relearning K4. Advanced topics
https://in.udacity.com/course/intro-to-descriptive-statistics-- https://www.youtube.com/watch?v=RtxI449ZjSc&feature=
ud827 relmfu
https://www.khanacademy.org/math/linear-algebra
P1. Competitions
K2. Online courses:
[email protected]
KP2WFEJ3RX
https://www.kaggle.com/competitions
https://www.coursera.org/specializations/deep-learning
http://www.image-net.org/challenges/LSVRC/
https://in.udacity.com/course/deep-learning--ud730
P2. Assessement of your skills
K3: Books:
https://datahack.analyticsvidhya.com/contest/all/
http://www.deeplearningbook.org/
P3. Github profile
http://neuralnetworksanddeeplearning.com/
http://deeplearning.stanford.edu/tutorial/
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 55
References
• http://machinelearningmastery.com/4-steps-to-get-started-in-machine-learning/
• https://docs.scipy.org/doc/scipy/reference/generated/scipy.optimize.minimize.html
• https://archive.ics.uci.edu/ml/datasets/Adult
• http://www.datasciencecentral.com/profiles/blogs/the-data-science-project-lifecycle
• https://project.inria.fr/deeplearning/files/2016/05/DLFrameworks.pdf
• http://www.wildml.com/2015/11/understanding-convolutional-neural-networks-for-nlp/
[email protected]
KP2WFEJ3RX • http://cs231n.stanford.edu/
• http://www.iro.umontreal.ca/~lisa/pointeurs/theano_scipy2010.pdf
• http://www.kdnuggets.com/software/text.html
• http://neuralnetworksanddeeplearning.com/
• deepmind.com/blog
• http://machinelearningmastery.com/sequence-classification-lstm-recurrent-neural-networks-python-keras/
• https://docs.google.com/viewer?a=v&pid=sites&srcid=ZGVmYXVsdGRvbWFpbnxhcmlqaXRsYWhhfGd4OjdhNzhiOTd
lZjM4NTRiOWY
• Machine learning, Deep learning courses in CourseEra by Andrew NG
• https://becominghuman.ai/cheat-sheets-for-ai-neural-networks-machine-learning-deep-learning-big-data-678c51b4b463
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action. 56
Thank you
[email protected]
KP2WFEJ3RX
Proprietary content. © Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited.
This file is meant for personal use by [email protected] only.
Sharing or publishing the contents in part or full is liable for legal action.