Updated DL Handbook 2023-24
Updated DL Handbook 2023-24
Course Objectives:
Course Outcomes:
CO1.: Understand the basics of Deep Learning and apply the tools to implement deep learning
applications
CO2.: Evaluate the performance of deep learning models (e.g., with respect to the bias-variance
trade- off, overfitting and underfitting, estimation of test error).
CO3.: To apply the technique of Convolution (CNN) and Recurrent Neural Network (RNN) for
implementing Deep Learning models
Unit I: Foundations of Deep learning (07 Hours)
What is machine learning and deep learning?, Supervised and Unsupervised Learning, bias
variance tradeoff, hyper parameters, under/overfitting regularization, Limitations of machine
learning, History of deep learning, Advantage and challenges of deep learning. Learning
representations from data , Understanding how deep learning works in three figures, Common
Architectural Principles of Deep Network, Architecture Design, Applications of Deep Learning,
Introduction and use of popular industry tools such as Tensor FLow, Keras, PyTorch, Caffe,
Shogun.
Unit II: Deep Neural Networks (DNNs) (07 Hours)
Introduction to Neural Networks :The Biological Neuron, The Perceptron, Multilayer Feed-
Forward Networks , Training Neural Networks :Backpropagation and Forward propagation
Activation Functions :Linear ,Sigmoid, Tannh, Hard Tanh, Softmax, Rectified Linear, Loss
Functions :Loss Function Notation , Loss Functions for Regression , Loss Functions for
Classification, Loss Functions for Reconstruction, Hyperparameters : Learning Rate,
Regularization, Momentum, Sparsity, Deep Feedforward Networks – Example of Ex OR, Hidden
Units, cost functions, error backpropagation, Gradient-Based Learning, Implementing Gradient
Descent, vanishing and Exploding gradient descent, Sentiment Analysis, Deep Learning with
Pytorch, Jupyter, colab.
Recurrent and Recursive Nets: Unfolding Computational Graphs, Recurrent Neural Networks,
Bidirectional RNNs, Encoder-Decoder Sequence-to-Sequence Architectures, Deep Recurrent
Networks, Recursive Neural Networks, The Challenge of Long-Term Dependencies, Echo State
Networks, Leaky Units and Other Strategies for Multiple Time Scales, The Long Short-Term
Memory and Other Gated RNNs, Optimization for Long-Term Dependencies, Explicit Memory.
Practical Methodology: Performance Metrics, Default Baseline Models, Determining Whether to
Gather More Data, Selecting Hyper parameters
Text Books:
1. Good fellow, I., Bengio, Y.,,Courville, A, ―Deep Learning, MIT Press, 2016.
2. Josh Patterson & Adam Gibson, ―Deep Learning
3. Charu Agarwal, ―Neural Networks and deep learning, A textbook
4. Nikhil Buduma, ―Fundamentals of Deep Learning, SPD
5. Francois chollet, ―Deep Learning with Python
References Books:
1. Richard S. Sutton and Andrew G. Barto, ―Reinforcement Learning: An Introduction
2. by SethWeidman, ―Deep Learning from Scratch: Building with Python from First Principles
O‘Reily
3. Francois Duval, ―Deep Learning for Beginners, Practical Guide with Python and Tensorflow
E-Books:
1. http://csis.pace.edu/ctappert/cs855-18fall/DeepLearningPractitionersApproach.pdf .
2. https://www.dkriesel.com/_media/science/neuronalenetze-en-zeta2-1col-dkrieselcom.pdf
Department of Computer Engineering
TEACHING PLAN
Academic Year: 2023-24(Semester: VIII)
Course Title: Deep Learning Subject Code: Class: B.E. Division: A & B
(DL) 410251
Term: II Date of commencement of classes: Date of conclusion of teaching:
11/12/2023 19/04/2024
Lecture Examination Scheme
Schedule:
3Hrs/ Week Theory:100 M Term Work Practical Oral
In Sem:30 M (1Hr)
End Sem:70 M -
(2Hrs.30 min.) 50 50
Subject Mrs. Sonal Previous 3 Years 2021-22 2022-23 2023-24
Teacher Fatangare & Mrs University Result
Jyoti Ragatwan NA
UNIT – I: Foundations of Deep learning (07 Hours)
What is machine learning and deep learning?, Supervised and Unsupervised Learning, bias variance tradeoff,
hyper parameters, under/overfitting regularization, Limitations of machine learning, History of deep learning,
Advantage and challenges of deep learning. Learning representations from data , Understanding how deep
learning works in three figures, Common Architectural Principles of Deep Network, Architecture Design,
Applications of Deep Learning, Introduction and use of popular industry tools such as Tensor FLow, Keras,
PyTorch, Caffe, Shogun.
PLAN ACTUAL
Lect. Date Topics Date Topics covered Reasons
No. for Deviation
1 What is machine learning and
deep learning?,Supervised
and Unsupervised Learning,
PLAN ACTUAL
Lect. Date Topics Date Topics covered Reasons
No. for Deviation
15 Introduction, CNN
architecture overview,
16 The Basic Structure of a
Convolutional Network-
Padding,
17 Strides, Typical Settings, the
ReLU layer, Pooling,
18 Fully Connected Layers
32 Generative adversarial
network (GAN),
33 Generative adversarial
network (GAN),
34 discriminator network,
generator network,
35 types of GAN, Applications
of GAN networks
Make up Classes
Contents Beyond syllabus
PLAN ACTUAL
Lect. Date Topics Date Topics covered Reasons
No. for Deviation
36 Introduction of deep
reinforcement learning,
37 Markov Decision Process,
basic framework of
reinforcement learning
38 challenges of reinforcement
learning, Dynamic
programming algorithms for
reinforcement learning
39 Q Learning and Deep Q-
Networks, Deep Q recurrent
networks
40 Q Learning and Deep Q-
Networks, Deep Q recurrent
networks
41 Simple reinforcement
learning for Tic-Tac-Toe.
42 Introduction of deep
reinforcement learning,
Make up Classes
SUMMARY
i. Keras
ii. Pytorch
iii. Caffe
iv. Shogun
v. Tensorflow
12. What is the Loss function? Explain Loss function for Regression
13. Draw and Explain the architecture of SOM ( Self Organizing Maps)
14. Draw and Explain the architecture of Single layer Feedforward Networks
15.What are loss functions?
6. What is Stride? What is the effect of high Stride on the feature map?
13. Why do we prefer Convolution Neural networks (CNN) over Artificial Neural networks (ANN)
12. Discuss the importance of pooling layer in Convolutional Neural Networks (CNN)
13. Describe the shape or dimension of input data for a recurrent neural networks
Instructions to Candidates:
1. Attempt Questions Q.1 OR Q.2, Q.3 OR Q.4.
2. Neat diagrams must be drawn wherever necessary.
3. Assume suitable data, if necessary.
1 PO1
A. What are Deep stacking networks? 5 CO1
PO5
PO1
B. Explain Convolution neural network 5 CO1
PO3
OR
2 PO1
A. Short Note on LSTM networks & GRU networks 5 CO1
PO5
PO1
B. Short note on applications of Deep learning, 5 CO1
PO12
OR
Instructions to Candidates:
1. Attempt Questions Q.1 OR Q.2, Q.3 OR Q.4.
2. Neat diagrams must be drawn wherever necessary.
3. Assume suitable data, if necessary.
B. 5 CO1 PO2
Explain Convolution neural networks
C. 5 CO1 PO5
Explain Recurrent neural networks
OR
B. 5 CO2 PO2
What is Activation Functions.?Explain any two
OR
4 A. 5 CO2 PO5
What are the methods for tuning hyperparameters
CO2 PO7
B. Explain Gradient Descent 5
CO2
C. What is the Loss function? Explain Loss function for 5 PO12
Regression
Instructions to Candidates:
1. Attempt Questions Q.1 OR Q.2, Q.3 OR Q.4.
2. Neat diagrams must be drawn wherever necessary.
3. Assume suitable data, if necessary.
Q.
Questions Marks CO PO
No.
OR
2 PO3
A. Explain how to train CNN with example 5 CO3
PO5
OR
Instructions to Candidates:
1. Attempt Questions Q.1 OR Q.2, Q.3 OR Q.4.
2. Neat diagrams must be drawn wherever necessary.
3. Assume suitable data, if necessary.
B. Explain the terms “Valid Padding” and “Same Padding” in 5 CO3 PO2
CNN.
C. What are the different types of Pooling? Explain their 5 CO3 PO3
characteristics
OR
2 A. Explain padding and strides with examples 5 CO3 PO4
B. Describe the shape or dimension of input data for a recurrent 5 CO4 PO7
neural networks
Instructions to Candidates:
1. Attempt Questions Q.3 OR Q.4, Q.5 OR Q.4, Q.5 OR Q.6 Q. 7 OR Q.8
2. Neat diagrams must be drawn wherever necessary.
3. Assume suitable data, if necessary.
1
A. Explain importance and different types of pooling. 6 CO3
PO3
OR
OR
OR
PO6
B. What are applications of GAN? 5 CO5
PO7
B. Define the Q function. How to take action given a Q function 6 CO6 P04
OR
PO4,
B. Explain Q Learning vs Deep Q Learning 6 CO6
PO5
Instructions to Candidates:
1. Attempt Questions Q.1 OR Q.2, Q.3 OR Q.4 Q.5 OR Q.6 Q. 7 OR Q.8
2. Neat diagrams must be drawn wherever necessary.
3. Assume suitable data, if necessary.