ca3dl
ca3dl
Deep learning is a subset of machine learning, which itself is a branch of artificial intelligence (AI). It focuses on
algorithms that mimic the structure and functioning of the human brain, known as artificial neural networks. These
networks consist of multiple layers, giving rise to the term "deep" learning because of the depth created by having many
layers of interconnected neurons.
Visualization of Backpropagation:
1. Forward Propagation:
Input → Hidden Layer → Output Layer → Compute Loss
Input → [Hidden Layer] → Output → Loss
2. Backward Propagation:
Error propagates back, weights get adjusted
Output Layer ← [Hidden Layer] ← Input
Types of Optimizers
Optimizers are categorized into different types based on how they update weights:
2. Adaptive Optimizers
• Momentum: Accelerates gradient descent by adding a fraction of the previous update.
• Nesterov Accelerated Gradient (NAG): Looks ahead to adjust the learning rate dynamically.
• Adagrad (Adaptive Gradient Algorithm): Adjusts learning rates individually for each parameter.
• RMSprop (Root Mean Square Propagation): Uses an exponentially decaying average of squared
gradients to normalize updates.
• Adam (Adaptive Moment Estimation): Combines Momentum and RMSprop to provide stable and
efficient updates.
• AdaMax: A variant of Adam using the infinity norm.
• Nadam (Nesterov-accelerated Adaptive Moment Estimation): Adam with Nesterov acceleration.
A loss function is a mathematical function that measures the difference between the actual (ground truth)
output and the predicted output of a model. The goal of training a neural network is to minimize this loss,
making the model more accurate.