Lecture4 - Convnets For CV Slide
Lecture4 - Convnets For CV Slide
Lecture 4:
Convolutional
architectures
Used Resources
same color
= Wij is the matrix connecting
same matrix of the ith input channel with the
connection jth feature map
Desiderata for networks for vision
Our goal is to design neural networks that are specifically adapted
for such problems
Example:
Discrete Convolution
Example:
with rows and columns flipped
Discrete Convolution
Why pooling?
֍ Introduces invariance to local translations
֍ Reduces the number of hidden units in hidden layer
Example: Pooling
Bishop’s book
Elastic Distortions
Can add ‘‘elastic’’ deformations (useful in character recognition)
Bishop’s book
Elastic Distortions
Can add ‘‘elastic’’ deformations (useful in character recognition)
Bishop’s book
Conv Nets: Examples
Optical Character Recognition, House Number and Traffic Sign
classification
Conv Nets: Examples
Pedestrian detection
Examples of “Hammer”
Deng et al. “Imagenet: a large scale hierarchical image database” CVPR 2009
Important Breakthroughs
Deep Convolutional Nets for Vision (Supervised)
Krizhevsky, A., Sutskever, I. and Hinton, G. E., ImageNet Classification with Deep
Convolutional Neural Networks, NIPS, 2012.
֍ Depth
֍ Width
֍ Parameter count
How to Choose Architecture
Many hyper-parameters:
Number of layers, number of feature maps
֍ Cross Validation
֍ Smarter Strategies
Layer 7: Full
Trained on Imagenet
dataset [Deng et al. CVPR’09] Layer 6: Full
Layer 4: Conv
Layer 3: Conv
Layer 4: Conv
Layer 3: Conv
Layer 4: Conv
Layer 3: Conv
(Layers 3 & 4)
Layer 7: Full
fully connected:
https://towardsdatascience.com/a-simple-guide-to-the-versions-of-the-inception-network-7fc52b863202
Larger kernel good for global features, and smaller kernel for local features.
number
of filters 1x1
3x3
5x5
number
of filters 1x1
3x3
5x5
֍ Width of inception modules ranges from 256 filters (in early modules)
to 1024 in top inception modules.
֍ Can remove fully connected layers on top completely
֍ Number of parameters is reduced to 5 million
֍ 6.7% top-5 validation error on Imagnet
Intuition: helps vanishing gradients; encourage reuse features (& hence reduce
parameter count);
Network is underperforming
֍ Compute flops and nr. params. → if too small, make net larger
֍ Visualize hidden units/params → fix optimization