VR Part2 Lecture 5 Annotated
VR Part2 Lecture 5 Annotated
Lecture 5: Transformers
Image Captioning
Encoder
Representation
Attention in LSTMs for Image Captioning
Sequence Modeling : Issues with RNN Architecture
Language Translation
http://jalammar.github.io/illustrated-transformer/
https://papers.nips.cc/paper/7181-attention-is-all-you-need.pdf
Self Attention- Motivation
Self Attention: A method to improve word embeddings.
Example : “The animal didn’t cross the street because it was too tired”
dmodel
= dk
Keys:
= dk
Values:
= dv
To control the input
values to softmax
function
http://jalammar.github.io/illustrated-transformer/
The
animal
The
animal
https://papers.nips.cc/paper/7181-attention-is-all-you-need.pdf
Advantages:
Parallelizable Computations
Long Range Dependency Modeling.
http://jalammar.github.io/illustrated-transformer/
Every position in the decoder to attend over all positions in the input
sequence