Sequence Learning Problem
Sequence Learning Problem
Problem
Unfolding Computational graph, RNN, Sequence Modeling
Conditioned on Context
Unfolding the Computational graph
❖ A computational graph is a way to formalize the structure of a set of
computations, such as those involved in mapping inputs and parameters to
outputs and loss.
❖ The idea of unfolding a recursive of recurrent computation into a
computational graph that has a repetitive structure, typically corresponding to
a chain of events. Unfolding this graph results in the sharing of parameters
across a deep network structure.
❖ For example consider the classical form of a dynamical system:
s(t) = f ( s(t-1); θ )
s(t) is called the state of the system / network at timestep i
Unfolding the Computational graph
❖ The above Equation is recurrent because the definition of s at time t refers back to
the same definition at time t - 1.
❖ For a finite number of time steps τ the graph can be unfolded by applying the
definition τ - 1 times. For example if we unfold equation for τ = 3 time steps, we
obtain
s(3) = f ( s(2); θ )
a(t)=b+Wh(t-1)+Ux(t)
h(t)=tanh(a(t))
o(t)=c+Vh(t)
This external context can come from a variety of sources, such as prior text,
metadata, or another related sequence.
Sequence Modeling vs Conditioned on Context
Sequence Modeling:
In general, sequence modeling involves training a model to predict or generate the
next element in a sequence based on previous elements. A classic example of
this is language modeling, where the task is to predict the next word in a sentence
given the previous words.
Conditioned on Contexts:
When the model is conditioned on a context, it means that the prediction of each
element in the sequence is influenced not only by the preceding elements in
the sequence but also by additional information provided as context.
Conditioned on output of the previous step