LSTM Material 1
LSTM Material 1
information.
• Information doesn’t not persist.
• But there are cases we need more context:
Recurrent Neural Network (RNN) Example: “I grew up in France…(100 words)… I
speak fluent French.” Recent information
• It’s a network with loops in them. suggests that the next word is probably the
• Information can persist. name of a language, but if we want to narrow
• Can be thought of multiple copies of the same down which language, we need the context of
network, passing the message to successor. France, from further back. It’s entirely possible
for the gap between the relevant information
and the point where it is needed to become
very large.
• Unfortunately, as that gap grows, RNNs
become unable to learn to connect the
information.
• Conclusion: RNN cannot handle long-term
dependencies.
It can connect information if there is less
content, as the content/words increases, it gets
difficult for it to connect the information.
Structure: