Natural Language Processing With Deep Learning 1 PDF
Natural Language Processing With Deep Learning 1 PDF
• Machine translation
• Spoken dialog systems
• Complex question answering
NLP in industry … is taking off
• Search (written and spoken)
• Online advertisement matching
• Automated/assisted translation
• Sentiment analysis for marketing or finance/trading
• Speech recognition
• Chatbots / Dialog agents
• Automating customer support
• Controlling devices
• Ordering goods
What’s special about human language?
A human language is a system specifically constructed to convey the
speaker/writer’s meaning
• Not just an environmental signal, it’s a deliberate communication
• Using an encoding which little kids can quickly learn (amazingly!)
A human language is a discrete/symbolic/categorical signaling system
• rocket = 🚀; violin = 🎻
• With very minor exceptions for expressive signaling
(“I loooove it.” “Whoomppaaa”)
• Presumably because of greater signaling reliability
• Symbols are not just an invention of logic / classical AI!
What’s special about human language?
The categorical symbols of a language can be encoded as a signal
for communication in several ways:
• Sound
• Gesture
• Images (writing)
The symbol is invariant across different encodings!
lab
2. What’s Deep Learning (DL)?
• Deep learning is a subfield of machine learning
3.3. APPROACH
#(λ ,y,x)
PM (y|x) ∝ ∏ θλ
Machine Learning vs. Deep Learning
ILSVRC
···
0.286
0.792
−0.177
−0.107
expect = 0.109
−0.542
0.349
0.271
0.487
Word similarities
1. frogs
2. toad litoria leptodactylidae
3. litoria
4. leptodactylidae
5. rana
6. lizard
7. eleutherodactylus
rana eleutherodactylus
http://nlp.stanford.edu/projects/glove/
Representations of NLP Levels: Morphology
• Traditional: Words are prefix stem suffix
made of morphemes un interest ed
• DL: !"#$%&!"'&()*$%&
• every morpheme is a vector !! ! "!
• a neural network combines
two vectors into one vector !"#$%&!"'&($%& )*$'(
• Luong et al. 2013 !! ! "!
!"!"# #$%&!"'&($%&
Question Answering
Temporal Q: What is the correct order of events?
a: PDGF binds to tyrosine kinases, then cells divide, then wound healing
b: Cells divide, then PDGF binds to tyrosine kinases, then wound healing
57 (9.74%)
Figure 3: Rules for determining the regular expressions for queries concerning two triggers. In each table, the condition
• DL: Again, a deep learning architecture can be used!
column decides the regular expression to be chosen. In the left table, we make the choice based on the path from the root to
the Wh- word in the question. In the right table, if the word directly modifies the main trigger, the DIRECT regular expression
• Facts are stored in vectors
is chosen. If the main verb in the question is in the synset of prevent, inhibit, stop or prohibit, we select the PREVENT regular
expression. Otherwise, the default one is chosen. We omit the relation label S AME from the expressions, but allow going
through any number of edges labeled by S AME when matching expressions to the structure.
Translation
The protests escalated over the weekend <EOS>
generated
0.1 0.2 0.4 0.5 0.2 -0.1 0.2 0.2 0.3 0.4 -0.2 -0.4 -0.3
0.3 0.6 0.4 0.5 0.6 0.6 0.6 0.6 0.6 0.4 0.6 0.6 0.5
0.1 -0.1 0.3 0.9 -0.1 -0.1 -0.1 -0.1 -0.1 -0.1 -0.1 -0.1 -0.1
-0.4 -0.7 -0.2 -0.3 -0.5 -0.7 -0.7 -0.7 -0.7 -0.7 -0.7 -0.7 -0.7
0.2 0.1 -0.3 -0.2 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1
Sentence 0.2
-0.2
0.2
0.6
0.1
0.3
0.2
0.6
0.2
-0.8
0.2
0.6
0.2
-0.1
-0.4
0.6
0.2
0.6
-0.1
0.6
0.2
0.4
0.3
0.6
0.2
0.6
meaning -0.1
0.1
-0.1
-0.7
-0.1
-0.7
-0.1
-0.4
-0.1
-0.5
-0.1
-0.7
-0.1
-0.7
-0.1
-0.7
-0.1
0.3
-0.1
0.3
-0.1
0.2
-0.1
-0.5
-0.1
-0.7
0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1
is built up
0.1 0.1
0.2 0.4 0.2 0.2 0.4 0.2 0.2 0.2 0.2 -0.1 -0.2 -0.4 0.2
0.6 -0.6 -0.3 0.4 -0.2 0.6 0.6 0.6 0.6 0.3 0.6 0.5 0.6
-0.1 0.2 -0.1 0.1 -0.3 -0.1 -0.1 -0.1 -0.1 -0.1 0.1 -0.5 -0.1
-0.7 -0.3 -0.4 -0.5 -0.4 -0.7 -0.7 -0.7 -0.7 -0.7 0.3 0.4 -0.7
0.1 0.4 0.2 -0.2 -0.2 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1
Source Die Proteste waren am Wochenende eskaliert <EOS> The protests escalated over the weekend Feeding in
sentence last word
Next week (Richard): how neural networks work and how they can
use these vectors for all NLP levels and many different applications