Lec4-5 - Modern Syntactic Analysis
Lec4-5 - Modern Syntactic Analysis
NPJohn-NNP liked
VP liked-VBD
NNP VBD NPdog-NN John dog
John liked DT Nominal dog-NN
the in
the Nominal
dog-NN
PPin-IN
NN IN NPpen-NN pen
dog in DT Nominal pen-NN the
the NN
pen RIAD SONBOL - NLP COURSE 4
Example
▪ Disadvantages:
◦ Some loss of expressivity
◦ Another formalism: what are the annotation
standards, how to build corpora, etc.?
VP
VP
PP
NP NP
Breaking Dawn has broken into the box office top ten
Example
S
VP
VP
PP
NP NP
Breaking Dawn has broken into the box office top ten
NP VP PP NP
Why Chunking?
▪ Used when full parse unnecessary
◦ Or infeasible or impossible (when?)
▪ Extraction of subcategorization frames
◦ Identify verb arguments
◦ e.g. VP NP
◦ VP NP NP
◦ VP NP to NP
▪ Information extraction: who did what to whom
▪ Summarization: Base information, remove mods
▪ Information retrieval: Restrict indexing to base NPs
Approaches
▪ Finite-state Approaches
◦ Grammatical rules in FSTs
◦ Cascade to produce more complex structure
▪ Machine Learning
◦ Similar to POS tagging
Finite-State Rule-Based Chunking
▪ Hand-crafted rules model phrases
◦ Typically application-specific
▪ Left-to-right longest match (Abney 1996)
◦ Start at beginning of sentence
◦ Find longest matching rule
◦ Greedy approach, not guaranteed optimal
Finite-State Rule-Based Chunking
▪ Chunk rules:
◦ Cannot contain recursion
◦ NP -> Det Nominal: Okay
◦ Nominal -> Nominal PP: Not okay
▪ Examples:
◦ NP → (Det) Noun* Noun
◦ NP → Proper-Noun
◦ VP → Verb
◦ VP → Aux Verb
Cascading FSTs
▪ Richer partial parsing
◦ Pass output of FST to next FST
▪ Approach:
◦ First stage: Base phrase chunking
◦ Next stage: Larger constituents (e.g. PPs, VPs)
◦ Highest stage: Sentences
Example
Chunking by Classification
▪ Model chunking as task similar to POS tagging
▪ Instance: tokens
▪ Labels:
◦ Simultaneously encode segmentation & identification
◦ IOB (or BIO tagging)
◦ Segment: B(eginning), I (nternal), O(utside)
◦ Identity: Phrase category: NP, VP, PP, etc.
◦ The morning flight from Denver has arrived
◦ NP-B NP-I NP-I PP-B NP-B VP-B VP-I
Features for Chunking
▪ What are good features?
◦ Preceding tags
◦ for 2 preceding words
◦ Words
◦ for 2 preceding, current, 2 following
◦ Parts of speech
◦ for 2 preceding, current, 2 following
▪ Vector includes those features + true label
Chunking as Classification
▪ Example
State-of-the-Art
▪ Base NP chunking: 0.96
▪ Complex phrases:
◦ Learning: 0.92-0.94
◦ Rule-based: 0.85-0.92
▪ Limiting factors:
◦ POS tagging accuracy
◦ Inconsistent labeling (parse tree extraction)
◦ Conjunctions
◦ Late departures and arrivals are common in winter
◦ Late departures and cancellations are common in winter
Named Entity Recognition