Unit No 3
Unit No 3
UNIT NO: 3
Q. No Question Marks
1 What is Parsing? Explain Chunking. 6
ANS
2 Write Short note on Hybrid of Rule Based and Probabilistic parsing. 6
ANS
3 Perform parsing using simple top down parsing for the sentence: 6
“The dogs cried”
Using the grammar given below:
S- NP VP
NP- ART N
NP- ART ADJ N
VP- V
VP- V NP
ANS
4 For CFGs given: 6
S – NP VP
VP – V NP
NP – Det N
Draw the Shift-Reducer parser in processing the sentence:
“The Woman Saw a puppy”
Use the following Lexical entries to create the chart parser.
The | a: Det
Woman | puppy: N
Saw: V
ANS
5 Explain the Statistical parsing. Describe usage of Probabilistic Context free Grammar(PCFG) in 6
NLP.
ANS
6 What are different parsing techniques? Explain Dependency parsing and Constituency parsing 6
ANS
7 What is POS tagging? Explain any one algorithm used for POS tagging. 6
ANS POS Tagging:
Parts of Speech tagging is a linguistic activity in Natural Language Processing (NLP)
wherein each word in a document is given a particular part of speech (adverb, adjective,
verb, etc.) or grammatical category.
Through the addition of a layer of syntactic and semantic information to the words, this
procedure makes it easier to comprehend the sentence’s structure and meaning.
NLP END SEM QUESTION BANK
Example of POS Tagging:
Input: “The quick brown fox jumps over the lazy dog.”
POS Tagging:
o “The” is tagged as determiner (DT)
o “quick” is tagged as adjective (JJ)
o “brown” is tagged as adjective (JJ)
o “fox” is tagged as noun (NN)
o “jumps” is tagged as verb (VBZ)
o “over” is tagged as preposition (IN)
o “the” is tagged as determiner (DT)
o “lazy” is tagged as adjective (JJ)
o “dog” is tagged as noun (NN)
We assumed that there are two states in the HMM and each of the state corresponds to
the selection of different biased coin.
Following matrix gives the state transition probabilities
Here,
We can also create an HMM model assuming that there are 3 coins or more.
This way, we can characterize HMM by the following elements
On the other side of coin, the fact is that we need a lot of statistical data to reasonably
estimate such kind of sequences.
However, to simplify the problem, we can apply some mathematical transformations
along with some assumptions.
The use of HMM to do a POS tagging is a special case of Bayesian interference.
Hence, we will start by restating the problem using Bayes’ rule, which says that the
above-mentioned conditional probability is equal to
We can eliminate the denominator in all these cases because we are interested in finding
the sequence C which maximizes the above value.
This will not affect our answer. Now, our problem reduces to finding the sequence C
that maximizes
Even after reducing the problem in the above expression, it would require large amount
of data.
We can make reasonable independence assumptions about the two probabilities in the
above expression to overcome the problem.
NLP END SEM QUESTION BANK