0% found this document useful (0 votes)
40 views4 pages

NPTEL NLP Assignment 10.bin

This document contains an assignment for a Natural Language Processing course consisting of 10 multiple-choice questions related to entity linking, keyphraseness, and information extraction techniques. Each question includes answer options, the correct answer, and a brief solution reference. The assignment aims to assess understanding of various concepts in NLP.

Uploaded by

d36078067
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
40 views4 pages

NPTEL NLP Assignment 10.bin

This document contains an assignment for a Natural Language Processing course consisting of 10 multiple-choice questions related to entity linking, keyphraseness, and information extraction techniques. Each question includes answer options, the correct answer, and a brief solution reference. The assignment aims to assess understanding of various concepts in NLP.

Uploaded by

d36078067
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

Natural Language Processing

Assignment 10
Type of Question: MCQ
Number of Questions: 10 Total Marks:(10×1)= 10

=====================================================

Question 1.

Different phases of entity linking are -


A) Candidate Selection -> Reference Disambiguation
B) Reference Disambiguation -> Candidate Selection -> Mention Identify
C) Mention Identify -> Candidate Selection -> Reference Disambiguation
D) All of the above
Answer: A, C

Solution: Theory. Slide 7; Lecture 1, Week 10 (Entity Linking 1).

===================================================

Question 2.

The text span s=“river” occurs in 700 different Wikipedia articles.

c1 223
c2 161
c3 78
c4 31
No Link 207

Calculate the keyphraseness of “Sea”.


A) 0.232
B) 0.886
C) 0.688
D) 0.704
Answer D
Solution: CF(si) / CF(s) = 223 +161+78+31 / 700 = 493 / 700 = 0.704

===================================================
Question 3.
What is the commonness of (s, c3) in the above question?
A) 0.765
B) 0.389
C) 0.158
D) 0.910

Answer C

Solution: 78/(223+161+78+31) = 78/493 = 0.158

===================================================

Question 4.

Relevant feature/s for a supervised model for predicting the topics to be linked
is/are:
A) Disambiguation Confidence
B) Relatedness
C) Link Probability
D) All of the above
Answer: D

Solution: Theory. Entity Linking Lecture II [Lecture Video 47]

===================================================

Question 5.

Which of the following problem exists in bootstrapping technique for Information


extraction are:

A) Sensitiveness towards the seed set


B) High precision
C) Less manual intervention
D) All of the above
Answer: A

Solution: Theory

===================================================
Question 6.

Which of the following is an advantage of unsupervised relation extraction:

A) Can work efficiently with small amount of hand-labeled data


B) Not easily generalizable to different relations
C) Need no training data.
D) Always perform better than supervised techniques.

Answer: C

Solution: Theory

===================================================

Question 7.

Which of the following is not a Hearst’s Lexico Syntactic Patterns for automatic
acquisition of hyponyms -
A) X or other Y
B) X and other Y
C) Y including X
D) X but not Y
Answer: D

Solution: Theory. Lecture Video 48 : Information Extraction - Introduction, Week


10, Lecture 3 - Slide 15/18.

===================================================

Question 8.

Advantage of Distant supervision over bootstrapping method

A) Need more data


B) Less human effort
C) Can handle noisy data better
D) No Advantage
Answer: C

Solution: Theory

===================================================
Question 9.

Consider a dataset with a very low number of relations - all of which are very
important. For a relation extraction task on that dataset, which of the following is the
most useful metric
A) Precision
B) Recall
C) Accuracy
D) F1-Score
Answer: B
Solution: Theory

===================================================

Question 10.

What is KeyPhraseness (wikipedia)?


A) Number of articles that mention a key phrase divided by the number of
wikipedia articles containing it.
B) Number of Wikipedia articles that use it as an anchor, divided by the
number of articles that mention it at all.
C) Number of articles that mention a key phrase times by the number of
wikipedia articles containing it.
D) Number of Wikipedia articles containing the key phrases times by number
of articles mentioning it.
Answer: B

Solution: Theory.

===================================================

You might also like