Week 2 Quiz
Week 2 Quiz
Week 2 Quiz
12 试题
1
point
1。
Let w1, w2, and w3 represent three words in the dictionary of an
inverted index. Suppose we have the following document
frequency distribution:
w1 1000
w2 100
w3 10
w1
w2
w3
1
point
2。
Assume we have the same scenario as in Question 1. If we enter
the query Q= “w1 w2” then the minimum possible number of
accumulators needed to score all the matching documents is:
1100
1000
10
100
1
point
3。
The gamma code for the term frequency of a certain document is
1110010. What is the term frequency of the document?
12
11
10
1
point
4。
When using an inverted index for scoring documents for queries, a
shorter query always uses fewer score accumulators than a longer
query.
True
False
1
point
5。
What is the advantage of tokenization (normalize and stemming)
before index?
1
point
6。
What can't an inverted index alone do for fast search?
1
point
7。
If Zipf's law does not hold, will an inverted index be much faster or
slower?
Faster
Slower
1
point
8。
In BM25, the TF after transformation has upper bound
k +1
1
point
9。
Which of the following are weighing heuristics for the vector space
model?
IDF weighting
1
point
10。
Which of the following integer compression has equal-length
coding?
Binary
γ -code
Unary
1
point
11。
Consider the following retrieval formula:
TF
IDF
1
point
12。
Suppose we compute the term vector for a baseball sports news
article in a collection of general news articles using TF-IDF
weighting. Which of the following words do you expect to have the
highest weight in this case?
baseball
computer
the
我(伟臣 沈)了解提交不是我自己完成的作业 将永远不会通过
此课程或导致我的 Coursera 帐号被关闭。
了解荣誉准则的更多信息
提交测试