0% found this document useful (0 votes)
119 views4 pages

Deep Learning - IIT Ropar - Unit 12 - Week 9

The document details the Week 9 assignment for the Deep Learning course at IIT Ropar, including questions related to word representation techniques and models such as PMI, one-hot encoding, and skip-gram. It outlines the correct answers and scores for various questions, highlighting concepts like cosine similarity, hierarchical softmax, and negative sampling. The assignment was submitted on September 25, 2024, and the due date had already passed.

Uploaded by

jaiakaash72
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
119 views4 pages

Deep Learning - IIT Ropar - Unit 12 - Week 9

The document details the Week 9 assignment for the Deep Learning course at IIT Ropar, including questions related to word representation techniques and models such as PMI, one-hot encoding, and skip-gram. It outlines the correct answers and scores for various questions, highlighting concepts like cosine similarity, hierarchical softmax, and negative sampling. The assignment was submitted on September 25, 2024, and the due date had already passed.

Uploaded by

jaiakaash72
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

10/31/24, 8:52 PM Deep Learning - IIT Ropar - - Unit 12 - Week 9

(https://swayam.gov.in) (https://swayam.gov.in/nc_details/NPTEL)

[email protected]

NPTEL (https://swayam.gov.in/explorer?ncCode=NPTEL) » Deep Learning - IIT Ropar (course)

Course Week 9 : Assignment 9


outline The due date for submitting this assignment has passed.
Due on 2024-09-25, 23:59 IST.
About
NPTEL ()
Assignment submitted on 2024-09-25, 18:34 IST
How does an 1) Let X be the co-occurrence matrix such that the (i, j) -th entry of X captures the 1 point
NPTEL PMI between the i-th and j -th word in the corpus. Every row of X corresponds to the
online representation of the i-th word in the corpus. Suppose each row of X is normalized (i.e., the L2
course norm of each row is 1) then the (i, j) -th entry of XX T captures the:
work? ()

Week 1 () PMI between word i and word j

Week 2 () Euclidean distance between word i and word j

Probability that word i


Week 3 ()

Cosine similarity between word i


week 4 ()
Yes, the answer is correct.
Score: 1
Week 5 ()
Accepted Answers:
Cosine similarity between word i
Week 6 ()

2) Consider the following corpus: "human machine interface for computer applications. 1 point
Week 7 ()
user opinion of computer system response time. user interface management system. system
engineering for improved response time". What is the size of the vocabulary of the above
Week 8 ()
corpus?

Week 9 () 13
14
One-hot
representation 15
s of words 16

https://onlinecourses.nptel.ac.in/noc24_cs114/unit?unit=115&assessment=297 1/4
10/31/24, 8:52 PM Deep Learning - IIT Ropar - - Unit 12 - Week 9

(unit? Yes, the answer is correct.


unit=115&less Score: 1
on=116) Accepted Answers:
15
Distributed
Representatio 3) You are given the one hot representation of two words below:
ns of words CAR= [1, 0, 0, 0, 0] , BUS= [0, 0, 0, 1, 0]

(unit?
unit=115&less What is the Euclidean distance between CAR and BUS?
on=117)
1.414
SVD for
learning word Yes, the answer is correct.
representation Score: 1
s (unit? Accepted Answers:
unit=115&less (Type: Range) 1.40,1.42
on=118)
1 point
SVD for
learning word 4) Let count(w, c) be the number of times the words w and c appear together in the 1 point
representation corpus (i.e., occur within a window of few words around each other). Further, let count(w) and
s (Contd.) count(c) be the total number of times the word w and c appear in the corpus respectively and
(unit? let N be the total number of words in the corpus. The PMI between w and c is then given by:
unit=115&less
on=119)
count(w,c)∗count(w)
log
Continuous N ∗count(c)

bag of words
count(w,c)∗count(c)
model (unit? log
N ∗count(w)
unit=115&less
on=120)
count(w,c)∗N
log
count(w)∗count(c)
Skip-gram
model (unit? Yes, the answer is correct.
unit=115&less Score: 1
on=121) Accepted Answers:
count(w,c)∗N
log
Skip-gram count(w)∗count(c)

model (Contd.)
(unit? 5) Consider a skip-gram model trained using hierarchical softmax for analyzing 1 point
unit=115&less
scientific literature. We observe that the word embeddings for 'Neuron' and 'Brain' are highly
on=122)
similar. Similarly, the embeddings for 'Synapse' and 'Brain' also show high similarity. Which of the
Contrastive following statements can be inferred?
estimation
(unit? 'Neuron' and 'Brain' frequently appear in similar contexts
unit=115&less The model's learned representations will indicate a high similarity between 'Neuron' and
on=123) 'Synapse'
Hierarchical The model's learned representations will not show a high similarity between 'Neuron' and
softmax (unit? 'Synapse'
unit=115&less
According to the model's learned representations, 'Neuron' and 'Brain' have a low cosine
on=124)
similarity
GloVe
Yes, the answer is correct.
representation Score: 1
s (unit?
Accepted Answers:
unit=115&less
'Neuron' and 'Brain' frequently appear in similar contexts
on=125)

https://onlinecourses.nptel.ac.in/noc24_cs114/unit?unit=115&assessment=297 2/4
10/31/24, 8:52 PM Deep Learning - IIT Ropar - - Unit 12 - Week 9

Evaluating The model's learned representations will indicate a high similarity between 'Neuron' and
word 'Synapse'
representation
6) Which of the following is a disadvantage of one hot encoding? 1 point
s (unit?
unit=115&less
It requires a large amount of memory to store the vectors
on=126)
It can result in a high-dimensional sparse representation
Relation
It cannot capture the semantic similarity between words
between SVD
and Word2Vec All of the above
(unit?
Yes, the answer is correct.
unit=115&less Score: 1
on=127)
Accepted Answers:
Lecture All of the above
Material for
Week 9 (unit? 7) Which of the following is true about the input representation in the CBOW model? 1 point
unit=115&less
on=128) Each word is represented as a one-hot vector

Week 9
Each word is represented as a continuous vector
Feedback Each word is represented as a sequence of one-hot vectors
Form: Deep Each word is represented as a sequence of continuous vectors
Learning - IIT
Ropar (unit? Yes, the answer is correct.
unit=115&less
Score: 1
on=192) Accepted Answers:
Each word is represented as a one-hot vector
Quiz: Week 9
: Assignment
8) What is the role of the softmax function in the skip-gram method? 1 point
9
(assessment?
To calculate the dot product between the target word and the context words
name=297)
To transform the dot product into a probability distribution
week 10 () To calculate the distance between the target word and the context words
To adjust the weights of the neural network during training
Week 11 ()
Yes, the answer is correct.
Score: 1
Week 12 ()
Accepted Answers:
To transform the dot product into a probability distribution
Download
Videos ()
9) We add incorrect pairs into our corpus to maximize the probability of words that 1 point
occur in the same context and minimize the probability of words that occur in different contexts.
Books ()
This technique is called-

Text Hierarchical softmax


Transcripts Contrastive estimation
()
Negative sampling

Problem Glove representations


Solving Yes, the answer is correct.
Session - Score: 1
July 2024 () Accepted Answers:
Negative sampling

https://onlinecourses.nptel.ac.in/noc24_cs114/unit?unit=115&assessment=297 3/4
10/31/24, 8:52 PM Deep Learning - IIT Ropar - - Unit 12 - Week 9

10) How does Hierarchical Softmax reduce the computational complexity of computing 1 point
the softmax function?

It replaces the softmax function with a linear function


It uses a binary tree to approximate the softmax function
It uses a heuristic to compute the softmax function faster
It does not reduce the computational complexity of computing the softmax function
Yes, the answer is correct.
Score: 1
Accepted Answers:
It uses a binary tree to approximate the softmax function

https://onlinecourses.nptel.ac.in/noc24_cs114/unit?unit=115&assessment=297 4/4

You might also like