NPTEL Natural Language Processing Week 7 Assignment Answers 2025

NPTEL Natural Language Processing Week 7 Assignment Answers 2025

1. Suppose you have a raw text corpus and you compute word co-occurrence matri› from there. Which of the following algorithms) can you utilize to learn word representations? (Choose all that apply)

a. CBOW
b. SVD
c. PCA
d. GloVe

Answer :- For Answers Click Here 

2. What is the method for solving word analogy questions like, given A, B and D, find C such that A:B::C:D, using word vectors?

a. Vc= Va + (Vb – Vd), then use cosine similarity to find the closest word of Vc.
b. Vc = Va + (Vd – Vb) then do dictionary lookup for Vc
C. Vc = Vd + (Va – Vb) then use cosine similarity to find the closest word of Vc.
d. Vc= Vd + (Va – Vb) then do dictionary lookup for Vc.
e. None of the above

Answer :- 

3. What is the value of PMI(W1, W2) for C(W1) = 100, C(W2) = 2500, C(W1, W2) = 320, N =
50000? N: Total number of documents.
C(wi): Number of documents, wi has appeared in.
C(Wi, wj): Number of documents where both the words have appeared in.
Note: Use base 2 in logarithm.

a. 4
b. 5
C. 6
d. 5.64

Answer :- 

4. Given two binary word vectors wi and wz as follows:
W1 = [1010011010]
W2 = [0011111100]
Compute the Dice and Jaccard similarity between them.

a. 6/11, 3/8
b. 10/11, 5/6
C. 4/9, 217
d. 5/9, 5/8

Answer :- 

5. Consider two probability distributions for two words be p and q. Compute their
similarity scores with KL-divergence.
p = [0.20, 0.75, 0.50]
q = [0.90, 0.10, 0.25)
Note: Use base 2 in logarithm.

a. 4.704, 1,720
b. 1.692, 0.553
c. 2.246, 1.412
d. 3.213, 2.426

Answer :- 

6.

Answer :- For Answers Click Here 

7. Which of the following types of relations can be captured by word2vec (CBOW or
Skipgram)?

  1. Analogy (A:B::C:?)
  2. Antonymy
  3. Polysemy
  4. All of the above
Answer :- 
Scroll to Top