NPTEL Deep Learning – IIT Ropar Week 12 Assignment Answers 2025

NPTEL Deep Learning – IIT Ropar Week 12 Assignment Answers 2025

1. What is the primary purpose of the attention mechanism in neural networks?

  • To reduce the size of the input data
  • To increase the complexity of the model
  • To eliminate the need for recurrent connections
  • To focus on specific parts of the input sequence
Answer :- For Answers Click Here 

2. Which of the following are the benefits of using attention mechanisms in neural networks?

  • Improved handling of long-range dependencies
  • Enhanced interpretability of model predictions
  • Ability to handle variable-length input sequences
  • Reduction in model complexity
Answer :- 

3. If we make the vocabulary for an encoder-decoder model using the given sentence. What will be the size of our vocabulary?
Sentence: Attention mechanisms dynamically identify critical input components, enhancing contextual understanding and boosting performance

  • 13
  • 14
  • 15
  • 16
Answer :- 

4. We are performing the task of Machine Translation using an encoder-decoder model. Choose the equation representing the Encoder model.

  • s0=CNN(xi)
  • s0=RNN(st−1,e(y^t−1))
  • s0=RNN(xit)
  • s0=RNN(ht−1,xit)
Answer :- 

5. Which of the following attention mechanisms is most commonly used in the Transformer model architecture?

  • Additive attention
  • Dot product attention
  • Multiplicative attention
  • None of the above
Answer :-  For Answers Click Here 

6. Which of the following is NOT a component of the attention mechanism?

  • Decoder
  • Key
  • Value
  • Query
  • Encoder
Answer :- 

7. In a hierarchical attention network, what are the two primary levels of attention?

  • Character-level and word-level
  • Word-level and sentence-level
  • Sentence-level and document-level
  • Paragraph-level and document-level
Answer :- 

8. Which of the following are the advantages of using attention mechanisms in encoderdecoder models?

  • Reduced computational complexity
  • Ability to handle variable-length input sequences
  • Improved gradient flow during training
  • Automatic feature selection
  • Reduced memory requirements
Answer :- 

9. In the encoder-decoder architecture with attention, where is the context vector typically computed?

  • In the encoder
  • In the decoder
  • Between the encoder and decoder
  • After the decoder
Answer :- 

10. Which of the following output functions is most commonly used in the decoder of an encoder-decoder model for translation tasks?

  • Softmax
  • Sigmoid
  • ReLU
  • Tanh
Answer :-  For Answers Click Here 
Scroll to Top