NPTEL Introduction to Machine Learning Week 7 Assignment Answers 2025

NPTEL Introduction to Machine Learning Week 7 Assignment Answers 2025

1. Which of the following statement(s) regarding the evaluation of Machine Learning models is/are true?

  • A model with a lower training loss will perform better on a validation dataset.
  • A model with a higher training accuracy will perform better on a validation dataset.
  • The train and validation datasets can be drawn from different distributions
  • The train and validation datasets must accurately represent the real distribution of data
Answer :- For Answers Click Here 

2. Suppose we have a classification dataset comprising of 2 classes A and B with 200 and 40 samples respectively. Suppose we use stratified sampling to split the data into train and test sets. Which of the following train-test splits would be appropriate?

  • Train-{A:50samples,B:10samples},Test-{A:150samples,B:30samples}
  • Train-{A:50samples,B:30samples},Test- {A:150samples,B:10samples}
  • Train- {A:150samples,B:30samples},Test- {A:50samples,B:10samples}
  • Train- {A:150samples,B:10samples},Test- {A:50samples,B:30samples}
Answer :- 

3. Suppose we are performing cross-validation on a multiclass classification dataset with N data points. Which of the following statement(s) is/are correct?

  • In k-fold cross-validation, we train k−1 different models and evaluate them on the same test set
  • In k-fold cross-validation, we train k different models and evaluate them on different test sets
  • In k-fold cross-validation, each fold should have a class-wise proportion similar to the given dataset.
  • In LOOCV (Leave-One-Out Cross Validation), we train N different models, using N−1 data points for training each model
Answer :- 

4. (Qns 4 to 7) For a binary classification problem we train classifiers and evaluate them to obtain confusion matrices in the following format:

Which of the following classifiers should be chosen to maximize the recall?

Answer :- 

5. For the confusion matrices described in Q4, which of the following classifiers should be chosen to minimize the False Positive Rate?

Answer :- 

6. For the confusion matrices described in Q4, which of the following classifiers should be chosen to maximize the precision?

Answer :- For Answers Click Here 

7. For the confusion matrices described in Q4, which of the following classifiers should be chosen to maximize the F1-score?

Answer :- 

8. Which of the following statement(s) regarding boosting is/are correct?

  • Boosting is an example of an ensemble method
  • Boosting assigns equal weights to the predictions of all the weak classifiers
  • Boosting may assign unequal weights to the predictions of all the weak classifiers
  • The individual classifiers in boosting can be trained parallelly
  • The individual classifiers in boosting cannot be trained parallelly
Answer :- 

9. Which of the following statement(s) about bagging is/are correct?

  • Bagging is an example of an ensemble method
  • The individual classifiers in bagging can be trained in parallel
  • Training sets are constructed from the original dataset by sampling with replacement
  • Training sets are constructed from the original dataset by sampling without replacement
  • Bagging increases the variance of an unstable classifier.
Answer :- 

10. Which of the following statement(s) about ensemble methods is/are correct?

  • Ensemble aggregation methods like bagging aim to reduce overfitting and variance
  • Committee machines may consist of different types of classifiers
  • Weak learners are models that perform slightly worse than random guessing
  • Stacking involves training multiple models and stacking their predictions into new training data
Answer :- For Answers Click Here 
Scroll to Top