NPTEL Data Mining Week 3 Assignment Answers 2025

NPTEL Data Mining Week 3 Assignment Answers 2025

1. Decision tree is an algorithm for:

A. Classification
B. Clustering
C. Association rule mining
D. Noise filtering

Answer :- For Answers Click Here 

2. Leaf nodes of a decision tree correspond to:

A. Attributes
B. Classes
C. Data instances
D. None of the above

Answer :- 

3. Non-leaf nodes of a decision tree correspond to:

A. Attributes
B. Classes
C. Data instances
D. None of the above

Answer :- 

4. Which of the following criteria is used to decide which attribute to split next ina decision tree:

A. Support
B. Confidence
C. Entropy
D. Scatter

Answer :- 

5. If we convert a decision tree to a set of logical rules, then:

A. the internal nodes in a branch are connected by AND and the branches by AND
B. the internal nodes in a branch are connected by OR and the branches by OR
C. the internal nodes in a branch are connected by AND and the branches by OR
D. the internal nodes in a branch are connected by OR and the branches by AND

Answer :- 

6. The purpose of pruning a decision tree is:

A. improving training set classification accuracy
B. improving generalization performance
C. dimensionality reduction
D. tree balancing

Answer :- For Answers Click Here 

7. Given the following training set for classification problem into two classes “fraud” and “normal”. There are two attributes A1 and A2 taking values 0 or 1. Splitting on which attribute in the root of a decision tree will lead to highest information gain?

A. A1
B. A2
C. There will be a tie among the attributes
D. Not enough information to decide

Answer :- 

8.

A. 0
B. –(4/10)xlog(4/10)-(6/10)xlog(6/10)
C. –log(4/10)-log(6/10)
D. 1

Answer :- 

9. Given the following training set for classification problem into two classes “fraud” and “normal”. There are two attributes A1 and A2 taking values 0 or 1. Splitting on attribute A1 in the root leads to an entropy reduction of?

A. –log(4/10)-log(6/10)
B. –(4/10)xlog(4/10)-(6/10)xlog(6/10) + (4/7)xlog(4/7) + (3/7)xlog(3/7)
C. –(4/10)xlog(4/10)-(6/10)xlog(6/10) + (4/7)xlog(4/7) + (3/7)xlog(3/7) + 1
D. 1

Answer :- 

10. Given the following training set for classification problem into two classes “fraud” and “normal”. There are two attributes A1 and A2 taking values 0 or 1. Splitting on attribute A2 in the root leads to an entropy reduction of?

A. –log(4/10)-log(6/10)
B. –(4/10)xlog(4/10)-(6/10)xlog(6/10) + (1/4)xlog(1/4) + (3/7)xlog(3/7)
C. –(4/10)xlog(4/10)-(6/10)xlog(6/10) + (3/4)xlog(3/4) + (1/4)xlog(1/4) + (1/6)xlog(1/6) +(5/6)xlog(5/6)
D. 1

Answer :- 

11. Decision trees can be used for:

A. Classification only
B. Regression only
C. Both classification and regression
D. Neither of classification and regression

Answer :- For Answers Click Here 
Scroll to Top