Decision Trees
Decision Trees
Decision Trees
This process of splitting is then repeated in a top-down, recursive manner until all, or the majority
of records have been classified under specific class labels.
Whether or not all data points are classified as homogenous sets is largely dependent on the
complexity of the decision tree.
Smaller trees are more easily able to attain pure leaf nodes—i.e. data points in a single class.
However, as a tree grows in size, it becomes increasingly
difficult to maintain this purity, and it usually results in too little
data falling within a given subtree.
Entropy and
Information Gini Impurity
Gain
Entropy and Information Gain
• Information gain represents the difference in entropy before and after a split on a given attribute.
• The attribute with the highest information gain will produce the best split as it’s doing the best job at
classifying the training data according to its target classification.
Gain
The feature having the most
The effort is to reduce the information is considered
entropy and maximize the important by the algorithm
information gain. and is used for training the
model.