Decision Tree For Classification (ID3 Information Gain Entropy)
Decision Tree For Classification (ID3 Information Gain Entropy)
Decision tree learning employs a recursive splitting strategy using greedy search
to identify the optimal split points within the tree. This process of splitting is
repeated from top to bottom until all records gets classified under a specific class
label or value. The complexity of the tree decides on well a decision tree can
There are different types of decision tree models to choose from based on there
learning and node splitting technique. ID3 (Iterative Dichotomiser 3), C4.5, CART
and Chi-Square are popular ones.
As node splitting is the key concept/step in decision tree algorithm let’s look at it in
detail. There are multiple ways to split a node and it can be broadly divided into
two categories based the type of target variable.
1. For each split, individually calculate the variance of each child node
(feature on which you want try split)
1. For each split, individually calculate the entropy of each child node (feature
on which you want try split)
Gini Impurity:
2. Calculate the Gini Impurity of each split as the weighted average Gini
Impurity of child nodes
REFERENCES:
https://www.analyticsvidhya.com/blog/2020/06/4-ways-split-decision-tree/
https://www.ibm.com/topics/decision-trees#:~:text=data mining
solutions-,Decision Trees,internal nodes and leaf nodes.
https://www.analyticsvidhya.com/blog/2021/08/decision-tree-algorithm/#Pruning