Lecture 8 PDF
Lecture 8 PDF
H(Y|X1) = ?
Information Gain
● IG(Humidity) = 0.151
● IG(Humidity) = 0.048
○ It is better to split on humidity rather than wind as
humidity has a higher information gain.
Gini Impurity
Gini Impurity
● Gini = 0.46
References
1. Constructing optimal binary decision trees is NP-complete."
Information Processing Letters 5.1 (1976): 15-17.
2. Entropy: https://www.cs.utexas.edu/~byoung/cs361/lecture32.pdf
3. http://www.cs.cmu.edu/~tom/10701_sp11/slides/DTreesAndOverfittin
g-1-11-2011_final.pdf
4. https://www.ke.tu-darmstadt.de/lehre/archiv/ws0809/mldm/dt.pdf