Module 10- Part 2- Boosting models
Module 10- Part 2- Boosting models
Boosting Models
AdaBoost, GBM, XGBoost
Supervised Unsupervised
Dimensionality
Regression Classification Clustering
Reduction
Part II
Pros and Cons
• AdaBoost is old: AdaBoost is a popular boosting technique introduced by Yoav Freund and Robert
Schapire in 1996.
• Learning rate shrinks the contribution of each tree. There is a trade-off between learning rate and
number of trees. Learning rate slows down the process even further, allowing for more and different
shaped trees to attack the residuals.
• Aggregation is done by adding the first tree predictions and a scaled (shrunk) version of the following
trees.
Pros:
Cons:
• XGBoost is more difficult to understand, visualize and to tune compared to AdaBoost and
random forests. There is a multitude of hyperparameters that can be tuned to increase
performance.