This document discusses various ensemble machine learning algorithms including bagging, boosting, and random forests. It explains that ensemble approaches average the predictions of multiple models to improve performance over a single model. Bagging trains models on random subsets of data and averages predictions. Random forests build on bagging by using random subsets of features to de-correlate trees. Boosting iteratively trains weak learners on weighted versions of the data that focus on previously misclassified examples. The document provides examples and comparisons of these ensemble techniques.