ML Lectures - 33 34
ML Lectures - 33 34
Lecture #33-34
Bias & Variance Tradeoff
Fig 1
Fig 2
10
Overfitting
▪ A natural temptation
o add complexity to the model
o in order to improve accuracy,
o can, in turn, lead to overfitting.
13
▪ The two classes are linearly separable, but noise has caused
one negative example to be mislabeled as positive.
▪ The high-order polynomial on the right overfits the data,
o ignoring the possibility of noise,
o in an attempt to avoid any error on the training set.
▪ The ideal solution often somewhere between the extremes
of linear classifiers and high-order polynomials.
▪ The best choice can be determined experimentally.
14
16
17
Regularization
▪ If λ is taken too large,
o only very simple models are allowed and
o we risk introducing bias.
▪ λ is optimized using cross-validation.
▪ In effect, this add-on parameter provides a warning alert
o to keep high variance in check
o while the original parameters are being optimized.
18