lecture03b_overfitting
lecture03b_overfitting
Martin Jaggi
Last updated on: September 24, 2024
credits to Mohammad Emtiyaz Khan & Rüdiger Urbanke
Motivation
Models can be too limited or they can be too rich. In the
first case we cannot find a function that is a good fit for the
data in our model. We then say that we underfit. In the
second case we have such a rich model family that we do not
just fit the underlying function but we in fact fit the noise
in the data as well. We then talk about an overfit. Both of
these phenomena are undesirable. This discussion is made
more difficult since all we have is data and so we do not know
a priori what part is the underlying signal and what part is
noise.
1 M =0
t
−1
0 x 1
1 M =0 1 M =1
t t
0 0
−1 −1
0 x 1 0 x 1
1 M =3 1 M =9
t t
0 0
−1 −1
0 x 1 0 x 1
1 N = 15 1 N = 100
t t
0 0
−1 −1
0 x 1 0 x 1
Additional Materials
Read about overfitting in the paper by Pedro Domingos (Sections 3 and 5
of “A few useful things to know about machine learning”).