Skip to content

Latest commit

 

History

History
44 lines (28 loc) · 3 KB

File metadata and controls

44 lines (28 loc) · 3 KB

Overfitting

Overfitting is an extremely important concept in machine learning, and it is very important to get it right!

Consider the following simple problem of approximating 5 dots (represented by x on the graphs below):

Linear model, 2 parameters Non-linear model, 7 parameters
Training error = 5.3 Training error = 0
Validation error = 5.1 Validation error = 20
  • On the left, we see a good straight line approximation. Because the number of parameters is adequate, the model gets the idea behind point distribution right.
  • On the right, the model is too powerful. Because we only have 5 points and model has 7 parameters, it can adjust in such a way as to pass through all points, making training error to be 0. However, it prevents the model from understanding the correct pattern behind data, thus validation error is very high.

Thus it is very important to strike a correct balance between richness of the model (number of parameters) and the number of training samples.

Why overfitting occurs

  • Not enough training data
  • Too powerful model
  • Too much noise in input data

How to detect overfitting

As you can see from the graph above, overfitting can be detected by very low training error, and high validation error. Normally during training we will see both training and validation errors starting to decrease, and then at some point validaton error might stop decreasing and start rising. This will be a sign of overfitting, and the indicator that we should probably stop training at this point (or at least make a snapshot of the model).

How to prevent overfitting

If you can see that overfitting occurs, you can do one of the following:

  • Increase the amount of training data
  • Decrease the complexity of the model
  • Use some regularization technique, such as Dropout, which we will consider later.

Overfitting and Bias-Variance Tradeoff

Overfitting is actually a case of more generic problem in statistics, called Bias-Variance Tradeoff. If we consider possible sources of error in our model, we can see two types of errors:

  • Bias errors are caused by our algorithm not being able to capture the relationship between training data correctly. It can result from the fact that our model is not powerful enough (underfitting).
  • Variance errors, which are caused by the model approximating noise in the input data instead of meaningful relationship (overfitting).

During training, bias error decreases (as our model learns to approximate the data), and variance error increases. It is important to stop training - either manually (when we detect overfitting) or automatically (by introducing regularization) to prevent overfitting.