Watch a model go from underfitting to just right to badly overfitting โ and discover why a model that's too clever on training data fails completely on new data!
Model is too simple โ misses the real pattern. High training error AND high test error. "The model didn't learn."
Model captures the true pattern without memorising noise. Low training error AND low test error. The goal!
Model memorises training data โ even the noise. Very low training error but HIGH test error. "It cheated."
Penalise model complexity. Forces simpler solutions. L1, L2, dropout โ all ways to prevent overfitting.
You mastered overfitting, learning curves, and regularisation!
When a model learns the training data too well โ including its noise โ it fails to generalise to new data.
A model that hasn't learned the pattern at all. High bias. A straight line through clearly curved data.
Keep some data hidden from training. Test error on this holdout reveals overfitting before deployment.
Add a penalty term to the loss function that grows with coefficient size. Forces simpler, smoother models.
High bias = underfit. High variance = overfit. The sweet spot balances both to minimise total error.
Split data into K folds. Train on K-1, test on 1. Repeat K times. More reliable than a single train/test split.