Machine learning
bias/variance trade-off
Jan 15, 2022     3 minutes read

1. Why an article about such a fairly straightforward subject as bias/variance trade-off?

IMHO, the subject is straightforward only superficially, so it may leave us with a false feeling of comprehension. From my statistician’s point of view, there is a little more to think about.

2. When do we face this dillema?

When we assess the quality of our model.

To do that, we use validation methods and metrics. One of the basic rules is that we calculate the metrics on the test set. A commonly used metric for regression problems is MSE (Mean Squared Error), which corresponds to variance of the error (for many algorithms error by definition has mean 0).

MSE consists of 3 components:

In short:

$$ Error(x_0) = Irreducible Error + Bias^2 + Variance $$

from The Elements of Statistical Learning, 2nd Edition by Hastie, Tibshirani, Friedman, eq. 7.9, which slowly becomes my all-time ml/statistics textbook.

3. Connotations

//TODO: connotations with learning curves