Bias and Variance

Bias and Variance

A good learner classifier should have a good generalisation error.

  • Generalisation: how well do we do on unseen data as opposed to the training data

The problems in the Machine Learning can be over-constrained and under-constrained.

  • Over-constrained: We have conflicting data to deal with. There are more equations than variables. In this case, the learner has insufficient flexibility to correctly predict all the training data. To solve this problem, we can minimise an error function, which means that we find a machine that explained the training data as best it can.
  • Under-constrained: There are many possible solutions that are consistent with the data . Need to choose a plausible solution.

Bias: the generalisation performance of the mean machine.

which $\hat{f}_{m}(\boldsymbol{x})$ is the mean predictor(machine) value. And the bias is defined as

(可以看成最终训练出的分类器在训练集上进行预测,得到的所有值的平均值与每一个原target计算error)

Variance: measures the expected variation from the average machine due to the fluctuations caused by the using a finite training set.

(训练出来的分类器,对每个训练集的输入进行预测,计算这些值的分布的方差)

Decomposition

The formulas of bias and variance are already defined above. Here we are going to show the decomposition.

The expected generalisation(平均泛化误差) is written as

The second term will vanish sicne the $\hat{f}_{m}(\boldsymbol{x})=\mathbb{E}_{\mathcal{D}}\left[\hat{f}\left(\boldsymbol{x} | \boldsymbol{w}_{\mathcal{D}}\right)\right]$. Finally we can rewrite the generalisation formula as

Bias-Variance Dilemma

The composition mentioned above encodes how sensitive the machine is to the data.

The dilemma arises because a simple machine will typically have a large bias, but small variance, while a complicated machine will have a small bias but large variance.

Reference

  1. Adam slide01 COMP6208
  2. Additional reading Bishop PRML Chapter 3.2