Bias and Variance of an Estimator
Consider the following estimators of the mean of a distribution. X is an i.i.d. sample from the distribution.
m1=N∑txt (this is the MLE)
m2=2x1+xN
m3=5
Now, draw a sample of size N (say N=3): X={6,1,5}
m1=4,m2=11/5,m3=5
If we consider the means to be random variables, each of them will have a variance.
Say we want to estimate Θ (here, Θ=μ of the distribution from which we are drawing X)
The desirable property of the estimator d of Θ is that the expected value of d must be equal to the quantity we want to estimate i.e. E[d]=Θ. d is then called the unbiased estimator.
The bias of an estimator 'd' is given by:
bΘ(d)=E[d]−Θ
If bΘ(d)=0, d is an unbiased estimator.
Is m2 an unbiased estimator of μ?
m2=2x1+xN E[m2]=E[2x1+xN]=21E[x1+xN]=21E[x1]+21E[xN]
Since E[xt]=μ (by definition), E[m2]=21μ+21μ=μ
Therefore, m2 is an unbiased estimator of the mean μ.
Is m3 an unbiased estimator of μ?
E[m3]=E[5]=5
Clearly, E[m3]−μ=0 iff μ=5. Therefore, m3 is not an unbiased estimator of the mean μ.
The variance of an estimator 'd' is given by
E[(d−E[d])2]
More data leads to lower variance.
m3 has the least variance (it is always 5!). m1 has a lower variance than m2.
The square error of an estimator is given by: E[(d−Θ)2]=(E[d]−Θ)2+E[(d−E[d])2]= Bias2+Variance
Last updated