Comparing Estimators
Estimators normally are a trade off between variance and biased predictions.
Mean Squared Error
Let theta_hat be an estimator of a parameter theta. The mean squared error of
theta_hat is denoted and defined by:
MSE(theta_hat) = E[(theta_hat - theta)^2]
If theta_hat is an unbiased estimator of theta its mean squared error is the
variance of theta.
The bias of theta_hat is denoted and defined by:
B(theta_hat) = E[theta_hat] - theta
Relative efficiency
Let theta_hat_1 and theta_hat_2 be two unbiased estimators of a parameter theta.
theta_hat_1 is more efficient that theta_hat_2 if:
Var[theta_hat_1] < Var[theta_hat_2]
The relative efficiency of theta_hat_1 to theta_hat_2 is denoted as:
Eff(theta_hat_1, theta_hat_2) = Var[theta_hat_2] / Var[theta_hat_1]
Cramer-Rao Lower Bound
This is the lower bound on the variance of all unbiased estimators (restrictions exist on distributions this applies to).
Depends on information theory. Proof is based on Cauchy-Schwartz inequality. I don't understand this at all tbh.
Var[tao_hat(theta)] >= ([tao_prime(theta)]^2) / (I_sub_n(theta))
I_sub_n(theta) := E[((a / a*theta) ln f(X_vec;theta))^2]
# in this context, := means "is defined as"
This does not hold if the parameter is in the support, like a uniform distribution.
A great use for this is find an unbiased estimator and check if it's variance achieves the lower bound. This means you have best case (uniformly minimum variance unbiased estimator).
Computational Simplifications
The math to prove this is totally wack:
CRLB = 1/(n/lambda^2)
Weak Law of Large Numbers
Let X_n be a sequence of random variables, iid, from any distribution with
mean u and variance o^2 < inf.
Then X_bar p_-> u. In English "Sample mean always converges in probability to
the true mean of the distribution."