Sample Distributions
If you take many random samples X1, X2, X3, ..., Xn
and create a histogram,
you end up with a sample distribution.
All variables are assumed to be independent and identically distributed (iid). Random sample = iid.
Estimator:
theta-hat = a random variable
If X is an estimator and E[X] = mue then X is an unbiased estimator.
Estimate:
theta-hat = an observation or number x
The expected value of a sample mean is the same as the expected value of the distribution itself (one random variable) because expected value of the whole sample is always an unbiased estimator in any distribution.
Unbiased is a trade off with small variance in a sample distribution. Having both is the goal.
V[sample mean] = o^2/n
Distributions of sums
You can use the moment generating function to uniquely identify distributions, this is useful when finding the distribution of a new random variable that is produced by taking the sum of all random variables in a sample distribution.
Not all random variable samples sum to a nice, named, distribution.
X1, X2, ..., Xn iid
Y = sum(i=1 to n) Xi
MY(t) = [MX(t)]^n
Examples:
Discrete
Bernoulli: M(t) = 1 - p + pe^t
sum of bernoulli sample will have binomial distribution:
Binomial: M(t) = (pe^t + 1 - p)^n
Continuous
Exponential: M(t) = lambda / (lambda - t)
sum of exponential sample will have gama distribution:
Gamma: M(t) = (beta/(beta - t))^a