Hi Karl!
(a)
If X1, X2, ... , Xn are iid Gamma(α,θ) then their common density is:
ƒ(x) = 1/(θαΓ(α)) xα-1e-x/θ.
Therefore, the joint likelihood is
L(x) = 1/(θnαΓ(α)n) Πxα-1e-Σx/θ = 1/θnαe-Σx/θ * Γ(α)-nΠxα-1
By The Factorization Theorem, T=Σx is sufficient for θ . This is because L(x) has been factored into the product of:
- g(T(x),θ) = 1/θnαe-Σx/θ which relies on x only through the sufficient statistic T, and
- h(x) = Γ(α)-nΠxα-1 which is a function of the data and known parameters.
To show T is complete we must show that there is no function of T which is an unbiased estimator of 0 except the constant function 0, by the definition of completeness. Note that because the Xi are all iid Gamma(α,θ), their sum T is Gamma(nα,θ). Consider the expectation of f(T), a function of T:
E(f(T)) = ∫ f(t) * 1/(θnαΓ(nα)) tnα-1e-t/θ dt, where t ranges from 0 to infinity, nα and θ positive.
Factoring out terms which do not rely on t yields:
∫ f(t) * tnα-1e-t/θ dt
which is the integral of f(t) and the product of two functions which are positive for all t>0: tnα-1 and e-t/θ.
Therefore, if this expectation is 0, then f(t) must be precisely equal to 0 almost surely, and by the definition of completeness, T(x) is a complete statistic.
Note that because we know T is Gamma(nα,θ), we know E(T) = nαθ, and so T/nα must be an unbiased estimator of θ. Lehmann-Scheffe tells us that unbiased estimators based on complete sufficient statistics must be UMVUE, and so T/nα is UMVU for θ.
(b)
An estimator is efficient if it meets the Cramer-Rao Lower Bound, which is a theoretical lower bound for any unbiased estimator of any parameter, equal to the inverse of the Fisher Information. The Fisher Information has many equivalent definitions, but it's probably easiest in this scenario to define it as negative the expected value of the derivative of the score function with respect to θ.
To answer this question, first we must find the variance of T/nα. We know the distribution of T is Gamma(nα,θ), so its variance is nαθ2, and therefore the variance of T/nα is θ2/nα.
Next, we must find the log-likelihood function for a single observation X:
L(x) = lnΓ(α) - αlnθ + (α-1)ln(x) - x/θ.
The derivative of this with respect to θ is the score function:
l(x) = -α/θ + x/θ2
And therefore, the derivative of the score function is
l`(x) = α/θ2 - 2x/θ3
Negative the expected value of l` is
-E(l`(x)) = -α/θ2 + 2/θ3E(X) = -α/θ2 + 2αθ/θ3 = α/θ2.
Because the Fisher Information scales with the sample size, the Fisher Information for a sample of size n is nα/θ2.
And the inverse of this is θ2/nα, which is precisely the variance of T/nα.
Therefore, our estimator T/nα is consistent.
(c)
An estimator is consistent if it converges in probability to the parameter it is used to estimate.
Notice that T/n is exactly the sample mean, and we have by the Weak law of Large Numbers:
T/n → E(x) = αθ as n → infinity
where "→" symbolizes convergence in probability. Dividing through by α yields the desired result.
(d)
An estimator is asymptotically efficient if it asymptotically achieves the Cramer-Rao Lower Bound. That is,
Var(T/nα) → θ2/nα as n → infinity.
Because we've already shown that Var(T/nα) = θ2/nα, T/nα is asymptotically efficient.