10

I would like to know if the statistic $$T(X_1,\ldots,X_n)=\frac{\sum_{i=1}^n (X_i-\bar{X}_n)^2}{n-1}$$ is complete for $\sigma^2$ in a $N(\mu,\sigma^2)$ setting.

Does this depend on whether $\mu$ is previously known or not? If $T$ is complete for $\sigma^2$, then by Lehmann-Scheffé it is UMVUE. But if $\mu$ were known, we could have considered $$W(X_1,\ldots,X_n)=\frac{\sum_{i=1}^n (X_i-\mu)^2}{n},$$whose variance equals the Cramer-Rao bound $2\sigma^4/n$, and is strictly less than $2\sigma^4/(n-1)=\text{Var}[T]$, so $T$ could not be UMVUE.

user39756
  • 381
  • 3
  • 11

3 Answers3

5

I think I solved my own question. Comments about this answer and new answers are welcome.

If $x_1,\ldots,x_n$ are observations in a $N(\mu,\sigma^2)$ population and $\mu$ is unknown, then $$f(x_1,\ldots,x_n|\mu,\sigma^2)=\left(\frac{1}{\sqrt{2\pi\sigma^2}}\right)^ne^{-\frac{n\mu^2}{2\sigma^2}}e^{\frac{\mu}{\sigma^2}\sum_{i=1}^nx_i-\frac{1}{2\sigma^2}\sum_{i=1}^n x_i^2}$$ (this shows that the normal family is a exponential family). As the image of the map $$(\mu,\sigma^2)\in \mathbb{R}\times\mathbb{R}^+\mapsto (\frac{\mu}{\sigma^2},-\frac{1}{2\sigma^2})$$ contains an open set of $\mathbb{R}^2$, by a theorem (for instance, see page 6 here), the statistic $U=(\sum_{i=1}^n X_i,\sum_{i=1}^n X_i^2)$ is sufficient and complete for $(\mu,\sigma^2)$. As $T$ is a function of $U$ and is centered for $\sigma^2$, by Lehmann-Scheffé $T$ is UMVUE for $\sigma^2$.

Now, if $\mu=\mu_0$ is known, $\mu$ does not belong to the parametric space anymore, therefore the "new" density function is $$f(x_1,\ldots,x_n|\sigma^2)=\left(\frac{1}{\sqrt{2\pi\sigma^2}}\right)^ne^{-\frac{1}{2\sigma^2}\sum_{i=1}^n(x_i-\mu_0)^2}$$(we have a new exponential family). As the image of the map $$\sigma^2\in\mathbb{R}^+\mapsto -\frac{1}{2\sigma^2}$$ contains an open subset of $\mathbb{R}$, our statistic $W$ is sufficient and complete for $\sigma^2$. Since it is in addition centered, $W$ is UMVUE for $\sigma^2$ by Lehmann-Scheffé.

user39756
  • 381
  • 3
  • 11
1

In your statistic $T(X_{1},\ldots,X_{n})$, $\bar{X}$ is used as the estimate of $\mu$. If you know the true value of $\mu$, then the estimator of the variance $W(X_{1},\ldots,X_{n})$ is preferable. $W$ is unbiased and has a lower variance than $T$. Thus, in the setting where $\mu$ is known, use $W$.

epp
  • 2,372
  • 2
  • 12
  • 31
  • I know that $W$ is preferable, but I would like to understand why $T$ is not complete when $\mu$ is known. – user39756 Dec 11 '16 at 16:09
  • 1
    I have been searching the net refreshing my memory on the Lehmann-Scheffe Theorem and Rao-Blackwell. I agree with the OP that the answer might be that T is not a complete sufficient statistic when mu is known. I think in may be that the parameter space changes. – Michael R. Chernick Dec 11 '16 at 16:44
0

W(X1,…,Xn) is not an unbiased estimator, so it is not a UMVUE. T actually is the UMVUE because it is a complete sufficient statistic and also an unbiased estimator for $\sigma^2$. Hence according to Lehmann-Scheffé, it is the UMVUE.