\[Var(\hat\theta) \le Var(\tilde \theta)\]
C.R. Rao; published initial results in Bulletin of the Calcutta Mathematical Society in 1945
Image source: Penn State University
David Blackwell; published the theorem independently of Rao in Annals of Mathematical Statistics in 1947
Image source: Wikipedia
\[E(\hat\theta) = \theta;\]
\[Var(\hat\theta) \leq Var(\tilde\theta).\]
* A more rigorous condition on the sufficient statistic \(U\) known as completeness is required for uniqueness of the MVUE, see the Lehmann-Sheffé theorem. The formal condition of completeness is beyond the scope of an undergraduate math-stat course, but as an example of an inlcomplete sufficient statistic, consider \(Y_{(n)}\) as an estimator of \(\theta\) when the population is \(UNIF(\theta,\theta+1)\). For all problems requiring you to find an MVUE, the sufficient statistic will be complete.
\[E(\hat\theta) = \theta;\]
\[Var(\hat\theta) \leq Var(\tilde\theta).\]
Proof:
\[E(\hat\theta) = E(E(\tilde\theta|U)) = E(\tilde\theta) = \theta\]
\[Var(\hat\theta) = Var(E(\tilde\theta|U))\]
\[\le Var(E(\tilde\theta|U))+E(\underbrace{Var(\tilde\theta|U)}_{\ge 0 \mbox{ always}}) = Var(\tilde\theta)\]
Rao-Blackwell theorem can be proved similarly for a more general case:
\[E(\hat\tau(\theta)) = \tau(\theta);\]
\[Var(\hat\tau(\theta)) \leq Var(\tilde\tau(\theta)).\]
Rao-Blackwellization is the process of improving upon an unbiased estimator by conditioning upon a sufficient statistic. For example:
\[\tilde\theta = \left\{\begin{array} {ll} 1 & Y_1 = 0 \\ 0 & Y_1 > 0\\ \end{array}\right.\]
We want to:
\[\tilde\theta = \left\{\begin{array} {ll} 1 & Y_1 = 0 \\ 0 & Y_1 > 0\\ \end{array}\right.\]
\[E(\tilde\theta) = 1\cdot P(\tilde\theta=1) + 0\cdot P(\tilde\theta = 1)\]
\[ = P(Y_1 =0) = \frac{e^{-\lambda}\lambda^0}{0!} = e^{-\lambda}=\theta\]
With \(U= \sum_{i=1}^n Y_i \sim POI(n\lambda)\):
\[\hat\theta = E(\tilde\theta|U=u)= 1\cdot P(\tilde\theta =1 |U=u)+ 0\cdot P(\tilde\theta=0|U=u)\]
\[=P(Y_1=0|U=u)=\frac{P(Y_1=0, U=u)}{P(U=u)}\]
\[=\frac{P(Y_1=0,\sum_{i=2}^n Y_i=u)}{P(U=u)}, \mbox{ where } Y_1 \perp\!\!\!\perp \sum_{i=2}^n Y_i\sim POI((n-1)\lambda) \]
\[=\frac{\frac{e^{-\lambda}\lambda^0}{0!}\times\frac{e^{-(n-1)\lambda}[(n-1)\lambda]^u}{u!}}{\frac{e^{-n\lambda}(n\lambda)^u}{u!}} = \left(\frac{n-1}{n}\right)^u\]
Suppose \(Y_1,...,Y_n\) are i.i.d. \(\sim EXP(\lambda)\) What is the MVUE of \(\lambda\)?