Rao–Blackwell theorem

Rao–Blackwell theorem

In statistics, the Rao–Blackwell theorem is a result which characterizes the transformation of an arbitrarily crude estimator into an estimator that is optimal by the mean-squared-error criterion or any of a variety of similar criteria.

The Rao-Blackwell theorem states that if "g"("X") is any kind of estimator of a parameter θ, then the conditional expectation of "g"("X") given "T"("X"), where "T" is a sufficient statistic, is typically a better estimator of θ, and is never worse. Sometimes one can very easily construct a very crude estimator "g"("X"), and then evaluate that conditional expected value to get an estimator that is in various senses optimal.

The theorem is named after Calyampudi Radhakrishna Rao and David Blackwell. The process of transforming an estimator using the Rao-Blackwell theorem is sometimes called Rao-Blackwellization. The transformed estimator is called the Rao-Blackwell estimator.

Definitions

*An estimator δ("X") is an "observable" random variable (i.e. a statistic) used for estimating some "unobservable" quantity. For example, one may be unable to observe the average height of "all" male students at the University of X, but one may observe the heights of a random sample of 40 of them. The average height of those 40--the "sample average"--may be used as an estimator of the unobservable "population average".

*A sufficient statistic "T"("X") is an "observable" random variable such that the conditional probability distribution of all observable data "X" given "T"("X") does not depend on any of the "unobservable" quantities such as the mean or standard deviation of the whole population from which the data "X" was taken. In the most frequently cited examples, the "unobservable" quantities are parameters that parametrize a known family of probability distributions according to which the data are distributed.: : In other words, a sufficient statistic "T(X)" is a statistic such that the conditional distribution of the data "X", given "T"("X"), does not depend on any unknown parameter.

*A Rao–Blackwell estimator δ1("X") of an unobservable quantity θ is the conditional expected value E(δ("X") | "T"("X")) of some estimator δ("X") given a sufficient statistic "T"("X"). Call δ("X") the "original estimator" and δ1("X") the "improved estimator". It is important that the improved estimator be "observable", i.e. that it not depend on θ. Generally, the conditional expected value of one function of these data given another function of these data "does" depend on θ, but the very definition of sufficiency given above entails that this one does not.

*The "mean squared error" of an estimator is the expected value of the square of its deviation from the unobservable quantity being estimated.

The theorem

Mean-squared-error version

One case of Rao–Blackwell theorem states:

:The mean squared error of the Rao–Blackwell estimator does not exceed that of the original estimator.

In other words

:operatorname{E}((delta_1(X)- heta)^2)leq operatorname{E}((delta(X)- heta)^2).,!

The essential tools of the proof besides the definition above are the law of total expectation and the fact that for any random variable "Y", E("Y"2) cannot be less than [E("Y")] 2. That inequality is a case of Jensen's inequality, although it may also be shown to follow instantly from the frequently mentioned fact that

: 0 leq operatorname{Var}(Y) = operatorname{E}((Y-operatorname{E}(Y))^2) = operatorname{E}(Y^2)-(operatorname{E}(Y))^2.,!

Convex loss generalization

The more general version of the Rao–Blackwell theorem speaks of the "expected loss"

:operatorname{E}(L(delta_1(X)))leq operatorname{E}(L(delta(X))),!

where the "loss function" "L" may be any convex function. For the proof of the more general version, Jensen's inequality cannot be dispensed with.

Properties

The improved estimator is unbiased if and only if the original estimator is unbiased, as may be seen at once by using the law of total expectation. The theorem holds regardless of whether biased or unbiased estimators are used.

The theorem seems very weak: it says only that the Rao-Blackwell estimator is no worse than the original estimator. In practice, however, the improvement is often enormous.

Example

Phone calls arrive at a switchboard according to a Poisson process at an average rate of λ per minute. This rate is not observable, but the numbers "X"1, ..., "X""n" of phone calls that arrived during "n" successive one-minute periods are observed. It is desired to estimate the probability "e"−λ that the next one-minute period passes with no phone calls.

An "extremely" crude estimator of the desired probability is

:delta_0=left{egin{matrix}1 & mbox{if} X_1=0 \0 & mbox{otherwise}end{matrix} ight},

i.e., this estimates this probability to be 1 if no phone calls arrived in the first minute and zero otherwise. Despite the apparent limitations of this estimator, the result given by its Rao–Blackwellization is a very good estimator.

The sum

: S_n = sum_{i=1}^n X_{i} = X_1+cdots+X_n,!

can be readily shown to be a sufficient statistic for λ, i.e., the "conditional" distribution of the data "X"1, ..., "X""n", given this sum, does not depend on λ. Therefore, we find the Rao–Blackwell estimator

:delta_1=operatorname{E}(delta_0|S_n).,!

After doing some algebra we have

:delta_1=left(1-{1 over n} ight)^{S_n}.,!

Since the average number n frac{S_n}{n} of calls arriving during the first "n" minutes is "n"λ, one might not be surprised if this estimator has a fairly high probability (if "n" is big) of being close to

:left(1-{1 over n} ight)^{nlambda}approx e^{-lambda}.

So δ1 is clearly a very much improved estimator of that last quantity. In fact, since "S"n is complete, then δ1 is the unique minimum variance unbiased estimator by the Lehmann-Scheffé theorem.


=Completeness and the Rao–Blackwell process=

Idempotence

In case the sufficient statistic is also a complete statistic, i.e., one which "admits no unbiased estimator of zero", the Rao–Blackwell process is idempotent. Using it to improve the already improved estimator does not do so, but merely returns as its output the same improved estimator.

Lehmann-Scheffé minimum variance

If the conditioning statistic is both complete and sufficient, and the starting estimator is unbiased, then the Rao-Blackwell estimator is the unique "best unbiased estimator".

ee also

* Sufficiency condition of a statistic
* Completeness condition of a statistic
* Basu's theorem - Another result on complete sufficient and ancillary statistics
* The Lehmann–Scheffé theorem


Wikimedia Foundation. 2010.

Игры ⚽ Нужна курсовая?

Look at other dictionaries:

  • Teorema de Rao-Blackwell — En estadística el teorema de Rao Blackwell permite resolver la transformación de un estimador crudo arbitrario en uno optimizado mediante el criterio del error cuadrático medio u otro similar. El teorema de Rao Blackwell establece que si g(X) es… …   Wikipedia Español

  • Calyampudi Radhakrishna Rao — Infobox Scientist name = C. R. Rao |300px caption = Calyampudi Radhakrishna Rao FRS birth date = Birth date and age|1920|9|10|mf=y birth place = Hadagali, State of Mysore, India residence = India nationality = death date = death place = field =… …   Wikipedia

  • David Blackwell — David Harold Blackwell Born April 24, 1919( …   Wikipedia

  • Lehmann–Scheffé theorem — In statistics, the Lehmann–Scheffé theorem, named after Erich Leo Lehmann and Henry Scheffé, states that any unbiased estimator based only on a complete, sufficient statistic is the unique best unbiased estimator of its expected value. The… …   Wikipedia

  • David Blackwell — (* 24. April 1919 in Centralia, Illinois; † 8. Juli 2010 in Berkeley) war ein US amerikanischer Mathematiker, der sich mit Statistik, Spieltheorie und Informationstheorie beschäftigte. David Blackwell, Seattle 1967 Inha …   Deutsch Wikipedia

  • Sufficient statistic — In statistics, a sufficient statistic is a statistic which has the property of sufficiency with respect to a statistical model and its associated unknown parameter, meaning that no other statistic which can be calculated from the same sample… …   Wikipedia

  • List of statistics topics — Please add any Wikipedia articles related to statistics that are not already on this list.The Related changes link in the margin of this page (below search) leads to a list of the most recent changes to the articles listed below. To see the most… …   Wikipedia

  • Sufficiency (statistics) — In statistics, sufficiency is the property possessed by a statistic, with respect to a parameter, when no other statistic which can be calculated from the same sample provides any additional information as to the value of the parameter cite… …   Wikipedia

  • List of mathematics articles (R) — NOTOC R R. A. Fisher Lectureship Rabdology Rabin automaton Rabin signature algorithm Rabinovich Fabrikant equations Rabinowitsch trick Racah polynomials Racah W coefficient Racetrack (game) Racks and quandles Radar chart Rademacher complexity… …   Wikipedia

  • Maximum likelihood — In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of a statistical model. When applied to a data set and given a statistical model, maximum likelihood estimation provides estimates for the model s… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”