Bayes estimator
From Wikipedia, the free encyclopedia
In decision theory and estimation theory, a Bayes estimator is an estimator or decision rule that maximizes the posterior expected value of a utility function or minimizes the posterior expected value of a loss function (also called posterior expected loss).
Contents |
[edit] Definition
Suppose an unknown parameter θ is known to have a prior distribution π. Let δ be an estimator of θ (based on some measurements), and let R(θ,δ) be a risk function, such as the mean squared error. The Bayes risk of δ is defined as Eπ{R(θ,δ)}, where the expectation is taken over the probability distribution of θ. An estimator δ is said to be a Bayes estimator if it minimizes the Bayes risk among all estimators. The estimator which minimizes the posterior expected loss for each x also minimizes the Bayes risk and therefore is a Bayes estimator.
If the prior is improper then an estimator which minimizes the posterior expected loss for each x is called a generalized Bayes estimator.
[edit] Examples
[edit] Minimum mean square error estimation
The most common risk function used for Bayesian estimation is the mean square error (MSE), also called squared error risk. The MSE is defined by
- ,
where the expectation is taken over the joint distribution of θ and x.
Using the MSE as risk, the Bayes estimate of the unknown parameter is simply the mean of the posterior distribution,
This is known as the minimum mean square error (MMSE) estimator. The Bayes risk, in this case, is the posterior variance.
[edit] Bayes estimators for conjugate priors
If there is no inherent reason to prefer one prior probability distribution over another, a conjugate prior is sometimes chosen for simplicity. A conjugate prior is defined as a prior distribution belonging to some parametric family, for which the resulting posterior distribution also belongs to the same family. This is an important property, since the Bayes estimator, as well as its statistical properties (variance, confidence interval, etc.), can all be derived from the posterior distribution.
Conjugate priors are especially useful for sequential estimation, where the posterior of the current measurement is used as the prior in the next measurement. In sequential estimation, unless a conjugate prior is used, the posterior distribution typically becomes more complex with each added measurement, and the Bayes estimator cannot usually be calculated without resorting to numerical methods.
Following are some examples of conjugate priors.
- If x|θ is normal, x|θ ~ N(θ,σ2, and the prior is normal, θ ~ N(μ,τ2), then the posterior is also normal and the Bayes estimator under MSE is given by
- If x1,...,xn are iid Poisson random variables xi|θ ~ P(θ), and if the prior is Gamma distributed θ ~ G(a,b), then the posterior is also Gamma distributed, and the Bayes estimator under MSE is given by
- If x1,...,xn are iid uniformly distributed xi|θ~U(0,θ), and if the prior is Pareto distributed θ~Pa(θ0,a), then the posterior is also Pareto distributed, and the Bayes estimator under MSE is given by
[edit] Alternative risk functions
Risk functions are chosen depending on how one measures the distance between the estimate and the unknown parameter. The MSE is the most common risk function in use, primarily due to its simplicity. However, alternative risk functions are also occasionally used. The following are several examples of such alternatives. We denote the posterior generalized distribution function by F.
- A "linear" loss function, with a > 0, which yields the posterior median as the Bayes' estimate:
- Another "linear" loss function, which assigns different "weights" a,b > 0 to over or sub estimation. It yields a quantile from the posterior distribution, and is a generalization of the previous loss function:
- The following loss function is trickier: it yields either the posterior mode, or a point close to it depending on the curvature and properties of the posterior distribution. Small values of the parameter K > 0 are recommended, in order to use the mode as an approximation (L > 0):
Other loss functions can be conceived, although the mean squared error is the most widely used and validated.
[edit] Generalized Bayes estimator
This article or section needs copy editing for grammar, style, cohesion, tone or spelling. You can assist by editing it now. A how-to guide is available. (April 2008) |
Improper prior has infinite mass and as a result the Bayes risk is usually infinite and has no meaning. However, the posterior expected loss usually exists, represented by-
where L is the loss function, a is an action and π(θ|x) is the posterior density.
A Generalized Bayes estimator, for a given x, is an action which minimizes the posterior expected loss (when the prior π(θ) is improper).
A useful example is location parameter estimation under L(a-θ) loss function:
Here θ is a location parameter and fx|θ=f(x-θ). It is common to use the improper prior π(θ)=1 in this case, specially when no other more subjective information is available. This yields,
π(θ|x)=π(θ)•fx|θ=f(x-θ), so the posterior expected loss is (by defining y=x-θ),
Defining C=a-x we get,
therefore the Generalized Bayes estimator is x+C where C is a constant minimizing E[L(y+C)].
Under MSE, as a private case, and the generalized Bayes estimator is δ(x)=x-E[y].
Assuming for example gaussian samples X|θ~N(θ,Ip) where X=(x1,...,xp) and θ=(θ1,...,θp) , then the generalized Bayes estimator of θ is δ(X)=X .
[edit] Empirical Bayes estimator
This article or section needs copy editing for grammar, style, cohesion, tone or spelling. You can assist by editing it now. A how-to guide is available. (April 2008) |
A Bayes estimator derived through the empirical Bayes method is called an empirical Bayes estimator. Empirical Bayes methods enable the use of auxiliary empirical data, from past observations, in the development of a Bayes estimator. This is under the assumption that the estimated parameters are from a common prior. Similarly, in compound decision problems (where simultaneous independent observations are being held) the data from current observations can be used.
Parametric empirical Bayes (PEB) is usually preferable since it is more applicable and more accurate on small amounts of data.[1]
Example for PEB estimation:
Given x1,...xn past observations with the conditional distribution f(xi|θi), the estimation of θn+1 based on xn+1 is required.
Assuming that θi have common prior with a specific parametric form (e.g. normal), we can use the past observations to determine the moments of that prior μπ and σπ (mean and variance)in the following way:
First we estimate the moments μm and σm of the marginal distribution of x1,...xn by,
Then we can use the following connection, where μf(θ) and σf(θ) are the moments of the conditional distribution,
Further assuming that μf(θ)=θ and σf(θ)=K is constant, we get:
So finally we get the estimated moments of the prior,
Now, if for example xi|θi~N(θi,1) and we assume a normal prior (which is conjugate prior in this case) so and we can calculate the Bayes estimator of θn+1 based on xn+1.
[edit] Properties
This article or section needs copy editing for grammar, style, cohesion, tone or spelling. You can assist by editing it now. A how-to guide is available. (April 2008) |
[edit] Admissibility of Bayes estimators
Bayes rules with finite Bayes risk are typically admissible:
- If a Bayes rule is unique then it is admissible. For example, as stated above, under mean squared error (MSE) the Bayes rule is unique and therefore admissible.
- For discrete θ, Bayes rules are admissible.
- For continues θ, if the risk function R(θ,δ) is continues in θ for every δ then the Bayes rules are admissible.
However, Generalized Bayes rules usually have infinite Bayes risk. These can be inadmissible and the verification of their admissibility can be difficult. For example, the generalized Bayes estimator of θ based on gaussian samples which is described in the "Generalized Bayes estimator" section above, is inadmissible for p>2 since it is well known that the James-Stein estimator has smaller risk for all θ.
[edit] Asymptotic efficiency of Bayes estimators
Suppose that x1,…,xn are iid samples with density f(xi|θ) and δn=δ(x1 ,…,xn) is Bayes estimator of θ. In addition, let be the true (unknown) value of θ. While Bayesian analysis assumes θ has density π(θ) and posterior density π(θ|X), for analyzing the asymptotic behavior of δ we regard θ0 as a deterministic parameter. Under specific conditions,[2] for large samples (large values of n), the posterior density of θ is approximately normal. This means that for large n the effect of the prior probability which was given to θ declines!
Moreover, if δ is the Bayes estimator under MSE then it is asymptotically unbiased and it converges in distribution to the normal distribution:
where I(θ0) is the fisher information of θ0. As a conclusion, the Bayes estimator δn under MSE is asymptotically efficient.
Another estimator which is asymptotically normal and efficient is the deterministic Maximum likelihood estimator (MLE), the relations between the two (for large samples) can be shown in the following simple example. Consider the estimator of θ based on binomial sample x~b(θ,n) where θ denotes the probability for success. Assuming the prior of θ is a Beta distribution, B(a,b), this is a conjugate prior and the posterior distribution is known to be B(a+x,b+n-x). So the Bayes estimator under MSE is,
The MLE in this case is x/n and so we get,
The last equation implies that, for n → ∞, the Bayes estimator (in the described problem) is close to the MLE. On the other hand when n is small the prior is more dominant.
[edit] See also
[edit] Notes
[edit] References
- Lehmann, E. L.; Casella, G. (1998). Theory of Point Estimation. Springer, 2nd ed. ISBN 0-387-98502-6.
- Berger, J.O. (1985). Statistical Decision Theory and Bayesian Analysis. Springer Verlag, New York, Second Edition.. ISBN ISBN 0-387-96098-8 and also ISBN 3-540-96098-8.