Jump to content

User:Slava3087

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Slava3087 (talk | contribs) at 23:14, 15 April 2008. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In classical statistical decision theory, where we are faced with the problem of estimating a deterministicparameter (vector) Failed to parse (unknown function "\bfx"): {\displaystyle {\bfx}} from observations , an estimator (estimation rule) is called minimax if its maximal risk is minimal among all estimators of . In a sense this is an estimator which performs best in the worst possible case allowed in the problem.


Problem Definition

Consider the problem of estimating a deterministic (not Bayesian) parameter belonging to some sete \delta^M(y)</math> is called minimax (or minmax) with respect to a loss function



we have an estimator that is used to estimate a parameter . We also assume a risk function , usually specified as the integral of a loss function. In this framework, is called minimax if it satisfies

.

An alternative criterion in the decision theoretic framework is the Bayes estimator in the presence of a prior distribution . An estimator is Bayes if it minimizes the average risk