|
|
Line 1: |
Line 1: |
| {{No footnotes|date = July 2010}}
| | When it comes to fiberglass ship repair the first-step is considering how serious the damage is. In some cases, severe damage will probably need a lot of repairs done by a professional. Though because of the mess and equipment required for even a little fiberglass boat repair cause many to choose to get an expert to repair any injury to a fiberglass boat rather than carrying it out themselves. <br><br>If you do choose to perform the fiberglass boat restoration yourself then you should con-sider studying the whole means of how fiberglass works because this assists you when fixing the boat. Even though bear in mind that there are several things that could fail when you"re doing fiberglass ship repairs all on your own if you don"t take the time to research the required repairs effectively. Although there are many companies out there that can help you with fiberglass boat repair. Browse here at [http://www.youtube.com/watch?v=Gq_cOkCxZBk fix garage door panel los angeles] to discover the inner workings of this enterprise. <br><br>If you choose to leave the repairs up to the skilled then you still need to take enough time to consider who is the proper person to take your ship to. You don"t want to leave your fiberglass boat repairs inside the hands of just anyone, instead you want to ensure they are a professional who knows exactly what they are doing. So that you can get the right professional for the work consider the following recommendations. <br><br>First see when the person really specializes in fiberglass boat re-pair. These individuals can provide you the level and type of care that your boat needs along with professionalism. The truth is, by taking your boat to somebody who focuses primarily on fiberglass boat repair you"ll be having the repair done at a faster pace with a higher level of performance then you could get with other people. <br><br>Also you want to locate a organization that"s willing to work with you through the repair process. This means the organization should know what work-you want done and be prepared to be practical. The company needs to have a top quality-of customer support. Because you will have the confidence they will provide you with a great degree of service this also helps to construct your confidence in them. This can be expected of anyone who would go to a specialist for fiberglass ship re-pair. Be taught more on an affiliated article directory by visiting [http://www.youtube.com/watch?v=iXLGdfa7tcM&feature=plcp read about garage doors oxnard]. <br><br>It is also a good idea to select a company that has been highly recommended. The simplest way to do that is to study the company online, which helps you to get more information about the company you"re about to do business with. By researching a company you can see if they have any negative states and what customer service opinions have to say in regards to the company. This could go quite a distance to giving the piece to you of mind that you are leaving your boat in the best hands. <br><br>It"s important that you make an effort to obtain the appropriate business to your fiberglass ship repair. This permits you to determine a business that will provide both you and your ship with the highest-level of attention while giving the very best price to you available. This tasteful [http://www.youtube.com/watch?v=u7klWknFJkE&feature=youtu.be garage door opener repair northridge] website has a pile of powerful aids for why to acknowledge this belief. Through study you will quickly find that there"s no lack of companies who are able to provide you with the quality degree of fiberglass ship fix that you need..<br><br>If you loved this post and you would like to get much more facts pertaining to [http://longtact8510.page.tl health reform] kindly pay a visit to our internet site. |
| {{refimprove|date = July 2010}}
| |
| In [[statistics]], the concept of being an '''invariant estimator''' is a criterion that can be used to compare the properties of different [[estimator]]s for the same quantity. It is a way of formalising the idea that an estimator should have certain intuitively appealing qualities. Strictly speaking, "invariant" would mean that the estimates themselves are unchanged when both the measurements and the parameters are transformed in a compatible way, but the meaning has been extended to allow the estimates to change in appropriate ways with such transformations. The term '''equivariant estimator''' is used in formal mathematical contexts that include a precise description of the relation of the way the estimator changes in response to changes to the dataset and parameterisation: this corresponds to the use of "[[Equivariant map|equivariance]]" in more general mathematics.
| |
| | |
| ==General setting==
| |
| ===Background===
| |
| | |
| In [[statistical inference]], there are several approaches to [[estimation theory]] that can be used to decide immediately what estimators should be used according to those approaches. For example, ideas from [[Bayesian inference]] would lead directly to [[Bayesian estimator]]s. Similarly, the theory of classical statistical inference can sometimes lead to strong conclusions about what estimator should be used. However, the usefulness of these theories depends on having a fully prescribed statistical model and may also depend on having a relevant loss function to determine the estimator. Thus a [[Bayesian inference|Bayesian analysis]] might be undertaken, leading to a posterior distribution for relevant parameters, but the use of a specific utility or loss function may be unclear. Ideas of invariance can then be applied to the task of summarising the posterior distribution. In other cases, statistical analyses are undertaken without a fully defined statistical model or the classical theory of statistical inference cannot be readily applied because the family of models being considered are not amenable to such treatment. In addition to these cases where general theory does not prescribe an estimator, the concept of invariance of an estimator can be applied when seeking estimators of alternative forms, either for the sake of simplicity of application of the estimator or so that the estimator is [[Robust statistics|robust]].
| |
| | |
| The concept of invariance is sometimes used on its own as a way of choosing between estimators, but this is not necessarily definitive. For example, a requirement of invariance may be incompatible with the requirement that the [[Bias of an estimator|estimator be mean-unbiased]]; on the other hand, the criterion of [[Median#Median-unbiased_estimators|median-unbiasedness]] is defined in terms of the estimator's sampling distribution and so is invariant under many transformations.
| |
| | |
| One use of the concept of invariance is where a class or family of estimators is proposed and a particular formulation must be selected amongst these. One procedure is to impose relevant invariance properties and then to find the formulation within this class that has the best properties, leading to what is called the optimal invariant estimator.
| |
| | |
| ===Some classes of invariant estimators===
| |
| There are several types of transformations that are usefully considered when dealing with invariant estimators. Each gives rise to a class of estimators which are invariant to those particular types of transformation.
| |
| *Shift invariance: Notionally, estimates of a [[location parameter]] should be invariant to simple shifts of the data values. If all data values are increased by a given amount, the estimate should change by the same amount. When considering estimation using a [[weighted average]], this invariance requirement immediately implies that the weights should sum to one. While the same result is often derived from a requirement for unbiasedness, the use of "invariance" does not require that a mean value exists and makes no use of any probability distribution at all.
| |
| *Scale invariance: Note that this is a topic not directly covered in [[scale invariance]].
| |
| *Parameter-transformation invariance: Here, the transformation applies to the parameters alone. The concept here is that essentially the same inference should be made from data and a model involving a parameter θ as would be made from the same data if the model used a parameter φ, where φ is a one-to-one transformation of θ, φ=''h''(θ). According to this type of invariance, results from transformation-invariant estimators should also be related by φ=''h''(θ). [[Maximum likelihood estimator]]s have this property.
| |
| *Permutation invariance: Where a set of data values can be represented by a statistical model that they are outcomes from [[independent and identically distributed]] [[random variables]], it is reasonable to impose the requirement that any estimator of any property of the common distribution should be permutation-invariant: specifically that the estimator, considered as a function of the set of data-values, should not change if items of data are swapped within the dataset.
| |
| | |
| The combination of permutation invariance and location invariance for estimating a location parameter from an [[independent and identically distributed]] dataset using a weighted average implies that the weights should be identical and sum to one. Of course, estimators other than a weighted average may be preferable.
| |
| | |
| ===Optimal invariant estimators===
| |
| Under this setting, we are given a set of measurements <math>x</math> which contains information about an unknown parameter <math>\theta</math>. The measurements <math>x</math> are modelled as a [[random vector|vector random variable]] having a [[probability density function]] <math>f(x|\theta)</math> which depends on a parameter vector <math>\theta</math>.
| |
| | |
| The problem is to estimate <math>\theta</math> given <math>x</math>. The estimate, denoted by <math>a</math>, is a function of the measurements and belongs to a set <math>A</math>. The quality of the result is defined by a [[loss function]] <math>L=L(a,\theta)</math> which determines a [[risk function]] <math>R=R(a,\theta)=E[L(a,\theta)|\theta]</math>. The sets of possible values of <math>x</math>, <math>\theta</math>, and <math>a</math> are denoted by <math>X</math>, <math>\Theta</math>, and <math>A</math>, respectively. | |
| | |
| ===In classification===
| |
| In [[statistical classification]], the rule which assigns a class to a new data-item can be consider to be a special type of estimator. A number of invariance-type considerations can be brought to bear in formulating [[prior knowledge for pattern recognition]].
| |
| | |
| ==Mathematical setting==
| |
| === Definition ===
| |
| An invariant estimator is an estimator which obeys the following two rules:{{citation needed|date=November 2010}}
| |
| # Principle of Rational Invariance: The action taken in a decision problem should not depend on transformation on the measurement used
| |
| # Invariance Principle: If two decision problems have the same formal structure (in terms of <math>X</math>, <math>\Theta</math>, <math>f(x|\theta)</math> and <math>L</math>), then the same decision rule should be used in each problem.
| |
| | |
| To define an invariant or equivariant estimator formally, some definitions related to groups of transformations are needed first. Let <math>X</math> denote the set of possible data-samples. A [[group of transformation]]s of <math>X</math>, to be denoted by <math>G</math>, is a set of (measurable) 1:1 and onto transformations of <math>X</math> into itself, which satisfies the following conditions:
| |
| | |
| # If <math>g_1\in G</math> and <math>g_2\in G</math> then <math>g_1 g_2\in G \,</math>
| |
| # If <math>g\in G</math> then <math>g^{-1}\in G</math>, where <math>g^{-1}(g(x))=x \, .</math> (That is, each transformation has an inverse within the group.)
| |
| # <math>e\in G</math> (i.e. there is an identity transformation <math> e(x)=x \, </math>)
| |
| | |
| Datasets <math>x_1</math> and <math>x_2</math> in <math>X</math> are equivalent if <math>x_1=g(x_2)</math> for some <math>g\in G</math>. All the equivalent points form an [[equivalence class]].
| |
| Such an equivalence class is called an [[orbit (group theory)|orbit]] (in <math>X</math>). The <math>x_0</math> orbit, <math>X(x_0)</math>, is the set <math>X(x_0)=\{g(x_0):g\in G\}</math>.
| |
| If <math>X</math> consists of a single orbit then <math>g</math> is said to be transitive.
| |
| | |
| A family of densities <math>F</math> is said to be invariant under the group <math>G</math> if, for every <math>g\in G</math> and <math>\theta\in \Theta</math> there exists a unique <math>\theta^*\in \Theta</math> such that <math>Y=g(x)</math> has density <math>f(y|\theta^*)</math>. <math>\theta^*</math> will be denoted <math>\bar{g}(\theta)</math>.
| |
| | |
| If <math>F</math> is invariant under the group <math>G</math> then the loss function <math>L(\theta,a)</math> is said to be invariant under <math>G</math> if for every <math>g\in G</math> and <math>a\in A</math> there exists an <math>a^*\in A</math> such that <math>L(\theta,a)=L(\bar{g}(\theta),a^*)</math> for all <math>\theta \in \Theta</math>. The transformed value <math>a^*</math> will be denoted by <math>\tilde{g}(a)</math>.
| |
| | |
| In the above, <math>\bar{G}=\{\bar{g}:g\in G\}</math> is a group of transformations from <math>\Theta</math> to itself and <math>\tilde{G}=\{\tilde{g}: g \in G\}</math> is a group of transformations from <math>A</math> to itself.
| |
| | |
| An estimation problem is invariant(equivariant) under <math>G</math> if there exist three groups <math>G, \bar{G}, \tilde{G}</math> as defined above.
| |
| | |
| For an estimation problem that is invariant under <math>G</math>, estimator <math>\delta(x)</math> is an invariant estimator under <math>G</math> if, for all <math>x\in X</math> and <math>g\in G</math>,
| |
| :<math>\delta(g(x)) = \tilde{g}(\delta(x)).</math>
| |
| | |
| === Properties ===
| |
| | |
| # The risk function of an invariant estimator, <math>\delta</math>, is constant on orbits of <math>\Theta</math>. Equivalently <math>R(\theta,\delta)=R(\bar{g}(\theta),\delta)</math> for all <math>\theta \in \Theta</math> and <math>\bar{g}\in \bar{G}</math>.
| |
| # The risk function of an invariant estimator with transitive <math>\bar{g}</math> is constant.
| |
| | |
| For a given problem, the invariant estimator with the lowest risk is termed the "best invariant estimator". Best invariant estimator cannot always be achieved. A special case for which it can be achieved is the case when <math>\bar{g}</math> is transitive.
| |
| | |
| === Example: Location parameter ===
| |
| | |
| Suppose <math>\theta</math> is a location parameter if the density of <math>X</math> is of the form <math>f(x-\theta)</math>. For <math> \Theta=A=\Bbb{R}^1 </math> and <math>L=L(a-\theta)</math>, the problem is invariant under <math>g=\bar{g}=\tilde{g}=\{g_c:g_c(x)=x+c, c\in \Bbb{R}\}</math>. The invariant estimator in this case must satisfy
| |
| :<math>\delta(x+c)=\delta(x)+c, \text{ for all } c\in \Bbb{R},</math>
| |
| thus it is of the form <math>\delta(x)=x+K</math> (<math>K\in \Bbb{R}</math>). <math>\bar{g}</math> is transitive on <math>\Theta</math> so the risk does not vary with <math>\theta</math>: that is, <math>R(\theta,\delta)=R(0,\delta)=\operatorname{E}[L(X+K)|\theta=0]</math>. The best invariant estimator is the one that brings the risk <math>R(\theta,\delta)</math> to minimum.
| |
| | |
| In the case that L is the squared error <math>\delta(x)=x-\operatorname{E}[X|\theta=0].</math>
| |
| | |
| === Pitman estimator ===
| |
| The estimation problem is that <math>X=(X_1,\dots,X_n)</math> has density <math>f(x_1-\theta,\dots,x_n-\theta)</math>, where ''θ'' is a parameter to be estimated, and where the [[loss function]] is <math>L(|a-\theta|)</math>. This problem is invariant with the following (additive) transformation groups:
| |
| :<math>G=\{g_c:g_c(x)=(x_1+c, \dots, x_n+c),c\in \Bbb{R}^1\},</math>
| |
| :<math>\bar{G}=\{g_c:g_c(\theta)=\theta + c,c\in \Bbb{R}^1\},</math>
| |
| :<math>\tilde{G}=\{g_c:g_c(a)=a + c,c\in \Bbb{R}^1\} .</math>
| |
| | |
| The best invariant estimator <math>\delta(x)</math> is the one that minimizes
| |
| :<math>\frac{\int_{-\infty}^{\infty}{L(\delta(x)-\theta)f(x_1-\theta,\dots,x_n-\theta)d\theta}}{\int_{-\infty}^{\infty}{f(x_1-\theta,\dots,x_n-\theta)d\theta}},</math>
| |
| and this is Pitman's estimator (1939).
| |
| | |
| For the squared error loss case, the result is
| |
| :<math>\delta(x)=\frac{\int_{-\infty}^{\infty}{\theta f(x_1-\theta,\dots,x_n-\theta)d\theta}}{\int_{-\infty}^{\infty}{f(x_1-\theta,\dots,x_n-\theta)d\theta}}.</math>
| |
| | |
| If <math>x \sim N(\theta 1_n,I)\,\!</math> (i.e. a [[multivariate normal distribution]] with independent, unit-variance components) then
| |
| :<math>\delta_{pitman} = \delta_{ML}=\frac{\sum{x_i}}{n}.</math>
| |
| | |
| If <math>x \sim C(\theta 1_n,I \sigma^2)\,\!</math> (independent components having a [[Cauchy distribution]] with scale parameter ''σ'') then | |
| <math>\delta_{pitman} \ne \delta_{ML}</math>,. However the result is
| |
| :<math>\delta_{pitman}=\sum_{k=1}^n{x_k\left[\frac{Re\{w_k\}}{\sum_{m=1}^{n}{Re\{w_k\}}}\right]}, \qquad n>1,</math> | |
| with
| |
| :<math>w_k = \prod_{j\ne k}\left[\frac{1}{(x_k-x_j)^2+4\sigma^2}\right]\left[1-\frac{2\sigma}{(x_k-x_j)}i\right].</math>
| |
| | |
| {{No footnotes|date=January 2011}}
| |
| | |
| ==References==
| |
| * {{cite book
| |
| |title=Statistical decision theory and Bayesian Analysis
| |
| |first=James O. |last=Berger |authorlink=James Berger (statistician)
| |
| |year=1985
| |
| |edition=2nd
| |
| |publisher=Springer-Verlag |location=New York
| |
| |ISBN=0-387-96098-8 |mr=0804611
| |
| }}{{Page needed|date=January 2011}}
| |
| * Freue, Gabriela V. Cohen (2007) "The Pitman estimator of the Cauchy location parameter", ''Journal of Statistical Planning and Inference'', 137, 1900–1913 {{doi|10.1016/j.jspi.2006.05.002}}
| |
| * Pitman, E.J.G. (1939) "The estimation of the location and scale parameters of a continuous population of any given form", ''[[Biometrika]]'', 30 (3/4), 391–421. {{jstor|2332656}}
| |
| * Pitman, E.J.G. (1939) "Tests of Hypotheses Concerning Location and Scale Parameters", ''[[Biometrika]]'', 31 (1/2), 200–215. {{jstor|2334983}}
| |
| | |
| {{DEFAULTSORT:Invariant Estimator}}
| |
| [[Category:Statistical theory]]
| |
| [[Category:Estimation theory]]
| |
| [[Category:Statistical terminology]]
| |
| [[Category:Invariant theory]]
| |