(redirected from Consistent estimator)
Also found in: Dictionary, Thesaurus, Medical.
Related to Consistent estimator: Efficient estimator, Unbiased estimator


A random variable or a function of it used to estimate population parameters.
McGraw-Hill Dictionary of Scientific & Technical Terms, 6E, Copyright © 2003 by The McGraw-Hill Companies, Inc.
The following article is from The Great Soviet Encyclopedia (1979). It might be outdated or ideologically biased.



in statistics, a function of the results of observations that is used to estimate an unknown parameter of the probability distribution of random variables that are under study. In English, a distinction is sometimes, but not always, made between the terms “estimator” and “estimate”: an estimate is the numerical value of the estimator for a particular sample.

Suppose, for example, that X1, . . . , Xn are independent random variables having the same normal distribution with the unknown mean a. Possible point estimators of a are the arithmetic mean of the observation results

and the sample median μ = μ(X1,..., Xn).

In choosing an estimator of a parameter θ, it is natural to select a function θ*(X1, . . . , Xn) of the observation results X1, . . . , Xn that is in some sense close to the true value of the parameter. By adopting some measure of the closeness of an estimator to the parameter being estimated, different estimators can be compared with respect to quality. A commonly used measure of closeness is the magnitude of the mean squared error

Eθ(θ* – θ)2 = Dθθ* + (θ - Eθθ*)2

which is expressed here in terms of the mathematical expectation Eθθ* and variance Dθθ* of the estimator.

The estimator θ* is said to be unbiased if Eθθ* = θ. In the class of all unbiased estimators, the best estimators from the standpoint of mean squared error are those that have for a given n the minimum possible variance for all θ. The estimator X̄ defined above for the parameter a of a normal distribution is the best unbiased estimator, since the variance of any other unbiased estimator a* of a satisfies the inequality Daa* DaX̄ = σ2/n, where σ2 is the variance of the normal distribution. If a minimum-variance unbiased estimator exists, an unbiased best estimator can also be found in the class of functions that depend only on a sufficient statistic.

In constructing estimators for large n, it is natural to assume that as n → ∞, the probability of deviations of θ* from the true value of θ that exceed some given number will be close to θ. Estimators with this property are said to be consistent. Unbiased estimators whose variance approaches θ as n→ ∞ are consistent. Because the rate at which the limit is approached plays an important role here, an asymptotic comparison of two estimators is made by considering the ratio of their asymptotic variances. In the example given above, the arithmetic mean X̄ is the best, and consequently the asymptotically best, estimator for the parameter a, whereas the sample median μ, although an unbiased estimator, is not asymptotically best, since

Nonetheless, the use of μ sometimes has advantages. If, for example, the true distribution is not exactly normal, the variance of X̄ may increase sharply while the variance of μ remains almost the same—that is, μ has the property known as robustness.

A widely used general method of obtaining estimators is the method of moments. In this technique, a certain number of sample moments are equated to the corresponding moments of the theoretical distribution, which are functions of the unknown parameters, and the equations obtained are solved for these parameters. The method of moments is convenient to use, but the estimators produced by it are not in general asymptotically best estimators. From the theoretical point of view, the maximum likelihood method is more important. It yields estimators that, under certain general conditions, are asymptotically best. The method of least squares is a special case of the maximum likelihood method.

An important supplement to the use of estimators is provided by the estimation of confidence intervals.


Kendall, M., and A. Stuart. Statisticheskie vyvody i sviazi. Moscow, 1973. (Translated from English.)
Cramér, H. Matematicheskie metody statistiki, 2nd ed. Moscow, 1975. (Translated from English.)


The Great Soviet Encyclopedia, 3rd Edition (1970-1979). © 2010 The Gale Group, Inc. All rights reserved.


A person who, by experience and training, is capable of estimating the probable cost of a building or portion thereof.
McGraw-Hill Dictionary of Architecture and Construction. Copyright © 2003 by McGraw-Hill Companies, Inc.
References in periodicals archive ?
and denote by [??] a consistent estimator of [psi].
I first illustrate the basic approach by testing mean incomes from two samples of different sizes where parts of the samples are overlapping; I then provide consistent estimators for testing marginal dominances.(5)
Although this estimator does not have a finite mean or variance, a consistent estimator for its asymptotic variance can be obtained by standard methods.
where [MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII] When [??] is unknown, one could plug in a consistent estimator of [??].
We can start with any consistent estimator [[??].sub.2] and improve it step by step using (8).
This is a consistent estimator of the asymptotic covariance of [Mathematical Expression Omitted] given in (6) in the sense that the conditional distribution of [Mathematical Expression Omitted] weakly converges to the unconditional distribution of [Mathematical Expression Omitted].(14)
Second, one substitutes the consistent estimator [b.sub.w] into a between group version of (4.5) to get
The sample value, F = pr[(y* = a + Xc) [is greater than or equal to] O], is a consistent estimator of [PSI].
The results are simple and consistent estimators of all four parameters based on five sample quantiles.
IM first introduced a uniform CI for [[theta].sub.0] under the assumption of asymptotic joint normality of [[??].sub.l], [[??].sub.u] and other assumptions, including superefficiency of the estimator of [DELTA] [equivalent to] [[theta].sub.u]--[[theta].sub.l], where [[??].sub.l], [[??].sub.u] are consistent estimators for [[theta].sub.l], [[theta].sub.u] respectively.
[Mathematical Expression Omitted] (16) it follows that [alpha][psi] and [beta][psi] are consistent estimators of [alpha] and [beta].

Full browser ?