# Chi-Square Distribution

Also found in: Dictionary, Thesaurus, Medical, Legal, Financial, Wikipedia.

## chi-square distribution

[′kī ¦skwer dis·trə′byü·shən]
(statistics)
The distribution of the sum of the squares of a set of variables, each of which has a normal distribution and is expressed in standardized units.

## Chi-Square Distribution

The probability distribution of the sum

of the squares of the normally distributed random variables X1, . . ., Xf, with zero mathematical expectation and unit variance is known as a chi-square distribution with f degrees of freedom. The distribution function for a chi-square random variable is

The first three moments (the mathematical expectation, the variance, and the third central moment) of χ2 are f, 2f, and 8f, respectively. The sum of two independent random variables and with f1 and f2 degrees of freedom has a chi-square distribution with f1 + f2 degrees of freedom.

Examples of chi-square distributions are the distributions of the squares of random variables that obey the Rayleigh and Maxwellian distributions. The Poisson distribution can be expressed in terms of a chi-square distribution with an even number of degrees of freedom:

If the number f of terms of the sum χ2 increases without bound, then, according to the central limit theorem, the distribution function of the standardized ratio converges to the standard normal distribution:

where

A consequence of this fact is another limit relation, which is convenient for calculating Ff(x) when f has large values:

In mathematical statistics, the chi-square distribution is used to construct interval estimates and statistical tests. Let Yi, . . ., Yn be random variables representing independent measurements of an unknown constant a. Suppose the measurement errors Yia are independent and are distributed identically normally. We have

E(Yia) = 0 E(Yia)2 = σ2

The statistical estimate of the unknown variance σ2 is then expressed by the equation

s2 = S2/(n – 1)

where

The ratio S22 obeys a chi-square distribution with f = n – 1 degrees of freedom. Let x1 and x2 be positive numbers that are solutions of the equations Ff(x1) = α/2 and Ff(x2) = 1 – α/2, where α is a specified number in the interval (0,1/2). In this case

P{x1 < S22 < x2} = P{S2/x2 < σ2 < S2/x1} = 1 – α

The interval (S2/x1, S2/x2 is called the confidence interval for σ2 with confidence coefficient 1 – α.

This method of constructing an interval estimate for σ2 is often used to test the hypothesis that , where is a given number. Thus, if belongs to the confidence interval indicated, then one concludes that the measurements do not contradict the hypothesis . If, however, or , then it must be assumed that or , respectively. This test corresponds to a significance level equal to α.

### REFERENCE

Cramer, H. Matematicheskie metody statistiki, 2nd ed. Moscow, 1975. (Translated from English.)

L. N. BOLSHEV

References in periodicals archive ?
K][*] denotes the expectation taken with respect to the chi-square distribution of K ~ [chi square]([upsilon]) given in Equation 10, and I([upsilon], [[?
Information about and tables for the chi-square distribution can be found in any elementary statistics text.
This is compared with the appropriate chi-square distribution to yield a chance probability.
The mean of the chi-square distribution with n degrees of freedom is exactly n, and the variance is 2n (Hays and Winkler, 1971).
Appendix table 9 contains probabilities in the tails of the chi-square distribution.
B] approximates a chi-square distribution with p-1 degrees of freedom, where p is the number of classes or levels of the hypothesized moderator:
Under the null hypothesis of no ARCH effects, the LM statistic follows a Chi-square distribution with degree of freedom (df) equal to the number of lags in the auxiliary regression.
001 the differences between our observed Monte Carlo chi-square distribution with Yates' correction, and the theoretical chi-square.
Various assumptions are implicit in the use of the chi-square distribution to assess the significance of the Pearson chi-square statistic (Delucchi 1983).
2] falls sufficiently close to the center of a Chi-square distribution with v degrees of freedom that it is very probable from such a distribution, validating the hypothesis being tested.
For example, some of these approaches use the chi-square distribution for hypothesis testing, a potentially inappropriate application with small samples (e.

Site: Follow: Share:
Open / Close