Chi-square distribution

From Wikipedia, the free encyclopedia

This article is about the mathematics of the chi-square distribution. For its uses in statistics, see chi-square test.
chi-square
Probability density function
Cumulative distribution function
Parameters k > 0\, degrees of freedom
Support x \in [0; +\infty)\,
Probability density function (pdf) \frac{(1/2)^{k/2}}{\Gamma(k/2)} x^{k/2 - 1} e^{-x/2}\,
Cumulative distribution function (cdf) \frac{\gamma(k/2,x/2)}{\Gamma(k/2)}\,
Mean k\,
Median approximately k-2/3\,
Mode k-2\, if k\geq 2\,
Variance 2\,k\,
Skewness \sqrt{8/k}\,
Excess Kurtosis 12/k\,
Entropy \frac{k}{2}\!+\!\ln(2\Gamma(k/2))\!+\!(1\!-\!k/2)\psi(k/2)
mgf (1-2\,t)^{-k/2} for 2\,t<1\,
Char. func. (1-2\,i\,t)^{-k/2}\,

In probability theory and statistics, the chi-square distribution (also chi-squared or χ2  distribution) is one of the theoretical probability distributions most widely used in inferential statistics, i.e. in statistical significance tests. It is useful because, under reasonable assumptions, easily calculated quantities can be proven to have distributions that approximate to the chi-square distribution if the null hypothesis is true.

If Xi are k independent, normally distributed random variables with means μi and variances \sigma_i^2, then the random variable

Q = \sum_{i=1}^k \left(\frac{X_i-\mu_i}{\sigma_i}\right)^2

is distributed according to the chi-square distribution. This is usually written

Q\sim\chi^2_k.\,

The chi-square distribution has one parameter: k - a positive integer that specifies the number of degrees of freedom (i.e. the number of Xi)

The chi-square distribution is a special case of the gamma distribution.

The best-known situations in which the chi-square distribution is used are the common chi-square tests for goodness of fit of an observed distribution to a theoretical one, and of the independence of two criteria of classification of qualitative data. However, many other statistical tests lead to a use of this distribution. One example is Friedman's analysis of variance by ranks.

Contents

[edit] Properties

The chi-square probability density function is

f(x;k)= \frac{(1/2)^{k/2}}{\Gamma(k/2)} x^{k/2 - 1} e^{-x/2}

where x \ge 0 and f(x;k) = 0 for x \le 0. Here Γ denotes the Gamma function. The cumulative distribution function is:

F(x;k)=\frac{\gamma(k/2,x/2)}{\Gamma(k/2)}\,

where γ(k,z) is the incomplete Gamma function.

Tables of this distribution — usually in its cumulative form — are widely available (see the External links below for online versions), and the function is included in many spreadsheets and all statistical packages.

If p independent linear homogeneous constraints are imposed on these variables, the distribution of X conditional on these constraints is \chi^2_{k-p}, justifying the term "degrees of freedom". The characteristic function of the Chi-square distribution is

\phi(t;k)=(1-2it)^{-k/2}.\,

The chi-square distribution has numerous applications in inferential statistics, for instance in chi-square tests and in estimating variances. It enters the problem of estimating the mean of a normally distributed population and the problem of estimating the slope of a regression line via its role in Student's t-distribution. It enters all analysis of variance problems via its role in the F-distribution, which is the distribution of the ratio of two independent chi-squared random variables divided by their respective degrees of freedom.

[edit] The normal approximation

If X\sim\chi^2_k, then as k tends to infinity, the distribution of X tends to normality. However, the tendency is slow (the skewness is \sqrt{8/k} and the kurtosis is 12 / k) and two transformations are commonly considered, each of which approaches normality faster than X itself:

Fisher showed that \sqrt{2X} is approximately normally distributed with mean \sqrt{2k-1} and unit variance.

Wilson and Hilferty showed in 1931 that \sqrt[3]{X/k} is approximately normally distributed with mean 1 − 2 / (9k) and variance 2 / (9k).

The expected value of a random variable having chi-square distribution with k degrees of freedom is k and the variance is 2k. The median is given approximately by

k-\frac{2}{3}+\frac{4}{27k}-\frac{8}{729k^2}.

Note that 2 degrees of freedom leads to an exponential distribution.

The information entropy is given by

H = \int_{-\infty}^\infty f(x;k)\ln(f(x;k)) dx = \frac{k}{2} + \ln  \left(   2 \Gamma   \left(    \frac{k}{2}   \right)  \right) + \left(1 - \frac{k}{2}\right) \psi(k/2).

where ψ(x) is the Digamma function.

[edit] Related distributions

Various chi and chi-square distributions
Name Statistic
chi-square distribution \sum_{i=1}^k \left(\frac{X_i-\mu_i}{\sigma_i}\right)^2
noncentral chi-square distribution \sum_{i=1}^k \left(\frac{X_i}{\sigma_i}\right)^2
chi distribution \sqrt{\sum_{i=1}^k \left(\frac{X_i-\mu_i}{\sigma_i}\right)^2}
noncentral chi distribution \sqrt{\sum_{i=1}^k \left(\frac{X_i}{\sigma_i}\right)^2}

[edit] See also

[edit] External links

Image:Bvn-small.png Probability distributionsview  talk  edit ]
Univariate Multivariate
Discrete: BernoullibinomialBoltzmanncompound PoissondegenerateGauss-Kuzmingeometrichypergeometriclogarithmicnegative binomialparabolic fractalPoissonRademacherSkellamuniformYule-SimonzetaZipfZipf-Mandelbrot Ewensmultinomial
Continuous: BetaBeta primeCauchychi-squareDirac delta functionErlangexponentialexponential powerFfadingFisher's zFisher-TippettGammageneralized extreme valuegeneralized hyperbolicgeneralized inverse GaussianHalf-LogisticHotelling's T-squarehyperbolic secanthyper-exponentialhypoexponentialinverse chi-squareinverse gaussianinverse gammaKumaraswamyLandauLaplaceLévyLévy skew alpha-stablelogisticlog-normalMaxwell-BoltzmannMaxwell speednormal (Gaussian)ParetoPearsonpolarraised cosineRayleighrelativistic Breit-WignerRiceStudent's ttriangulartype-1 Gumbeltype-2 GumbeluniformVoigtvon MisesWeibullWigner semicircleWilks' lambda DirichletKentmatrix normalmultivariate normalvon Mises-FisherWigner quasiWishart
Miscellaneous: Cantorconditionalexponential familyinfinitely divisiblelocation-scale familymarginalmaximum entropyphase-typeposteriorpriorquasisamplingsingular