Binomial proportion confidence interval

From Wikipedia, the free encyclopedia

In statistics, a binomial proportion confidence interval is a confidence interval for a proportion in a statistical population. It uses the proportion estimated in a statistical sample and allows for sampling error. There are several formulas for a binomial confidence interval, but all of them rely on the assumption of a binomial distribution. A simple example of a binomial distribution is the number of heads observed when a coin is flipped ten times. In general, a binomial distribution applies when an experiment is repeated a fixed number of times, each trial of the experiment has two possible outcomes (labeled arbitrarily success and failure), the probability of success is the same for each trial, and the trials are statistically independent.

Contents

[edit] Normal approximation interval

The simplest and most commonly used formula for a binomial confidence interval relies on approximating the binomial distribution with a normal distribution. This approximation is justified by the central limit theorem. The formula is

\hat p \pm z_{1- \alpha /2} \sqrt{ \frac{\hat p \left ( 1- \hat p \right )}{n}}

where \hat p is the proportion estimated from the statistical sample, z1 − α / 2 is the 1 − α / 2 percentile of a standard normal distribution, and n is the sample size.

The central limit theorem applies well to a binomial distribution, even with a sample size less than 30, as long as the proportion is not too close to 0 or 1. For very extreme probabilities, though, a sample size of 30 or more may still be inadequate. The normal approximation fails totally when the sample proportion is exactly zero or exactly one. A frequently cited rule of thumb is that the normal approximation works well as long as np > 5 and n(1 − p) > 5; see however Brown et al. 2001.

An important theoretical derivation of this confidence interval involves the inversion of a hypothesis test. Under this formulation, the confidence interval represents those values of the population parameter that would have large p-values if they were tested as a hypothesized population proportion. The collection of values, θ, for which the normal approximation is valid can be represented as

\left \{\ \theta \ \bigg| \ Z_{\alpha / 2} \le \frac{{\hat p - \theta}}{{\sqrt{\hat p \left ( {1-\hat p} \right ) / n}}} \le Z_{1-\alpha / 2} \right \}

Since the test in the middle of the inequality is a Wald test, the normal approximation interval is sometimes called the Wald interval.

[edit] Wilson score interval

The Wilson interval is an improvement (the actual coverage probability is closer to the nominal value) over the normal approximation interval and was first developed in Wilson (1927).


\frac{{ {\hat p + \frac{{1}}{{2n}}
 z_{1- \alpha / 2}^2  \pm z_{1- \alpha / 2}
\sqrt {\frac{{\hat p\left( {1 - \hat p} \right)}}{n} + \frac{{z_{1- \alpha / 2}^2}}
{{4n^2}} }} }}
{{ {1 + \frac{{1}}{n}} z_{1- \alpha / 2}^2 }}

This interval has good properties even for a small number of trials and/or an extreme probability. The center of the Wilson interval


\frac{{ {\hat p + \frac{{1}}{{2n}}
 z_{1- \alpha / 2}^2 } }}
{{ {1 + \frac{{1}}{n}} z_{1- \alpha / 2}^2 }}

can be shown to be a weighted average of \hat p = X/n and 1/2, with \hat p receiving greater weight as the sample size increases. For the 95% interval, the Wilson interval is nearly identical to the normal approximation interval using \scriptstyle \tilde p\, =\,(X+2)/(n+4) instead of \hat p.

The Wilson interval can be derived as (The following formula may be wrong. It's identical to the way the Normal approximation is derived)

\left \{ \ \theta \ \bigg| \ Z_{\alpha / 2} \le \frac{{\hat p - \theta}}{{\sqrt{\theta \left ( {1-\theta} \right ) / n}}} \le Z_{1-\alpha / 2}\ \right \}

The test in the middle of the inequality is a score test, so the Wilson interval is sometimes called the Wilson score interval.

[edit] Clopper-Pearson interval

The Clopper-Pearson interval is an early and very common method for calculating exact binomial confidence intervals (Clopper and Pearson 1934). This method uses the cumulative probabilities of the binomial distribution. The Clopper-Pearson interval can be written as

\left \{\ \theta \ \Big| \ P  \left [ \mathrm{Bin} \left ( n; \theta \right ) \le X \right ] \ge \alpha /2 \right \}\ \bigcap\ \left \{\ \theta\ \Big|\ P  \left [ \mathrm{Bin} \left ( n; \theta \right ) \ge X \right ] \ge \alpha /2 \ \right \}

where X is the number of successes observed in the sample and Bin(n; θ) is a binomial random variable with n trials and probability of success θ.

Because of a relationship between the cumulative binomial distribution and the beta distribution, the Clopper-Pearson interval is sometimes presented in an alternate format that uses percentiles from the beta distribution. The beta distribution is, in turn, related to the F-distribution so a third formulation of the Clopper-Pearson interval uses F percentiles.

The Clopper-Pearson interval is an exact interval since it is based directly on the binomial distribution rather than any approximation to the binomial distribution. This interval, however, can be conservative because of the discrete nature of the binomial distribution.

[edit] Comparison of different intervals

There are several research papers that compare these and other confidence intervals for the binomial proportion. A good starting point is Agresti and Coull (1998) which points out that exact methods such as the Clopper-Pearson interval may not work as well as certain approximations.But it is still used today for many studies.

[edit] Web-based calculators

There are numerous web sites that will calculate a binomial proportion confidence interval.

[edit] References

  • Agresti, A., and Coull, B. Approximate is better than 'exact' for interval estimation of binomial proportions. The American Statistician 52: 119-126, 1998.
  • Brown, L. D., Cai, T. T., and DasGupta, A. Interval Estimation for a Binomial Proportion. Statistical Science 16(2): 101-117, 2001.
  • Clopper, C. and Pearson, S. The use of confidence or fiducial limits illustrated in the case of the binomial. Biometrika 26: 404-413, 1934.
  • Wilson, E. B. Probable inference, the law of succession, and statistical inference. Journal of the American Statistical Association 22: 209-212, 1927.
Languages