Conjugate prior

From Wikipedia, the free encyclopedia

In Bayesian probability theory, a class of prior probability distributions p(θ) is said to be conjugate to a class of likelihood functions p(x|θ) if the resulting posterior distributions p(θ|x) are in the same family as p(θ). For example, the Gaussian family is conjugate to itself (or self-conjugate): if the likelihood function is Gaussian, choosing a Gaussian prior will ensure that the posterior distribution is also Gaussian. The concept, as well as the term "conjugate prior", was introduced by Howard Raiffa and Robert Schlaifer in their work on Bayesian decision theory.[1] A similar concept had been discovered independently by George Alfred Barnard.[2]

Consider the general problem of inferring a distribution for a parameter θ given some datum or data x. From Bayes' theorem, the posterior distribution is calculated from the prior p(θ) and the likelihood function \theta \mapsto p(x\mid\theta)\! as

p(\theta|x) = \frac{p(x|\theta) \, p(\theta)}   {\int p(x|\theta) \, p(\theta) \, d\theta}. \!

Let the likelihood function be considered fixed; the likelihood function is usually well-determined from a statement of the data-generating process. It is clear that different choices of the prior distribution p(θ) may make the integral more or less difficult to calculate, and the product p(x|θ) × p(θ) may take one algebraic form or another. For certain choices of the prior, the posterior has the same algebraic form as the prior (generally with different parameters). Such a choice is a conjugate prior.

A conjugate prior is an algebraic convenience: otherwise a difficult numerical integration may be necessary.

All members of the exponential family have conjugate priors. See Gelman et al.[3] for a catalog.

Contents

[edit] Example

For a random variable which is a Bernoulli trial with unknown probability of success q in [0,1], the usual conjugate prior is the beta distribution with

p(q=x) = {x^{\alpha-1}(1-x)^{\beta-1} \over \Beta(\alpha,\beta)}

where α and β are chosen to reflect any existing belief or information (α = 1 and β = 1 would give a uniform distribution) and Β(αβ) is the Beta function acting as a normalising constant.

If we then sample this random variable and get s successes and f failures, we have

P(s,f|q=x) = {s+f \choose s} x^s(1-x)^f,
p(q=x|s,f) = {{{s+f \choose s} x^{s+\alpha-1}(1-x)^{f+\beta-1} / \Beta(\alpha,\beta)} \over  \int_{y=0}^1 \left({s+f \choose s} y^{s+\alpha-1}(1-y)^{f+\beta-1} / \Beta(\alpha,\beta)\right) dy} = {x^{s+\alpha-1}(1-x)^{f+\beta-1} \over \Beta(s+\alpha,f+\beta)} ,

which is another Beta distribution with a simple change to the parameters. This posterior distribution could then be used as the prior for more samples, with the parameters simply adding each extra piece of information as it comes.

[edit] Table of conjugate distributions

[edit] Discrete likelihood distributions

Likelihood Model parameters Conjugate prior distribution Prior hyperparameters Posterior hyperparameters
Bernoulli p (probability) Beta \alpha,\ \beta\! \alpha + \sum_{i=1}^n x_i,\ \beta + n - \sum_{i=1}^n x_i\!
Binomial p (probability) Beta \alpha,\ \beta\! \alpha + \sum_{i=1}^n x_i,\ \beta + \sum_{i=1}^n N_i - x_i\!
Poisson λ (rate) Gamma \alpha,\ \beta\! \alpha + \sum_{i=1}^n x_i,\ \beta + n\!
Multinomial p (probability vector) Dirichlet \vec{\alpha}\! \vec{\alpha}+\sum_{i=1}^n\vec{x}^{\,(i)}\!
Geometric p0 (probability) Beta \alpha,\ \beta\! \alpha + n,\ \beta + \sum_{i=1}^n x_i\!

[edit] Continuous likelihood distributions

Likelihood Model parameters Conjugate prior distribution Prior hyperparameters Posterior hyperparameters
Uniform U(0,\theta)\! Pareto X_{m},\ k\! \min\{\,X_{(n)},X_{m}\}, k+n\!
Exponential λ (rate) Gamma \alpha,\ \beta\! \alpha+n,\ \beta+\sum_{i=1}^n x_i\!
Normal
with known variance σ2
μ (mean) Normal \mu_0, \sigma_0^2\! (\frac{\mu_0}{\sigma_0^2} + \frac{\sum_{i=1}^n x_i}{\sigma^2})/(\frac{1}{\sigma_0^2} + \frac{n}{\sigma^2}), (\frac{1}{\sigma_0^2} + \frac{n}{\sigma^2})^{-1}
Normal
with known mean μ
σ2 (variance) Scaled inverse chi-square \nu,\ \sigma^2\! \nu+n,\frac{\nu\sigma^2+\sum_{i=1}^n (x_i-\mu)^2}{\nu+n}\!
Normal
with known mean μ
τ (precision) Gamma \alpha,\ \beta\! \alpha+\frac{n}{2},\beta+\frac{\sum_{i=1}^n (x_i-\mu)^2}{2}\!
Normal μ and σ2
Assuming dependence
Normal-Scaled inverse gamma m,\ u,\ \nu,\ \sigma^2\ \frac{um+n\bar{x}}{u+n},u+n, \nu+n, \frac{\nu\sigma^2}{\nu+n}+\frac{(n-1)S^2}{\nu+n}+\frac{nu(\bar{x}-m)^2}{(n+u)(\nu+n)}, where \bar{x} is sample mean and S2 is the sample variance.
Normal μ and τ
Assuming dependence
Normal-Gamma m,\ u,\ \alpha,\ \beta\ \frac{um+n\bar{x}}{u+n},u+n, \alpha+\frac{n}{2}, \beta+\frac{nS^2}{2}+\frac{nu(\bar{x}-m)^2}{2(n+u)}, where \bar{x} is sample mean and S2 is the sample variance.
Multivariate normal μ (mean vector) Multivariate normal
Multivariate normal Σ (variance matrix) inverse-Wishart
Pareto k (shape) Gamma \alpha, \beta\! \alpha+n,\ \beta+\sum_{i=1}^n \ln\frac{x_i}{x_{\mathrm{m}}}\!
Pareto xm (location) Pareto

[edit] Notes

  1. ^ Howard Raiffa and Robert Schlaifer. Applied Statistical Decision Theory. Division of Research, Graduate School of Business Administration, Harvard University, 1961.
  2. ^ Jeff Miller et al. Earliest Known Uses of Some of the Words of Mathematics, "conjugate prior distributions". Electronic document, revision of November 13, 2005, retrieved December 2, 2005.
  3. ^ Andrew Gelman, John B. Carlin, Hal S. Stern, and Donald B. Rubin. Bayesian Data Analysis, 2nd edition. CRC Press, 2003. ISBN 1-58488-388-X