Jensen–Shannon divergence
From Wikipedia, the free encyclopedia
In probability theory and statistics, the Jensen-Shannon divergence is a popular method of measuring the similarity between two probability distributions. It is also known as information radius (IRad)[1] or total divergence to the average[2]. It is based on the Kullback-Leibler divergence, with the notable (and useful) difference that it is always a finite value.
[edit] Definition
Consider the set of probability distributions where A is a set provided with some σ-algebra.
Jensen-Shannon divergence (JSD) is a symmetrized and smoothed version of the Kullback-Leibler divergence. It is defined by
where
[edit] See also
Kullback-Leibler divergence for details calculating the Jensen-Shannon divergence.
[edit] References
- ^ Hinrich Schütze; Christopher D. Manning (1999). Foundations of Statistical Natural Language Processing. Cambridge, Mass: MIT Press, p. 304. ISBN 0-262-13360-1.
- ^ Dagan, Ido; Lillian Lee, Fernando Pereira (1997). "Similarity-Based Methods For Word Sense Disambiguation". Proceedings of the Thirty-Fifth Annual Meeting of the Association for Computational Linguistics and Eighth Conference of the European Chapter of the Association for Computational Linguistics: pp. 56–63.
- Jensen-Shannon Divergence and Hilbert space embedding, Bent Fuglede and Flemming Topsøe University of Copenhagen, Department of Mathematics [1]
- J. Lin. Divergence measures based on the shannon entropy. IEEE Trans. on Information Theory, 37(1):145--151, January 1991.
- Y. Ofran & B. Rost. Analysing Six Types of Protein-Protein Interfaces. 2003.