Negentropy
From Wikipedia, the free encyclopedia
In 1943 Erwin Schrödinger used the concept of “negative entropy” in his popular-science book What is life?. The actual term “negentropy” was later coined by Léon Brillouin. In 1974, Albert Szent-Györgyi proposed replacing the term negentropy with syntropy.
Schrödinger introduced the concept when explaining that a living system exports entropy in order to maintain its own entropy at a low level. By using the term "Negentropy", he could express this fact in a more "positive" way: A living system imports negentropy and stores it.
Actually, negentropy is a misconception derived from entropy. Negentropy is the loss of entropy in a system by having more entropy flowing out than in. An example in thermodynamics, where entropy is heat divided by temperature, is the flow of energy from hot to cold via another medium. This medium will be in an energetic equilibrium, that is, emit as much energy as it receives. Yet, emission occurs at a lower temperature than immission. That means, that MORE entropy is emitted than is received, resulting in a net loss of entropy in the transisting medium. However, the net total amount of entropy rises. (As the temperature by which the heat is conducted is now lower than before.)
In a note to What is Life? Schrödinger explains his usage of this term.
- Let me say first, that if I had been catering for them [physicists] alone I should have let the discussion turn on free energy instead. It is the more familiar notion in this context. But this highly technical term seemed linguistically too near to energy for making the average reader alive to the contrast between the two things. (Erwin Schrödinger)
[edit] Information theory
In information theory, “negentropy” is used as a measure of distance to normality. Consider a signal with a certain distribution. If the signal is Gaussian, the signal is said to have a normal distribution. Negentropy is always positive, is invariant by any linear invertible change of coordinates, and vanishes iff the signal is Gaussian.
Negentropy is defined as
- J(px) = S(φx) − S(px)
where S(φx) stands for the Gaussian density with the same mean and variance as px and S(px) is the differential entropy:
Negentropy is used in statistics and signal processing. It is related to network entropy, which is used in Independent Component Analysis.
P. Comon, Independent Component Analysis - a new concept?, Signal Processing, 36:287-314, 1994.
[edit] Organisation Theory
Ashby (1952) and von Bertalanffy (1950) both considered living systems to be negentropic, running against the arrow of time, seeking to create more order from disorder and conserving free energy. Therefore it is important to distinguish between the organism, which is negentropic, and its physical environment. In risk management, negentropy is the force that seeks to achieve effective organisational behaviour and lead to a steady predictable state (Foster 2006).