The negentropy has different meanings in information theory and theoretical biology. In a biological context, the negentropy (also negative entropy, syntropy, extropy, ectropy or entaxy[1]) of a living system is the entropy that it exports to keep its own entropy low; it lies at the intersection of entropy and life. In other words, negentropy is reverse entropy. It means things becoming more orderly. By 'order' is meant organisation, structure and function: the opposite of randomness or chaos. The concept and phrase "negative entropy" was introduced by Erwin Schrdinger in his 1944 popular-science book What is Life?[2] Later, Lon Brillouin shortened the phrase to negentropy,[3][4] to express it in a more "positive" way: a living system imports negentropy and stores it.[5] In 1974, Albert Szent-Gyrgyi proposed replacing the term negentropy with syntropy. That term may have originated in the 1940s with the Italian mathematician Luigi Fantappi, who tried to construct a unified theory of biology and physics. Buckminster Fuller tried to popularize this usage, but negentropy remains common.
In a note to What is Life? Schrdinger explained his use of this phrase.
In 2009, Mahulikar & Herwig redefined negentropy of a dynamically ordered sub-system as the specific entropy deficit of the ordered sub-system relative to its surrounding chaos.[6] Thus, negentropy has SI units of (J kg1 K1) when defined based on specific entropy per unit mass, and (K1) when defined based on specific entropy per unit energy. This definition enabled: i) scale-invariant thermodynamic representation of dynamic order existence, ii) formulation of physical principles exclusively for dynamic order existence and evolution, and iii) mathematical interpretation of Schrdinger's negentropy debt.
In information theory and statistics, negentropy is used as a measure of distance to normality.[7][8][9] Out of all distributions with a given mean and variance, the normal or Gaussian distribution is the one with the highest entropy. Negentropy measures the difference in entropy between a given distribution and the Gaussian distribution with the same mean and variance. Thus, negentropy is always nonnegative, is invariant by any linear invertible change of coordinates, and vanishes if and only if the signal is Gaussian.
Negentropy is defined as
where S ( x ) {displaystyle S(varphi _{x})} is the differential entropy of the Gaussian density with the same mean and variance as p x {displaystyle p_{x}} and S ( p x ) {displaystyle S(p_{x})} is the differential entropy of p x {displaystyle p_{x}} :
Negentropy is used in statistics and signal processing. It is related to network entropy, which is used in independent component analysis.[10][11]
There is a physical quantity closely linked to free energy (free enthalpy), with a unit of entropy and isomorphic to negentropy known in statistics and information theory. In 1873, Willard Gibbs created a diagram illustrating the concept of free energy corresponding to free enthalpy. On the diagram one can see the quantity called capacity for entropy. This quantity is the amount of entropy that may be increased without changing an internal energy or increasing its volume.[12] In other words, it is a difference between maximum possible, under assumed conditions, entropy and its actual entropy. It corresponds exactly to the definition of negentropy adopted in statistics and information theory. A similar physical quantity was introduced in 1869 by Massieu for the isothermal process[13][14][15] (both quantities differs just with a figure sign) and then Planck for the isothermal-isobaric process.[16] More recently, the MassieuPlanck thermodynamic potential, known also as free entropy, has been shown to play a great role in the so-called entropic formulation of statistical mechanics,[17] applied among the others in molecular biology[18] and thermodynamic non-equilibrium processes.[19]
In 1953, Lon Brillouin derived a general equation[20] stating that the changing of an information bit value requires at least kT ln(2) energy. This is the same energy as the work Le Szilrd's engine produces in the idealistic case. In his book,[21] he further explored this problem concluding that any cause of this bit value change (measurement, decision about a yes/no question, erasure, display, etc.) will require the same amount of energy.
Read more here:
- DotNetNuke Skins - Home - February 20th, 2017 [February 20th, 2017]
- Greydon Square Extropy Lyrics | Genius Lyrics - March 11th, 2017 [March 11th, 2017]
- Extropy - Evernote User Forum - March 11th, 2017 [March 11th, 2017]
- Extropy SHIFT> - March 11th, 2017 [March 11th, 2017]
- TMG 2010 rewriting original host headers when ... - Extropy - July 15th, 2017 [July 15th, 2017]
- Sabretooth Clan | Father Sebastiaan - February 15th, 2018 [February 15th, 2018]
- College of Arts and Humanities The central page for the ... - May 11th, 2018 [May 11th, 2018]
- Knights of Unicron (SG) - Transformers Wiki - TFWiki.net - May 22nd, 2018 [May 22nd, 2018]
- Unicron/Shattered Glass - Transformers Wiki - July 6th, 2018 [July 6th, 2018]
- Knights of Unicron (SG) - Transformers Wiki - July 6th, 2018 [July 6th, 2018]
- Removing Dead Exchange 2010 Servers | www.extropy.com - July 6th, 2018 [July 6th, 2018]
- Exchange 2010 - www.extropy.com - November 11th, 2018 [November 11th, 2018]
- Active Directory, NTP and VMware | www.extropy.com - April 9th, 2019 [April 9th, 2019]
- Markus Guentner: Extropy Album Review | Pitchfork - December 27th, 2021 [December 27th, 2021]
- Whatever Happened to the Transhumanists? - Gizmodo - August 5th, 2022 [August 5th, 2022]