Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Negentropy

From Wikipedia, the free encyclopedia
Measure of distance to normality
Not to be confused withNegative entropy[clarification needed Should this be "For biological contexts..."?].
"Syntropy" redirects here. For other uses, seeSyntropy (software).

Ininformation theory andstatistics,negentropy is used as a measure of distance to normality. The concept and phrase "negative entropy" was introduced byErwin Schrödinger in his 1944 popular-science bookWhat is Life?[1] Later,FrenchphysicistLéon Brillouin shortened the phrase tonéguentropie (negentropy).[2][3] In 1974,Albert Szent-Györgyi proposed replacing the termnegentropy withsyntropy. That term may have originated in the 1940s with the Italian mathematicianLuigi Fantappiè, who tried to construct a unified theory ofbiology andphysics.Buckminster Fuller tried to popularize this usage, butnegentropy remains common.

In a note toWhat is Life? Schrödinger explained his use of this phrase.

... if I had been catering for them [physicists] alone I should have let the discussion turn onfree energy instead. It is the more familiar notion in this context. But this highly technical term seemed linguistically too near toenergy for making the average reader alive to the contrast between the two things.

Information theory

[edit]
This sectionis missing information about the mathematical treatment of negentropy in information theory. Please expand the section to include this information. Further details may exist on thetalk page.(December 2024)

Ininformation theory andstatistics, negentropy is used as a measure of distance to normality.[4][5][6] Out of alldistributions with a given mean and variance, the normal orGaussian distribution is the one with the highestentropy. Negentropy measures the difference in entropy between a given distribution and the Gaussian distribution with the same mean and variance. Thus, negentropy is always nonnegative, is invariant by any linear invertible change of coordinates, and vanishesif and only if the signal is Gaussian.

Negentropy is defined as

J(px)=S(φx)S(px){\displaystyle J(p_{x})=S(\varphi _{x})-S(p_{x})\,}

whereS(φx){\displaystyle S(\varphi _{x})} is thedifferential entropy of the Gaussian density with the samemean andvariance aspx{\displaystyle p_{x}} andS(px){\displaystyle S(p_{x})} is the differential entropy ofpx{\displaystyle p_{x}}:

S(px)=px(u)logpx(u)du{\displaystyle S(p_{x})=-\int p_{x}(u)\log p_{x}(u)\,du}

Negentropy is used instatistics andsignal processing. It is related to networkentropy, which is used inindependent component analysis.[7][8]

The negentropy of a distribution is equal to theKullback–Leibler divergence betweenpx{\displaystyle p_{x}} and a Gaussian distribution with the same mean and variance aspx{\displaystyle p_{x}} (seeDifferential entropy § Maximization in the normal distribution for a proof). In particular, it is always nonnegative.

Correlation between statistical negentropy and Gibbs' free energy

[edit]
Willard Gibbs’ 1873available energy (free energy) graph, which shows a plane perpendicular to the axis ofv (volume) and passing through point A, which represents the initial state of the body. MN is the section of the surface ofdissipated energy. Qε and Qη are sections of the planesη = 0 andε = 0, and therefore parallel to the axes of ε (internal energy) and η (entropy) respectively. AD and AE are the energy and entropy of the body in its initial state, AB and AC itsavailable energy (Gibbs energy) and itscapacity for entropy (the amount by which the entropy of the body can be increased without changing the energy of the body or increasing its volume) respectively.

There is a physical quantity closely linked tofree energy (free enthalpy), with a unit of entropy and isomorphic to negentropy known in statistics and information theory. In 1873,Willard Gibbs created a diagram illustrating the concept of free energy corresponding tofree enthalpy. On the diagram one can see the quantity calledcapacity for entropy. This quantity is the amount of entropy that may be increased without changing an internal energy or increasing its volume.[9] In other words, it is a difference between maximum possible, under assumed conditions, entropy and its actual entropy. It corresponds exactly to the definition of negentropy adopted in statistics and information theory. A similar physical quantity was introduced in 1869 byMassieu for theisothermal process[10][11][12] (both quantities differs just with a figure sign) and by thenPlanck for theisothermal-isobaric process.[13] More recently, the Massieu–Planckthermodynamic potential, known also asfree entropy, has been shown to play a great role in the so-called entropic formulation ofstatistical mechanics,[14] applied among the others in molecular biology[15] and thermodynamic non-equilibrium processes.[16]

J=SmaxS=Φ=klnZ{\displaystyle J=S_{\max }-S=-\Phi =-k\ln Z\,}
where:
S{\displaystyle S} isentropy
J{\displaystyle J} is negentropy (Gibbs "capacity for entropy")
Φ{\displaystyle \Phi } is theMassieu potential
Z{\displaystyle Z} is thepartition function
k{\displaystyle k} theBoltzmann constant

In particular, mathematically the negentropy (the negative entropy function, in physics interpreted as free entropy) is theconvex conjugate ofLogSumExp (in physics interpreted as the free energy).

Brillouin's negentropy principle of information

[edit]

In 1953,Léon Brillouin derived a general equation[17] stating that the changing of an information bit value requires at leastkTln2{\displaystyle kT\ln 2} energy. This is the same energy as the workLeó Szilárd's engine produces in the idealistic case. In his book,[18] he further explored this problem concluding that any cause of this bit value change (measurement, decision about a yes/no question, erasure, display, etc.) will require the same amount of energy.

See also

[edit]

Notes

[edit]
  1. ^Schrödinger, Erwin,What is Life – the Physical Aspect of the Living Cell, Cambridge University Press, 1944
  2. ^Brillouin, Leon: (1953) "Negentropy Principle of Information",J. of Applied Physics, v.24(9), pp. 1152–1163
  3. ^Léon Brillouin,La science et la théorie de l'information, Masson, 1959
  4. ^Aapo Hyvärinen,Survey on Independent Component Analysis, node32: Negentropy, Heli University of Technology Laboratory of Computer and Information Science
  5. ^Aapo Hyvärinen and Erkki Oja,Independent Component Analysis: A Tutorial, node14: Negentropy, Helsinki University of Technology Laboratory of Computer and Information Science
  6. ^Ruye Wang,Independent Component Analysis, node4: Measures of Non-Gaussianity
  7. ^P. Comon, Independent Component Analysis – a new concept?,Signal Processing,36 287–314, 1994.
  8. ^Didier G. Leibovici and Christian Beckmann,An introduction to Multiway Methods for Multi-Subject fMRI experiment, FMRIB Technical Report 2001, Oxford Centre for Functional Magnetic Resonance Imaging of the Brain (FMRIB), Department of Clinical Neurology, University of Oxford, John Radcliffe Hospital, Headley Way, Headington, Oxford, UK.
  9. ^Willard Gibbs,A Method of Geometrical Representation of the Thermodynamic Properties of Substances by Means of Surfaces,Transactions of the Connecticut Academy, 382–404 (1873)
  10. ^Massieu, M. F. (1869a). Sur les fonctions caractéristiques des divers fluides.C. R. Acad. Sci. LXIX:858–862.
  11. ^Massieu, M. F. (1869b). Addition au precedent memoire sur les fonctions caractéristiques.C. R. Acad. Sci. LXIX:1057–1061.
  12. ^Massieu, M. F. (1869),Compt. Rend.69 (858): 1057.
  13. ^Planck, M. (1945).Treatise on Thermodynamics. Dover, New York.
  14. ^Antoni Planes, Eduard Vives,Entropic Formulation of Statistical MechanicsArchived 2008-10-11 at theWayback Machine, Entropic variables and Massieu–Planck functions 2000-10-24 Universitat de Barcelona
  15. ^John A. Scheilman,Temperature, Stability, and the Hydrophobic Interaction,Biophysical Journal73 (December 1997), 2960–2964, Institute of Molecular Biology, University of Oregon, Eugene, Oregon 97403 USA
  16. ^Z. Hens and X. de Hemptinne,Non-equilibrium Thermodynamics approach to Transport Processes in Gas Mixtures, Department of Chemistry, Catholic University of Leuven, Celestijnenlaan 200 F, B-3001 Heverlee, Belgium
  17. ^Leon Brillouin, The negentropy principle of information,J. Applied Physics24, 1152–1163 1953
  18. ^Leon Brillouin,Science and Information theory, Dover, 1956
Look upnegentropy in Wiktionary, the free dictionary.
Retrieved from "https://en.wikipedia.org/w/index.php?title=Negentropy&oldid=1260868587"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp