physical nature of information - weizmann

137
Physical Nature of Information G. Falkovich August 4, 2021 How to receive, send and forget information Contents 1 Thermodynamics and statistical physics 5 1.1 Basic notions ........................... 5 1.2 Legendre transform ........................ 12 1.3 Microcanonical distribution ................... 15 1.4 Canonical distribution and fluctuations ............. 18 2 Appearance of irreversibility 22 2.1 Evolution in the phase space ................... 23 2.2 Kinetic equation and H-theorem ................. 25 2.3 Phase-space mixing and entropy growth ............ 29 2.4 Baker map ............................. 36 2.5 Entropy decrease and non-equilibrium fractal measures .... 39 3 Physics of information 43 3.1 Central limit theorem and large deviations ........... 43 3.2 Information as a choice ...................... 47 3.3 Communication Theory ...................... 50 3.4 Correlations in the signals .................... 53 3.5 Mutual information as a universal tool ............. 55 3.6 Hypothesis testing and relative entropy ............. 61 1

Upload: others

Post on 14-Jan-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Physical Nature of Information - Weizmann

Physical Nature of Information

G. Falkovich

August 4, 2021

How to receive, send and forget information

Contents

1 Thermodynamics and statistical physics 51.1 Basic notions . . . . . . . . . . . . . . . . . . . . . . . . . . . 51.2 Legendre transform . . . . . . . . . . . . . . . . . . . . . . . . 121.3 Microcanonical distribution . . . . . . . . . . . . . . . . . . . 151.4 Canonical distribution and fluctuations . . . . . . . . . . . . . 18

2 Appearance of irreversibility 222.1 Evolution in the phase space . . . . . . . . . . . . . . . . . . . 232.2 Kinetic equation and H-theorem . . . . . . . . . . . . . . . . . 252.3 Phase-space mixing and entropy growth . . . . . . . . . . . . 292.4 Baker map . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 362.5 Entropy decrease and non-equilibrium fractal measures . . . . 39

3 Physics of information 433.1 Central limit theorem and large deviations . . . . . . . . . . . 433.2 Information as a choice . . . . . . . . . . . . . . . . . . . . . . 473.3 Communication Theory . . . . . . . . . . . . . . . . . . . . . . 503.4 Correlations in the signals . . . . . . . . . . . . . . . . . . . . 533.5 Mutual information as a universal tool . . . . . . . . . . . . . 553.6 Hypothesis testing and relative entropy . . . . . . . . . . . . . 61

1

Page 2: Physical Nature of Information - Weizmann

4 Applications of Information Theory 684.1 Distribution from information . . . . . . . . . . . . . . . . . . 694.2 Exorcizing Maxwell demon . . . . . . . . . . . . . . . . . . . . 734.3 Flies and spies . . . . . . . . . . . . . . . . . . . . . . . . . . . 774.4 Rate Distortion and Information Bottleneck . . . . . . . . . . 844.5 Information is money . . . . . . . . . . . . . . . . . . . . . . . 87

5 Renormalization group and entanglement entropy 915.1 Analogy between quantum mechanics and statistical physics . 915.2 Renormalization group and information loss . . . . . . . . . . 935.3 Quantum information . . . . . . . . . . . . . . . . . . . . . . . 1005.4 Thermalization, Area law and Bekenstein bound . . . . . . . . 110

6 Stochastic processes 1136.1 Random walk and diffusion . . . . . . . . . . . . . . . . . . . 1136.2 Brownian motion . . . . . . . . . . . . . . . . . . . . . . . . . 1166.3 General fluctuation-dissipation relation . . . . . . . . . . . . . 1196.4 Stochastic Web surfing and Google’s PageRank . . . . . . . . 126

7 Conclusion 1297.1 Take-home lessons . . . . . . . . . . . . . . . . . . . . . . . . 1297.2 Epilogue . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133

8 Appendix 1: Continuous distributions 135

9 Appendix 2: On Zipf law 136

2

Page 3: Physical Nature of Information - Weizmann

Preface

The book grew out of the one-semester course, initially intended as a partinggift to those leaving physics for greener pastures and wondering what isworth taking with them. Statistically, most of the former physicists usestatistical physics, because this discipline (and this course) answers the mostfrequent question: How much can we do and say about something we do notknow? The course was then taught in different institutions and countries,each time attended by a motley mix of students, postdocs and faculty fromphysics, mathematics, engineering, computer science, economics and biology.Eventually, it evolved into a meeting place where we learn from each otherusing the universal language of information theory, which is a statisticalphysics in disguise, albeit transparent.

The simplest way to answer the above question is called thermodynamics.It is a phenomenology that deals only with visible manifestations of the hid-den, using symmetries and conservation laws to restrict possible outcomesand focusing on mean values ignoring fluctuations. More sophisticated ap-proach derives the statistical laws by explicitly averaging over the hiddendegrees of freedom. Those laws justify thermodynamics and describe theprobabilities of fluctuations.

We first recall basics of thermodynamics and statistical physics and theirdouble focus on what we have (energy) and what we don’t (knowledge).When ignorance exceeds knowledge, the right strategy is to measure igno-rance. Entropy does that. We learn how irreversible entropy change appearsfrom reversible flows in phase space via dynamical chaos. We shall under-stand that entropy is not a property of a system, but of our knowledge ofthe system. It is then natural to use the language of the information theoryrevealing the universality of the approach, which is essentially based on thesimple trick of adding many random numbers. Building on that basis, onedevelops several versatile instruments, of which the mutual information andits quantum sibling, entanglement entropy, are presently most widely appliedto the description of subjects ranging from bacteria and neurons to marketsand quantum computers. We then discuss the so far most sophisticated wayto forget information - renormalization group. Forgetting is a fascinatingactivity — one learns truly fundamental things this way. We end with thestochastic thermodynamics and the generalizations of the second law.

The course uses only elementary tools, but from all three fields of mathe-matics — geometry, algebra and analysis — which correspond respectively

3

Page 4: Physical Nature of Information - Weizmann

to studying space, time and continuum in the physical world. We shall learntwo complementary ways of thinking: continuous flows and discrete combi-natorics (thus employing both brain hemispheres). Together, they produce apowerful and universal tool, applied everywhere, from computer science andmachine learning to biophysics and economics. The course is panoramic,combining into a reasonably coherent whole the subjects that are taughtin much details in different departments: thermodynamics and statisticalmechanics (taught in physics and engineering), dynamical chaos (taught inphysics and applied mathematics), information and communication theories(as taught in computer science and engineering). At the end, recognizing theinformational nature of physics and breaking the barriers of specializationis also of value for physicists. People working on quantum computers andthe entropy of black holes use the same tools as those designing self-drivingcars and market strategies, studying molecular biology, animal behavior andhuman languages, and figuring out how the brain works. Last, I felt com-pelled to tell the story worth telling: how we discover the limits imposed byuncertainty on engines, communications and computations.

Physicists familiar with thermodynamics and statistical physics can startfrom the second Chapter. Readers who are also familiar with the basics ofkinetics and dynamical chaos can directly go to the third Chapter, consultingthe material from the first two when it is referred to. On the other hand,readers from computer science, engineering or mathematics may benefit fromreading the first two chapters as they provide some unifying framework for therest of the course. The notes in full contain the amount of material exceedingthat for a standard one-semester course, so that lecturers can choose whatis more appropriate for their audience. The set of about 30 problems withdetailed solutions will be provided for the problem-solving sessions and theexams.Small-print parts can be omitted upon the first reading.

4

Page 5: Physical Nature of Information - Weizmann

1 Thermodynamics and statistical physics

One can teach monkey to differentiate, integration requires humans.G Kotkin

In any natural science, laws appear usually by induction: from data andparticular cases to a general law and from processes to state functions. Thelatter step requires integration (to pass, for instance, from the Newton equa-tions of mechanics to the Hamiltonian or from thermodynamic equations ofstate to thermodynamic potentials). It is much easier to differentiate than tointegrate, and so deduction (or postulation approach) is usually more simpleand elegant. It also provides a good vantage point for generalizations andappeals to our brain, which likes to hypothesize before receiving any data.In such an approach, one starts from postulating a variational principle forsome function of the state of the system. Then one deduces from that princi-ple the laws that govern changes when one passes from state to state. In thefirst two sections, such a deduction is presented for thermodynamics1, whilethe last two sections serve as a brief reminder of statistical physics.

1.1 Basic notions

Our knowledge is always partial. If we study macroscopic systems, somedegrees of freedom remain hidden. For small sets of atoms or sub-atomicparticles, their quantum nature prevents us from knowing precise values oftheir momenta and coordinates simultaneously. We believe that we found theway around the partial knowledge in mechanics, electricity and magnetism,where we have closed description of the explicitly known degrees of freedom.Even in those cases our knowledge is partial, but we restrict our descriptiononly to things that can be considered independent of the unknown. Forexample, planets are large complex bodies, and yet the motion of their centersof mass in the limit of large distances allows for a closed description of celestialmechanics. Already the next natural problem — how to describe a planetrotation — needs the account of many extra degrees of freedom, such as, forinstance, oceanic flows (which slow down rotation by tidal forces).

Yet even when we have a closed description, our knowledge is partialbecause of a finite precision of our measurements. This has dramatic conse-quences, when there is an instability, so that small variation of initial data

1We follow the spirit of the book H. B. Callen, Thermodynamics (1965).

5

Page 6: Physical Nature of Information - Weizmann

leads to large deviation in evolution. In a sense, every new decimal in preci-sion is a new degree of freedom for an unstable system.

In this course we shall deal with observable manifestations of the hiddendegrees of freedom. While we do not know their state, we do know theirnature, whether those degrees of freedom are related to moving particles,spins, bacteria or market traders. That means that we know the symmetriesand conservation laws of the system. Thermodynamics studies restrictionson the possible macroscopic properties that follow from the symmetries of thefundamental laws. Therefore, thermodynamics does not predict numericalvalues but rather sets inequalities and establishes relations among differentproperties.

Thermodynamics started with physical systems, where the basic sym-metry is invariance of the fundamental laws with respect to time shifts2.Evolution of an isolated physical system is usually governed by the Hamil-tonian (the energy written in canonical variables), whose time-independencemeans energy conservation. Therefore, we wish to ascribe to the states ofthe system the values of the internal energy E. We focus on the states inde-pendent of the way they are prepared. We call such states equilibrium, theyare completely characterized by the static values of observable variables.

Passing from state to state under external action involves the energychange, which generally consists of two parts: the energy change of visibledegrees of freedom (which we shall call work) and the energy change of hiddendegrees of freedom (which we shall call heat). To be able to measure energychanges in principle, we need adiabatic processes where there is no heatexchange. Ascribing to every state its energy (up to an additive constantcommon for all states) hinges on our ability to relate any two equilibriumstates A and B by an adiabatic process either A → B or B → A, whichallows to measure the difference in the internal energies by the work W doneby the system. Now, if we encounter a process where the energy change isnot equal to the work done, we call the difference the heat exchange δQ:

dE = δQ− δW . (1)

This statement is known as the first law of thermodynamics, it reflects ourbelief in energy conservation: if the visible energy balance does not hold then

2Be careful trying to build thermodynamics for biological or social-economic systems,since generally the laws that govern them are not time-invariant. For example, the metabo-lism of the living beings changes with age, and the number of market regulations generallyincreases (as well as the total money mass, albeit not necessarily in our pockets).

6

Page 7: Physical Nature of Information - Weizmann

the energy of the hidden must change. The energy is a function of state sowe use differential, but we use δ for heat and work, which aren’t differentialsof any function. Heat exchange and work depend on the path taken fromA to B, that is they refer to particular forms of energy transfer (not energycontent).

The basic problem of thermodynamics is the determination of the equili-brium state that eventually results after all internal constraints are removedin a closed composite system. The problem is solved with the help of extre-mum principle: there exists a quantity S called entropy which is a functionof the parameters of any composite system. The values assumed by the pa-rameters in the absence of an internal constraint maximize the entropy overthe manifold of constrained equilibrium states (Clausius 1865).

Thermodynamic limit. Traditionally, thermodynamics have dealt withextensive parameters whose value grows linearly with the number of degreesof freedom. Extensive quantities are number of particles N , electric chargeand magnetic moment, etc. Energy usually is extensive in the thermodyn-amic limit. That does not mean that it is additive, that is the energy ofa composite system is not generally the sum of the parts, because of aninteraction energy. To treat energy as an additive variable we make two as-sumptions: i) assume that the forces of interaction are short-range and actonly along the boundary, ii) take thermodynamic limit V → ∞ where onecan neglect surface terms that scale as V 2/3 in comparison with the bulkterms that scale as V .

In that limit, thermodynamic entropy is also an extensive variable3, whichis a homogeneous first-order function of all the extensive parameters:

S(λE, λV, . . .) = λS(E, V, . . .) . (2)

This function (called also fundamental relation) is everything one needs toknow to solve the basic problem (and others) in thermodynamics.

Of course, (2) does not mean that S(E) is a linear function when otherparameters fixed: S(λE, V, . . .) 6= λS(E, V, . . .). On the contrary, we shallsee in a moment that it is a convex function. Nor entropy is necessary amonotonic function of energy4. Yet for every interval of a definite deriva-

3We shall see later that non-extensive parts of entropy are also important.4An example of the two-level system in Section 1.4 shows that S(E) could be non-

monotonic for systems with a finite phase space.

7

Page 8: Physical Nature of Information - Weizmann

tive sign, say (∂E/∂S)X > 0, we can solve S = S(E, V, . . .) uniquely forE(S, V, . . .) which is an equivalent fundamental relation. We assume thefunctions S(E,X) and E(S,X) to be continuous differentiable. An efficientway to treat partial derivatives is to use jacobians

∂(u, v)

∂(x, y)=∂u

∂x

∂v

∂y− ∂v

∂x

∂u

∂y,

(∂u

∂x

)y

=∂(u, y)

∂(x, y).

Then(∂S

∂X

)E

= 0⇒(∂E

∂X

)S

= −∂(ES)

∂(XS)

∂(EX)

∂(EX)= −

(∂S

∂X

)E

(∂E

∂S

)X

= 0 .

Differentiating the last relation one more time we get

(∂2E/∂X2)S = −(∂2S/∂X2)E(∂E/∂S)X ,

since the derivative of the second factor is zero as it is at constant X. Wethus see that in the case (∂E/∂S)X > 0 the equilibrium is defined by theenergy minimum instead of the entropy maximum (very much like circlecan be defined as the figure of either maximal area for a given perimeter orminimal perimeter for a given area).

It is important that the equilibrium curve S(E) is convex, which gua-rantees stability of a homogeneous state. Indeed, if our system would breakspontaneously into two halves with a bit different energies, the entropy mustdecrease: 2S(E) > S(E+ ∆) +S(E−∆) = 2S(E) +S ′′∆2/2, which requiresS ′′ < 0 (that argument does not work for systems with long-range interactionwhere energy is non-additive). On the figure, unconstrained equilibrium sta-tes lie on the curve while all other states lie below. One can reach the stateA either maximizing entropy at a given energy or minimizing energy at agiven entropy:

A

S

E

8

Page 9: Physical Nature of Information - Weizmann

One can work either in energy or entropy representation but ought to becareful not to mix the two.

Experimentally, one usually measures changes thus finding derivatives(called equations of state). The partial derivatives of an extensive varia-ble with respect to its arguments (also extensive parameters) are intensiveparameters5. For example, for the energy one writes

∂E

∂S≡ T (S, V,N) ,

∂E

∂V≡ −P (S, V,N)

∂E

∂N≡ µ(S, V,N) , . . . (3)

These relations are called the equations of state and they serve as definitionsfor temperature T , pressure P and chemical potential µ, corresponding tothe respective extensive variables are S, V,N . We shall see later that en-tropy is the missing information, so that temperature is the energetic priceof information. Our entropy is dimensionless, so that T is assumed to bemultiplied by the Boltzmann constant k = 1.3 ·10−23J/K and have the samedimensionality as the energy. From (3) we write

dE = δQ− δW = TdS − PdV + µdN . (4)

Entropy is thus responsible for hidden degrees of freedom (i.e. heat) whileother extensive parameters describe macroscopic degrees of freedom. We seethat in equilibrium something is maximal for hidden degrees of freedom butthis ”something” is not their energy.

Let us see how the entropy maximum principle solves the basic problem.Consider two simple systems separated by a rigid wall which is impermeablefor anything but heat. The whole composite system is closed that is E1 +E2 =const. The entropy change under the energy exchange,

dS =∂S1

∂E1

dE1 +∂S2

∂E2

dE2 =dE1

T1

+dE2

T2

=(

1

T1

− 1

T2

)dE1 , (5)

must be positive which means that energy flows from the hot subsystem tothe cold one (T1 > T2 ⇒ ∆E1 < 0). We see that our definition (3) is inagreement with our intuitive notion of temperature. When equilibrium isreached, dS = 0 which requires T1 = T2. If fundamental relation is known,then so is the function T (E, V ). Two equations, T (E1, V1) = T (E2, V2) and

5In thermodynamics we have only extensive and intensive variables, because we takethermodynamic limit N →∞, V →∞ keeping N/V finite.

9

Page 10: Physical Nature of Information - Weizmann

E1 + E2 =const completely determine E1 and E2. In the same way one canconsider movable wall and get P1 = P2 in equilibrium. If the wall allows forparticle penetration we get µ1 = µ2 in equilibrium.

Both energy and entropy are homogeneous first-order functions of itsvariables: S(λE, λV, λN) = λS(E, V,N) and E(λS, λV, λN) = λE(S, V,N)(here V and N stand for the whole set of extensive macroscopic parameters).Differentiating the second identity with respect to λ and taking it at λ = 1one gets the Euler equation

E = TS − PV + µN . (6)

The equations of state are homogeneous of zero order, for instance,

T (λE, λV, λN) = T (E, V,N) .

That confirms that the temperature, pressure and chemical potential are thesame for a portion of an equilibrium system as for the whole system. It mayseem that a thermodynamic description of a one-component system requiresoperating functions of three intensive variables. Let us show that the homo-geneity leaves only two independent parameters. For example, the chemicalpotential µ can be found as a function of T and P . Indeed, differentiating(6) and comparing with (4) one gets the so-called Gibbs-Duhem relation (inthe energy representation) Ndµ = −SdT + V dP or for quantities per mole,s = S/N and v = V/N : dµ = −sdT + vdP . In other words, one can chooseλ = 1/N and use first-order homogeneity to get rid of the variable N , for in-stance: E(S, V,N) = NE(s, v, 1) = Ne(s, v). In the entropy representation,

S = E1

T+ V

P

T−N µ

T,

the Gibbs-Duhem relation is again states that because dS = (dE + PdV −µdN)/T then the sum of products of the extensive parameters and the dif-ferentials of the corresponding intensive parameters vanish:

Ed(1/T ) + V d(P/T )−Nd(µ/T ) = 0 . (7)

Processes. While thermodynamics is fundamentally about states it isalso used for describing processes that connect states. Particularly importantquestions concern performance of engines and heaters/coolers. Heat engineworks by delivering heat from a reservoir with some higher T1 via some system

10

Page 11: Physical Nature of Information - Weizmann

to another reservoir with T2 doing some work in the process6. If the entropyof the hot reservoir decreases by some ∆S1 then the entropy of the cold onemust increase by some ∆S2 ≥ ∆S1. The work W is the difference betweenthe heat given by the hot reservoir Q1 = T1∆S1 and the heat absorbed bythe cold one Q2 = T2∆S2 (assuming both processes quasi-static). Engineefficiency is the fraction of heat used for work that is

W

Q1

=Q1 −Q2

Q1

= 1− T2∆S2

T1∆S1

≤ 1− T2

T1

. (8)

T

Q

W

T2

1

1

2Q

One cannot overestimate the historical importance of this simple relation:empirical studies of the limits of engine efficiency lead to the distillation ofthe entropy concept. Indeed, maximal work is achieved for minimal entropychange ∆S2 = ∆S1, which happens for reversible (quasi-static) processes —if, for instance, a gas works by moving a piston then the pressure of thegas and the work are less for a fast-moving piston than in equilibrium. Theefficiency is larger when the temperatures differ more.

Similarly, refrigerator/heater is something that does work to transfer heatfrom cold to hot systems. The performance is characterized by the ratio oftransferred heat to the work done. For the cooler, the efficiency is Q2/W ≤T2/(T1 − T2), for the heater it is Q1/W ≤ T1/(T1 − T2). Now, the efficiencyis large when the temperatures are close, as it requires almost no work totransfer heat.

Summary of formal structure: The fundamental relation is equivalentto the three equations of state (3). If only two equations of state are giventhen Gibbs-Duhem relation may be integrated to obtain the third relationup to an integration constant; alternatively one may integrate molar relationde = Tds − Pdv to get e(s, v), again with an undetermined constant ofintegration.

Example 1: Consider a system that is characterized solely by its energy,which can change between zero and Emax. The equation of state is the energy-temperature relation E/Emax = (1 + eε/T )−1, which tends to 1/2 at T ε and isexponentially small at T ε. In Section 1.3, we shall identify this with a set ofelements with two energy levels, 0 and ε. To find the fundamental relation in the

6Look under the hood of your car to appreciate the level of idealization achieved inthat definition.

11

Page 12: Physical Nature of Information - Weizmann

entropy representation, we integrate the equation of state:

1

T=dS

dE=

1

εlnEmax − E

E⇒ S(E) =

Emaxε

lnEmax

Emax − E+E

εlnEmax − E

E. (9)

Example 2: Consider an ideal monatomic gas characterized by two equationsof state (found, say, experimentally with R ' 8.3 J/mole K ' 2 cal/mole K ):

PV = NRT , E = 3NRT/2 . (10)

The extensive parameters here are E, V,N so we want to find the fundamentalequation in the entropy representation, S(E, V,N). We write (6) in the form

S = E1

T+ V

P

T−N µ

T. (11)

Here we need to express intensive variables 1/T, P/T, µ/T via extensive variables.The equations of state (10) give us two of them:

P

T=NR

V=R

v,

1

T=

3NR

2E=

3R

e. (12)

Now we need to find µ/T as a function of e, v using Gibbs-Duhem relation inthe entropy representation (7). Using the expression of intensive via extensivevariables in the equations of state (12), we compute d(1/T ) = −3Rde/2e2 andd(P/T ) = −Rdv/v2, and substitute into (7):

d

T

)= −3

2

R

ede− R

vdv ,

µ

T= C − 3R

2ln e−R ln v ,

s =1

Te+

P

Tv − µ

T= s0 +

3R

2ln

e

e0+R ln

v

v0. (13)

Here e0, v0 are parameters of the state of zero internal energy used to determine

the temperature units, and s0 is the constant of integration.

1.2 Legendre transform

Let us emphasize that the fundamental relation always relates extensivequantities. Therefore, even though it is always possible to eliminate, say,S from E = E(S, V,N) and T = T (S, V,N) getting E = E(T, V,N), thisis not a fundamental relation and it does not contain all the information.Indeed, E = E(T, V,N) is actually a partial differential equation (becauseT = ∂E/∂S) and even if it can be integrated the result would contain unde-termined function of V,N . Still, it is easier to measure, say, temperature than

12

Page 13: Physical Nature of Information - Weizmann

entropy so it is convenient to have a complete formalism with an intensiveparameter as operationally independent variable and an extensive parameteras a derived quantity. This is achieved by the Legendre transform: We wantto pass from the relation Y = Y (X) to that in terms of P = ∂Y/∂X. Yet it isnot enough to eliminate X and consider the function Y = Y [X(P )] = Y (P ),because such function determines the curve Y = Y (X) only up to a shiftalong X:

X

Y Y

X

For example, the single function Y = P 2/4 correspond to the family offunctions Y = (X+C)2 for arbitrary C. To fix the shift, we specify for everyP the position ψ(P ) where the straight line tangent to the curve interceptsthe Y -axis: ψ = Y − PX:

Y

XP

X

ψ

P

Y = Ψ +

In this way we consider the curve Y (X) as the envelope of the familyof the tangent lines characterized by the slope P and the intercept ψ. Thefunction ψ(P ) = Y [X(P )] − PX(P ) completely defines the curve; here onesubstitutes X(P ) found from P = ∂Y (X)/∂X. The function ψ(P ) is theLegendre transform of Y (X). From dψ = −PdX − XdP + dY = −XdPone gets −X = ∂ψ/∂P i.e. the inverse transform is the same up to a sign:Y = ψ +XP .

The transform is possible when for every X there is one P , that is P (X)is monotonic and Y (X) is convex, ∂P/∂X = ∂2Y/∂X2 6= 0. Sign-definitesecond derivative means that the function is either concave or convex. Thisis the second time we meet convexity, which can be also related to stability.Indeed, for the function E(S), one-to-one correspondence between S andT = ∂E/∂S guarantees uniformity of the temperature across the system.Convexity and concavity will play an important role in this course.

13

Page 14: Physical Nature of Information - Weizmann

Different thermodynamics potentials suitable for different physicalsituations are obtained replacing different extensive parameters by the re-spective intensive parameters.

Free energy F = E − TS (also called Helmholtz potential) is that partialLegendre transform of E which replaces the entropy by the temperature asan independent variable: dF (T, V,N, . . .) = −SdT −PdV + µdN + . . .. It isparticularly convenient for the description of a system in a thermal contactwith a heat reservoir because then the temperature is fixed and we have onevariable less to care about. The maximal work that can be done under aconstant temperature (equal to that of the reservoir) is minus the differentialof the free energy. Indeed, this is the work done by the system and the thermalreservoir. That work is equal to the change of the total energy

d(E + Er) = dE + TrdSr = dE − TrdS = d(E − TrS) = d(E − TS) = dF .

In other words, the free energy F = E − TS is that part of the internalenergy which is free to turn into work, the rest of the energy TS we mustkeep to sustain a constant temperature. The equilibrium state minimizes F ,not absolutely, but over the manifold of states with the temperature equal tothat of the reservoir. Indeed, consider F (T,X) = E[S(T,X), X]−TS(T,X),then (∂E/∂X)S = (∂F/∂X)T that is they turn into zero simultaneously.Also, in the point of extremum, one gets (∂2E/∂X2)S = (∂2F/∂X2)T i.e.both E and F are minimal in equilibrium. Monatomic gas at fixed T,Nhas F (V ) = E − TS(V ) = −NRT lnV+const. If a piston separates equalamounts N , then the work done in changing the volume of a subsystem fromV1 to V2 is ∆F = NRT ln[V2(V − V2)/V1(V − V1)].

Enthalpy H = E+PV is that partial Legendre transform of E which replacesthe volume by the pressure dH(S, P,N, . . .) = TdS + V dP + µdN + . . .. It isparticularly convenient for situation in which the pressure is maintained constantby a pressure reservoir (say, when the vessel is open into atmosphere). Just asthe energy acts as a potential at constant entropy and the free energy as potentialat constant temperature, so the enthalpy is a potential for the work done by thesystem and the pressure reservoir at constant pressure. Indeed, now the reservoirdelivers pressure which can change the volume so that the differential of the totalenergy is

d(E + Er) = dE − PrdVr = dE + PrdV = d(E + PrV ) = d(E + PV ) = dH .

Equilibrium minimizes H under the constant pressure. On the other hand, theheat received by the system at constant pressure (and N) is the enthalpy change:

14

Page 15: Physical Nature of Information - Weizmann

δQ = dQ = TdS = dH. Compare it with the fact that the heat received by thesystem at constant volume (and N) is the energy change since the work is zero.

One can replace both entropy and volume obtaining (Gibbs) thermodynamicspotential G = E−TS+PV which has dG(T, P,N, . . .) = −SdT+V dP+µdN+. . .and is minimal in equilibrium at constant temperature and pressure. From (6) weget (remember, they all are functions of different variables):

F = −P (T, V )V + µ(T, V )N , H = TS + µN , G = µ(T, P )N . (14)

When there is a possibility of change in the number of particles (because our

system is in contact with some particle source having a fixed chemical potential)

then it is convenient to use the grand canonical potential Ω(T, V, µ) = E−TS−µNwhich has dΩ = −SdT − PdV −Ndµ. The grand canonical potential reaches its

minimum under the constant temperature and chemical potential.

Since the Legendre transform is invertible, all potentials are equivalentand contain the same information. The choice of the potential for a givenphysical situation is that of convenience: we usually take what is fixed as avariable to diminish the number of effective variables.

The next two sections present a brief reminder of classical Boltzmann-Gibbs statistical mechanics. Here we introduce microscopic statistical des-cription in the phase space and describe two principal ways (microcanonicaland canonical) to derive thermodynamics from statistical mechanics.

1.3 Microcanonical distribution

Consider a closed system with the fixed number of particles N and the energyE0. Boltzmann assumed that all microstates with the same energy have equalprobability (ergodic hypothesis) which gives the microcanonical distribution:

ρ(p, q) = Aδ[E(p1 . . . pN , q1 . . . qN)− E0] . (15)

Usually one considers the energy fixed with the accuracy ∆ so that the mi-crocanonical distribution is

ρ =

1/Γ for E ∈ (E0, E0 + ∆)0 for E 6∈ (E0, E0 + ∆) ,

(16)

where Γ is the volume of the phase space occupied by the system

Γ(E, V,N,∆) =∫E<H<E+∆

d3Npd3Nq . (17)

15

Page 16: Physical Nature of Information - Weizmann

For example, for N noninteracting particles (ideal gas) the states with theenergy E =

∑p2/2m are in the p-space near the hyper-sphere with the

radius√

2mE. Remind that the surface area of the hyper-sphere with theradius R in 3N -dimensional space is 2π3N/2R3N−1/(3N/2− 1)! and we have

Γ(E, V,N,∆) ∝ E3N/2−1V N∆/(3N/2− 1)! ≈ (E/N)3N/2V N∆ . (18)

To link statistical physics with thermodynamics one must define the fun-damental relation i.e. a thermodynamic potential as a function of respectivevariables. For microcanonical distribution, Boltzmann introduced the en-tropy as

S(E, V,N) = ln Γ(E, V,N) . (19)

This is one of the most important formulas in physics7 (on a par with F =ma ,E = mc2 and E = hω).

Noninteracting subsystems are statistically independent. That meansthat the statistical weight of the composite system is a product - indeed, forevery state of one subsystem we have all the states of another. If the weightis a product then the entropy is a sum. For interacting subsystems, this istrue only for short-range forces in the thermodynamic limit N →∞.

Consider two subsystems, 1 and 2, that can exchange energy. Let’s seehow statistics solves the basic problem of thermodynamics (to define equili-brium) that we treated above in (5). Assume that the indeterminacy in theenergy of any subsystem, ∆, is much less than the total energy E. Then

Γ(E) =E/∆∑i=1

Γ1(Ei)Γ2(E − Ei) . (20)

We denote E1, E2 = E− E1 the values that correspond to the maximal termin the sum (20). To find this maximum, we compute the derivative of it,which is proportional to (∂Γ1/∂Ei)Γ2+(∂Γ2/∂Ei)Γ1 = (Γ1Γ2)[(∂S1/∂E1)E1

−(∂S2/∂E2)E2

]. Then the extremum condition is evidently (∂S1/∂E1)E1=

(∂S2/∂E2)E2, that is the extremum corresponds to the thermal equilibrium

where the temperatures of the subsystems are equal. The equilibrium isthus where the maximum of probability is. It is obvious that Γ(E1)Γ(E2) ≤Γ(E) ≤ Γ(E1)Γ(E2)E/∆. If the system consists of N particles and N1, N2 →∞ then S(E) = S1(E1) +S2(E2) +O(logN) where the last term is negligiblein the thermodynamic limit.

7It is inscribed on the Boltzmann’s gravestone.

16

Page 17: Physical Nature of Information - Weizmann

The same definition (entropy as a logarithm of the number of states) istrue for any system with a discrete set of states. For example, consider theset of N particles (spins, neurons), each with two energy levels 0 and ε. Ifthe energy of the set is E then there are L = E/ε upper levels occupied.The statistical weight is determined by the number of ways one can choose Lout of N : Γ(N,L) = CL

N = N !/L!(N − L)!. We can now define entropy (i.e.find the fundamental relation): S(E,N) = ln Γ. At the thermodynamic limitN 1 and L 1, it gives S(E,N) ≈ N ln[N/(N − L)] + L ln[(N − L)/L],which coincides with (9). The entropy as a function of energy is drawn inthe Figure:

E

0

T=+0

ε

T=

T=−0

N

T=−

S

The entropy is symmetric about E = Nε/2 and is zero at E = 0, Nε whenall the particles are in the same state.. The equation of state (temperature-energy relation) is T−1 = ∂S/∂E ≈ ε−1 ln[(N − L)/L]. We see that whenE > Nε/2 then the population of the higher level is larger than of thelower one (inverse population as in a laser) and the temperature is negative.Negative temperature may happen only in systems with the upper limit ofenergy levels and simply means that by adding energy beyond some levelwe actually decrease the entropy i.e. the number of accessible states. Thatexample with negative temperature is to help you to disengage from theeveryday notion of temperature and to get used to the physicist idea oftemperature as the derivative of energy with respect to entropy.

Available (non-equilibrium) states lie below the S(E) plot. The entropy max-

imum corresponds to the energy minimum for positive temperatures and to the

energy maximum for the negative temperatures. Imagine now that the system

with a negative temperature is brought into contact with the thermostat (having

positive temperature). To equilibrate with the thermostat, the system needs to

acquire a positive temperature. A glance on the figure shows that our system must

give away energy (a laser generates and emits light). If this is done adiabatically

slow, that is along the equilibrium curve, the system first decreases the tempe-

17

Page 18: Physical Nature of Information - Weizmann

rature further until it passes through minus/plus infinity to positive values and

eventually reaches the temperature of the thermostat. That is negative tempera-

tures are actually ”hotter” than positive. By itself though the system is stable

since ∂2S/∂E2 = −N/L(N −L)ε2 < 0 at any temperature. Stress that there is no

volume in S(E,N) that is we consider only subsystem or only part of the degrees

of freedom. Indeed, real particles have kinetic energy unbounded from above and

can correspond only to positive temperatures [negative temperature and infinite

energy give infinite Gibbs factor exp(−E/T )].

The derivation of thermodynamic fundamental relation S(E, . . .) in themicrocanonical ensemble is thus via the number of states or phase volume.

1.4 Canonical distribution and fluctuations

Consider a small subsystem or a system in a contact with a thermostat, whichcan be thought of as consisting of infinitely many copies of our system — thisis so-called canonical ensemble, characterized by N, V, T . Let us derive thecanonical distribution from the microcanonical. Here our system can haveany energy and the question arises what is the probability W (E). Let usfind first the probability of the system to be in a given microstate a with theenergy E. Since all the states of the thermostat are equally likely to occur,then the probability should be directly proportional to the statistical weightof the thermostat Γ0(E0 − E), where we assume E E0, expand (in theexponent!) Γ0(E0−E) = exp[S0(E0−E)] ≈ exp[S0(E0)−E/T )] and obtain

wa(E) = Z−1 exp(−E/T ) , (21)

Z =∑a

exp(−Ea/T ) . (22)

Note that there is no trace of the thermostat left except for the temperature.The normalization factor Z(T, V,N) is a sum over all states accessible to thesystem and is called the partition function.

The probability to have a given energy is the probability of the state (21)times the number of states i.e. the statistical weight of the subsystem:

W (E) = Γ(E)wa(E) = Γ(E)Z−1 exp(−E/T ) . (23)

Here the weight Γ(E) grows with E very fast for large N . But as E → ∞the exponent exp(−E/T ) decays faster than any power. As a result, W (E)is concentrated in a very narrow peak and the energy fluctuations around E

18

Page 19: Physical Nature of Information - Weizmann

are very small. For example, for an ideal gas W (E) ∝ E3N/2 exp(−E/T ).Let us stress again that the Gibbs canonical distribution (21) tells that theprobability of a given microstate exponentially decays with the energy of thestate while (23) tells that the probability of a given energy has a peak.

An alternative and straightforward way to derive the canonical distributionis to use consistently the Gibbs idea of the canonical ensemble as a virtual set,of which the single member is the system under consideration and the energy ofthe total set is fixed. The probability to have our chosen system in the state awith the energy Ea is then given by the average number of systems na in thisstate divided by the total number of systems N . Any set of occupation numbersna = (n0, n1, n2 . . .) satisfies obvious conditions∑

a

na = N ,∑a

Eana = E = εN . (24)

Any given set is realized in Wna = N !/n0!n1!n2! . . . number of ways and theprobability to realize the set is proportional to the respective W :

na =

∑naWna∑Wna

, (25)

where summation goes over all the sets that satisfy (24). We assume that inthe limit when N,na → ∞ the main contribution into (25) is given by the mostprobable distribution that is maximum of W (we actually look at the maximumof lnW which is the same yet technically simpler) under the constraints (24).Using the method of Lagrangian multipliers we look for the extremum of lnW −α∑a na − β

∑aEana. Using the Stirling formula lnn! = n lnn − n we write

lnW = N lnN−∑a na lnna. We thus need to find the value n∗a which corresponds

to the extremum of∑a na lnna−α

∑a na−β

∑aEana. Differentiating we obtain:

lnn∗a = −α− 1− βEa which gives

n∗aN

=exp(−βEa)∑a exp(−βEa)

. (26)

The parameter β is given implicitly by the relation

E

N= ε =

∑aEa exp(−βEa)∑a exp(−βEa)

. (27)

Of course, physically ε(β) is usually more relevant than β(ε).

To get thermodynamics from the Gibbs distribution one needs to definethe free energy because we are under a constant temperature. This is done

19

Page 20: Physical Nature of Information - Weizmann

via the partition function Z (which is of central importance since macroscopicquantities are generally expressed via the derivatives of it):

F (T, V,N) = −T lnZ(T, V,N) . (28)

To prove that, differentiate the identity Z = exp(−F/T ) =∑a exp(−Ea/T )

with respect to temperature, which gives

F = E + T

(∂F

∂T

)V

,

equivalent to F = E − TS in thermodynamics.One can also relate statistics and thermodynamics by defining entropy.

Remind that for a closed system Boltzmann defined S = ln Γ while theprobability of state was wa = 1/Γ. In other words, the entropy was minusthe log of probability. For a subsystem at fixed temperature both energyand entropy fluctuate. What should be the thermodynamic entropy: meanentropy −〈lnwa〉 or entropy at a mean energy lnwa(E)? For a system thathas a Gibbs distribution, lnwa is linear in Ea, so that the entropy at a meanenergy is the mean entropy, and we recover the standard thermodynamicrelation:

S = − 〈lnwa〉 = −∑

wa lnwa =∑

wa(Ea/T + lnZ) (29)

= E/T + lnZ = (E − F )/T = − lnwa(E) = S(E) .

Even though the Gibbs entropy, S = −∑wa lnwa is derived here forequilibrium, this definition can be used for any set of probabilities wa, sinceit provides a useful measure of our ignorance about the system, as we shallsee later.

Are canonical and microcanonical descriptions equivalent? Of course, not.The descriptions are equivalent only when fluctuations are neglected and con-sideration is restricted to mean values. That takes place in thermodynamics,where the distributions just produce different fundamental relations betweenthe mean values: S(E,N) for microcanonical, F (T,N) for canonical, Ω(T, µ)for grand canonical. These relations are related by the Legendre transforms.How operationally one checks, for instance, the equivalence of canonical andmicrocanonical energies? One takes an isolated system at a given energy E,measures the derivative ∂E/∂S, then puts it into the thermostat with thetemperature equal to that ∂E/∂S; the energy now fluctuates but the mean

20

Page 21: Physical Nature of Information - Weizmann

energy must be equal to E (as long as system is macroscopic and all theinteractions are short-range).

As far as fluctuations are concerned, there is a natural hierarchy: microca-nonical distribution neglects fluctuations in energy and number of particles,canonical distribution neglects fluctuations in N but accounts for fluctuationsin E, and eventually grand canonical distribution accounts for fluctuationsboth in E and N .

To describe fluctuations one needs to expand the respective thermodynamic po-tential around the mean value, using the second derivatives ∂2S/∂E2 and ∂2S/∂N2

(which must be negative for stability). That will give Gaussian distributions ofE − E and N − N . Of course, the probability distribution (23) is generally non-Gaussian, but in the thermodynamic limit it can be approximated by Gaussiannot far from the (very sharp) maximum. A straightforward way to find the energy

variance (E − E)2 is to differentiate with respect to β the identity E − E = 0. Forthis purpose one can use canonical distribution and get

∂β

∑a

(Ea − E)eβ(F−Ea) =∑a

(Ea − E)

(F + β

∂F

∂β− Ea

)eβ(F−Ea) − ∂E

∂β= 0 ,

(E − E)2 = −∂E∂β

= T 2CV . (30)

Magnitude of fluctuations is determined by the second derivative of the respectivethermodynamic potential:

∂2S

∂E2=

∂E

1

T= − 1

T 2

∂T

∂E= − 1

T 2CV.

This is natural: the sharper the extremum (the higher the second derivative)the better system parameters are confined to the mean values. Since both Eand CV are proportional to N then the relative fluctuations are small indeed:(E − E)2/E2 ∝ N−1. Note that any extensive quantity f =

∑Ni=1 fi which is a sum

over independent subsystems (i.e. fifk = fifk) have a small relative fluctuation:

(f2 − f2)

f2=

∑(f2i − f2

i )

(∑fi)2

∝ 1

N.

Let us repeat this important distinction: all thermodynamic potentialsare equivalent for the description of mean values but respective statistical dis-tributions are different. System that can exchange energy and particles witha thermostat has its extensive parameters E and N fluctuating and the grand

21

Page 22: Physical Nature of Information - Weizmann

canonical distribution describes those fluctuations. The choice of descriptionis dictated only by convenience in thermodynamics because it treats onlymean values. But if we want to describe the whole statistics of the system inthermostat, we need to use canonical distribution, not the micro-canonicalone. That does not mean that one cannot learn everything about a weaklyfluctuating system in thermal equilibrium by considering it isolated (micro-canonically). Indeed, we can determine CV (and other second derivatives)for an isolated system and then will know the mean squared fluctuation ofenergy when we bring the system into a contact with a thermostat.

2 Appearance of irreversibility

Ou sont les neiges d’antan?Francois Villon

After we recalled thermodynamics and statistical physics, it is time forreflection. The main puzzle here is how irreversible entropy growth appearsout of reversible laws of mechanics. If we screen the movie of any evolutionbackwards, it will be a legitimate solution of the equations of motion. Willit have its entropy decreasing? Can we also decrease entropy by employingthe Maxwell demon who can distinguish fast molecules from slow ones andselectively open a window between two boxes to increase the temperaturedifference between the boxes and thus decrease entropy?

These conceptual questions have been already posed in the 19 century.It took the better part of the 20 century to answer these questions, resolvethe puzzles and make statistical physics conceptually trivial (and technicallymuch more powerful). This required two things: i) better understandingdynamics and revealing the mechanism of randomization called dynamicalchaos, ii) consistent use of the information theory which turned out to bejust another form of statistical physics. This Chapter is devoted to the firstsubject, the next Chapter — to the second one. Here we describe how ir-reversibility and relaxation to equilibrium essentially follows from necessityto consider ensembles (regions in phase space) due to incomplete knowledge.Initially small regions spread over the whole phase space under reversibleHamiltonian dynamics, very much like flows of an incompressible liquid aremixing. Such spreading and mixing in phase space correspond to the appro-ach to equilibrium. On the contrary, to deviate a system from equilibrium,one adds external forcing and dissipation, which makes its phase flow com-

22

Page 23: Physical Nature of Information - Weizmann

pressible and distribution non-uniform. Difference between equilibrium andnon-equilibrium distributions in phase space can then be expressed by thedifference between incompressible and compressible flows.

2.1 Evolution in the phase space

So far we said precious little about how physical systems actually evolve. Letus focus on a broad class of energy-conserving systems that can be describedby the Hamiltonian evolution. Every such system is characterized by itsmomenta p and coordinates q, together comprising the phase space. Wedefine probability for a system to be in some ∆p∆q region of the phasespace as the fraction of time it spends there: w = limT→∞∆t/T . Assumingthat the probability to find it within the volume dpdq is proportional tothis volume, we introduce the statistical distribution in the phase space asdensity: dw = ρ(p, q)dpdq. By definition, the average with the statisticaldistribution is equivalent to the time average:

f =∫f(p, q)ρ(p, q)dpdq = lim

T→∞

1

T

∫ T

0f(t)dt . (31)

The main idea is that ρ(p, q) for a subsystem does not depend on the initialstates of this and other subsystems so it can be found without actually sol-ving equations of motion. We define statistical equilibrium as a state wheremacroscopic quantities are equal to the mean values. Assuming short-rangeforces we conclude that different macroscopic subsystems interact weakly andare statistically independent so that the distribution for a composite systemρ12 is factorized: ρ12 = ρ1ρ2.

Since we usually do not know exactly the coordinates and momenta ofall particles, we consider the ensemble of identical systems starting fromdifferent points in some domain of the phase space. In a flow with thevelocity v = (p, q) the density changes according to the continuity equation:∂ρ/∂t+ div (ρv) = 0. For not very long time, the motion can be consideredconservative and described by the Hamiltonian dynamics: qi = ∂H/∂pi andpi = −∂H/∂qi, so that

∂ρ

∂t=∑i

∂H∂pi

∂ρ

∂qi− ∂H∂qi

∂ρ

∂pi≡ ρ,H .

Here the Hamiltonian generally depends on the momenta and coordinatesof the given subsystem and its neighbors. Hamiltonian flow in the phase

23

Page 24: Physical Nature of Information - Weizmann

space is incompressible, it conserves area in each plane pi, qi and the totalvolume: div v = ∂qi/∂qi + ∂pi/∂pi = 0. That gives the Liouville theorem:dρ/dt = ∂ρ/∂t + (v∇)ρ = −ρdiv v = 0. The statistical distribution is thusconserved along the phase trajectories of any subsystem. As a result, ρ is anintegral of motion and it must be expressed solely via the integrals of motion.Since in equilibrium ln ρ is an additive quantity then it must be expressedlinearly via the additive integrals of motions which for a general mechanicalsystem are momentum P(p, q), the momentum of momentum M(p, q) andenergy E(p, q) (again, neglecting interaction energy of subsystems):

ln ρa = αa + βEa(p, q) + c ·Pa(p, q) + d ·M(p, q) . (32)

Here αa is the normalization constant for a given subsystem while the se-ven constants β, c,d are the same for all subsystems (to ensure additivityof integrals) and are determined by the values of the seven integrals of mo-tion for the whole system. We thus conclude that the additive integrals ofmotion is all we need to get the statistical distribution of a closed system(and any subsystem), those integrals replace all the enormous microscopicinformation. Considering subsystem which neither moves nor rotates we aredown to the single integral, energy, which corresponds to the Gibbs’ canonicaldistribution:

ρ(p, q) = A exp[−βE(p, q)] . (33)

It was obtained for any macroscopic subsystem of a very large system, whichis the same as any system in the contact with thermostat. Note one subtlety:On the one hand, we considered subsystems weakly interacting to have theirenergies additive and distributions independent. On the other hand, preciselythis weak interaction is expected to drive a complicated evolution of anysubsystem, which makes it visiting all regions of the phase space, thus makingstatistical description possible. Particular case of (33) is a microcanonical(constant) distribution, which is evidently invariant under the Hamiltonianevolution of an isolated system due to Liouville theorem.

Assuming that the system spends comparable time in different availablestates (ergodic hypothesis) we conclude that since the equilibrium must bethe most probable state, then it corresponds to the entropy maximum. Inparticular, the canonical equilibrium distribution (33) corresponds to themaximum of the Gibbs entropy, S = −

∫ρ ln ρ dpdq, under the condition

of the given mean energy E =∫ρ(p, q)E(p, q) dpdq. Indeed, requiring zero

variation δ(S + βE) = 0 we obtain (33). For an isolated system with a

24

Page 25: Physical Nature of Information - Weizmann

fixed energy, the entropy maximum corresponds to a uniform micro-canonicaldistribution.

2.2 Kinetic equation and H-theorem

How the system comes to the equilibrium and reaches the entropy maximum?What often causes confusion here is that the dynamics (classical and quan-tum) of any given system is time reversible. The Hamiltonian evolution des-cribed above is an incompressible flow in the phase space, div v = 0, so it con-serves the total Gibbs entropy: dS/dt = −

∫dx ln ρ∂ρ

∂t=∫dx ln ρ div ρv =

−∫dx (v∇)ρ = −

∫dx ρ div v = 0. How then the entropy can grow? Boltz-

mann answered this question by deriving the equation on the one-particlemomentum probability distribution. Such equation must follow from inte-grating the N -particle Liouville equation over all N coordinates and N − 1momenta. Consider the phase-space probability density ρ(x, t) in the spacex = (P,Q), where P = p1 . . .pN and Q = q1 . . .qN. For the system

with the Hamiltonian H =∑ip2i2m

+∑i<j U(qi − qj), the evolution of the

density is described by the following Liouville equation:

∂ρ(P,Q, t)

∂t= ρ(P,Q, t),H =

− N∑i

pi2m

∂qi+∑i<j

θij

ρ(P,Q, t) , (34)

where

θij = θ(qi,pi,qj,pj) =∂U(qi − qj)

∂qi

(∂

∂pi− ∂

∂pj

).

For a reduced description of the single-particle distribution over momentaρ(p, t) =

∫ρ(P,Q, t)δ(p1 − p) dp1 . . . dpNdq1 . . . dqN , we integrate (34):

∂ρ(p, t)

∂t=∫θ(q,p; q′,p′)ρ(q,p; q′,p′) dqdq′dp′ , (35)

This equation is apparently not closed since the rhs contains two-particleprobability distribution. If we write respective equation on that two-particledistribution integrating the Liouville equation over N − 2 coordinates andmomenta, the interaction θ-term brings three-particle distribution, etc. Con-sistent procedure is to assume a short-range interaction and a low density,so that the mean distance between particles much exceeds the radius of inte-raction. In this case we may assume for every binary collision that particles

25

Page 26: Physical Nature of Information - Weizmann

come from large distances and their momenta are not correlated. Statisticalindependence then allows one to replace the two-particle momenta distribu-tion by the product of one-particle distributions.

Such derivation is cumbersome,8 but it is easy to write the general formthat such a closed equation must have. For a dilute gas, only two-particlecollisions need to be taken into account in describing the evolution of thesingle-particle distribution over moments ρ(p, t). Consider the collision oftwo particles having momenta p,p1: p

p1 1p

p

For that, they must come to the same place, yet we shall assume that theparticle velocity is independent of the position and that the momenta of twoparticles are statistically independent, that is the probability is the productof single-particle probabilities: ρ(p,p1) = ρ(p)ρ(p1). These very strong as-sumptions constitute what is called the hypothesis of molecular chaos. Undersuch assumptions, the number of such collisions (per unit time per unit vo-lume) must be proportional to probabilities ρ(p)ρ(p1) and depend both oninitial momenta p, p1 and the final ones p′, p′1:

w(p,p1; p′,p′1)ρ(p)ρ(p1) dpdp1dp′dp′1 . (36)

One may believe that (36) must work well when the distribution functionevolves on a time scale much longer than that of a single collision. Weassume that the medium is invariant with respect to inversion r→ −r whichgives the detailed equilibrium:

w ≡ w(p,p1; p′,p′1) = w(p′,p′1; p,p1) ≡ w′ . (37)

We can now write the rate of the probability change as the difference betweenthe number of particles coming and leaving the given region of phase spacearound p by integrating over all p1p

′p′1:

∂ρ

∂t=

∫(w′ρ′ρ′1 − wρρ1) dp1dp

′dp′1 . (38)

We now use the probability normalization which states the sum of transitionprobabilities over all possible states, either final or initial, is unity and so the

8see e.g. http://www.damtp.cam.ac.uk/user/tong/kintheory/kt.pdf

26

Page 27: Physical Nature of Information - Weizmann

sums are equal to each other:∫w(p,p1; p′,p′1) dp′dp′1 =

∫w(p′,p′1; p,p1) dp′dp′1 . (39)

Using (39) we transform the second term (38) and obtain the famous Boltz-mann kinetic equation (1872):

∂ρ

∂t=∫w′(ρ′ρ′1 − ρρ1) dp1dp

′dp′1 ≡ I , (40)

H-theorem. Let us look at the evolution of the entropy

dS

dt= −

∫ ∂ρ

∂tln ρ dp = −

∫I ln ρ dp , (41)

The integral (41) contains the integrations over all momenta so we may ex-ploit two interchanges, p1 ↔ p and p,p1 ↔ p′,p′1:

dS

dt=∫w′(ρρ1 − ρ′ρ′1) ln ρ dpdp1dp

′dp′1

=1

2

∫w′(ρρ1 − ρ′ρ′1) ln(ρρ1) dpdp1dp

′dp′1

=1

2

∫w′ρρ1 ln

ρρ1

ρ′ρ′1dpdp1dp

′dp′1 ≥ 0 , (42)

Here we may add the integral∫w′(ρρ1 − ρ′ρ′1) dpdp1dp

′dp1/2 = 0 and thenuse the inequality x lnx− x+ 1 ≥ 0 with x = ρρ1/ρ

′ρ′1.Even if we use scattering probabilities obtained from mechanics reversible

in time, w(−p,−p1;−p′,−p′1) = w(p′,p′1; p,p1), our use of molecular chaoshypothesis made the kinetic equation irreversible. Equilibrium realizes theentropy maximum and so the distribution must be a steady solution of theBoltzmann equation. Indeed, the collision integral turns into zero by virtueof ρ0(p)ρ0(p1) = ρ0(p′)ρ0(p′1), since ln ρ0 is the linear function of the inte-grals of motion as was explained in Sect. 2.1. All this is true also for theinhomogeneous equilibrium in the presence of an external force.

One can look at the transition from (34) to (40) from a temporal view-point. N -particle distribution changes during every collision when particlesexchange momenta. On the other hand, changing the single-particle distribu-tion requires many collisions. In a dilute system with short-range interaction,the collision time is much shorter than the time between collisions, so thetransition is from a fast-changing function to a slow-changing one.

27

Page 28: Physical Nature of Information - Weizmann

Let us summarize the present state of confusion. The full entropy of theN -particle distribution is conserved. Yet the one-particle entropy grows. Isthere a contradiction here? Is not the full entropy a sum of one-particleentropies? The answer (”no” to both questions) requires introduction of thecentral notion of this course - mutual information - and will be given inSection 3.6 below. For now, a brief statement will suffice: We broke timereversibility when we assumed particles uncorrelated before the collision. Ifone starts from a set of uncorrelated particles and let them interact, then theinteraction will build correlations and the total distribution will change, butthe total entropy will not. However, the single-particle entropy will generallygrow, since the Boltzmann equation is valid for an uncorrelated initial state(and for some time after). Motivation for choosing such an initial statefor computing one-particle evolution is that it is most likely in any genericensemble. Yet that would make no sense to run the Boltzmann equationbackwards from a correlated state, which is statistically a very unlikely initialstate, since it requires momenta to be correlated in such a way that a definitestate is produced after time t. In other words, Boltzmann equation describesat a macroscopic level (of one-particle distribution) not all but most of themicroscopic (N -particle) evolutions.

One may think that for dilute systems there must be a regular expansions of dif-ferent quantities in powers of density. In particular, molecular chaos factorizationof two-particle distribution ρ12 = ρ(q1,p1;q2,p2) via one-particle distributionsρ1 and ρ2 is expected to be just the first term of such an expansion:

ρ12 = ρ1ρ2 +

∫dq3dp3J123ρ1ρ2ρ3 + . . . .

In reality such (so-called cluster) expansion is well-defined only for equilibrium

distributions. For non-equilibrium distributions, starting from some term (depen-

ding on the space dimensionality), all higher terms diverge. The same divergencies

take place if one tries to apply the expansion to kinetic coefficients like diffusivity,

conductivity or viscosity, which are non-equilibrium properties by their nature.

These divergencies can be related to the fact that non-equilibrium distributions do

not fill the phase space, as described below in Section 2.5. Obtaining finite results

requires re-summation and brings logarithmic terms. As a result, kinetic coeffi-

cients and other non-equilibrium properties are non-analytic functions of density.

Boltzmann equation looks nice, but corrections to it are ugly, when one deviates

from equilibrium.

28

Page 29: Physical Nature of Information - Weizmann

2.3 Phase-space mixing and entropy growth

We have seen that one-particle entropy can grow even when the full N -particle entropy is conserved. But thermodynamics requires the full entropyto grow. To accomplish that, let us return to the full N -particle distributionand recall that we have an incomplete knowledge of the system. That meansthat we always measure coordinates and momenta within some intervals, i.e.characterize the system not by a point in phase space but by a finite regionthere. We shall see that quite general dynamics stretches this finite domaininto a very thin convoluted strip whose parts can be found everywhere inthe available phase space, say on a fixed-energy surface. The dynamics thusprovides a stochastic-like element of mixing in phase space that is responsiblefor the approach to equilibrium, say to uniform microcanonical distribution.Yet by itself this stretching and mixing does not change the phase volumeand entropy. Another ingredient needed is the necessity to continually treatour system with finite precision, which follows from the insufficiency of in-formation. Such consideration is called coarse graining and it, together withmixing, it is responsible for the irreversibility of statistical laws and for theentropy growth.

The dynamical mechanism of the entropy growth is the separation oftrajectories in phase space so that trajectories started from a small neig-hborhood are found in larger and larger regions of phase space as timeproceeds. Denote again by x = (P,Q) the 6N -dimensional vector of theposition and by v = (P, Q) the velocity in the phase space. The relativemotion of two points, separated by r, is determined by their velocity diffe-rence: δvi = rj∂vi/∂xj = rjσij. We can decompose the tensor of velocityderivatives into an antisymmetric part (which describes rotation) and a sym-metric part Sij = (∂vi/∂xj + ∂vj/∂xi)/2 (which describes deformation). Weare interested here in deformation because it is the mechanism of the entropygrowth. The vector initially parallel to the axis j turns towards the axis iwith the angular speed ∂vi/∂xj, so that 2Sij is the rate of variation of theangle between two initially mutually perpendicular small vectors along i andj axes. In other words, 2Sij is the rate with which rectangle deforms intoparallelograms: S

Sδyx

yxy

δy

δx

δx

29

Page 30: Physical Nature of Information - Weizmann

texp(S t)

exp(S t)xx

yy

Figure 1: Deformation of a phase-space element by a permanent strain.

Arrows in the Figure show the velocities of the endpoints. The symmetrictensor Sij can be always transformed into a diagonal form by an orthogonaltransformation (i.e. by the rotation of the axes), so that Sij = Siδij. Accor-ding to the Liouville theorem, a Hamiltonian dynamics is an incompressibleflow in the phase space, so that the trace of the tensor, which is the rateof the volume change, must be zero: Trσij =

∑i Si = div v = 0 — that

some components are positive, some are negative. Positive diagonal compo-nents are the rates of stretching and negative components are the rates ofcontraction in respective directions. Indeed, the equation for the distancebetween two points along a principal direction has a form: ri = δvi = riSi .The solution is as follows:

ri(t) = ri(0) exp[∫ t

0Si(t

′) dt′]. (43)

For a time-independent strain, the growth/decay is exponential in time. Onerecognizes that a purely straining motion converts a spherical element into anellipsoid with the principal diameters that grow (or decay) in time. Indeed,consider a two-dimensional projection of the initial spherical element i.e. a

circle of the radius R at t = 0. The point that starts at x0, y0 =√R2 − x2

0

goes into

x(t) = eS11tx0 ,

y(t) = eS22ty0 = eS22t√R2 − x2

0 = eS22t√R2 − x2(t)e−2S11t ,

x2(t)e−2S11t + y2(t)e−2S22t = R2 . (44)

The equation (44) describes how the initial circle turns into the ellipse whoseeccentricity increases exponentially with the rate |S11 − S22|. In a multi-dimensional space, any sphere of initial conditions turns into the ellipsoiddefined by

∑6Ni=1 x

2i (t)e

−2Sit =const.Of course, as the system moves in the phase space, both the strain va-

lues and the orientation of the principal directions change, so that expanding

30

Page 31: Physical Nature of Information - Weizmann

direction may turn into a contracting one and vice versa. Since we do notwant to go into details of how the system interacts with the environment,then we consider such evolution as a kind of random process. The questionis whether averaging over all values and orientations gives a zero net result.It may seem counter-intuitive at first, but in a general case an exponentialstretching persists on average and the majority of trajectories separate. Phy-sicists think in two ways: one in space and another in time (unless they arerelativistic and live in a space-time)9.

Let us first look at separation of trajectories from a temporal perspective,going with the flow: even when the average rate of separation along a givendirection, Λi(t) =

∫ t0 Si(t

′)dt′/t, is zero, the average exponent of it is largerthan unity (and generally growing with time):

limt→∞

∫ t

0Si(t

′)dt′ = 0 , limT→∞

1

T

∫ T

0dt exp

[∫ t

0Si(t

′)dt′]≥ 1 . (45)

This is because the intervals of time with positive Λ(t) give more contributioninto the exponent than the intervals with negative Λ(t). That follows fromthe concavity of the exponential function. In the simplest case, when Λ isuniformly distributed over the interval −a < Λ < a, the average Λ is zero,while the average exponent is (1/2a)

∫−aa eΛdΛ = (ea − e−a)/2a > 1.

Looking from a spatial perspective, consider the simplest flow field: two-dimensional10 pure strain, which corresponds to an incompressible saddle-point flow: vx = λx, vy = −λy. Here we have one expanding direction andone contracting direction, their rates being equal. The vector r = (x, y)(the distance between two close trajectories) can look initially at any di-rection. The evolution of the vector components satisfies the equationsx = vx and y = vy. Whether the vector is stretched or contracted aftersome time T depends on its orientation and on T . Since x(t) = x0 exp(λt)and y(t) = y0 exp(−λt) = x0y0/x(t) then every trajectory is a hyperbole. Aunit vector initially forming an angle ϕ with the x axis will have its length[cos2 ϕ exp(2λT )+sin2 ϕ exp(−2λT )]1/2 after time T . The vector is stretchedif cosϕ ≥ [1 + exp(2λT )]−1/2 < 1/

√2, i.e. the fraction of stretched directi-

ons is larger than half. When along the motion all orientations are equally

9”Time and space are modes by which we think and not conditions in which we live”A. Einstein

10Two-dimensional phase space corresponds to the trivial case of one particle movingalong a line, yet it is great illustrative value. Also, remember that the Liouville theoremis true in every pi − qi plane projection.

31

Page 32: Physical Nature of Information - Weizmann

x

y

xx(T)x(0)

y

y(0)

y(T)ϕ0

Figure 2: Left panel: streamlines of the saddle-point flow. Right panel:motion along a streamline. For ϕ = ϕ0 the initial and final points aresymmetric relative to the diagonal: x(0) = y(T ) and y(0) = x(T ). Ifϕ < ϕ0 = arccos[1 + exp(2λT )]−1/2 > π/4, then the distance from the originincreases.

probable, the net effect is stretching, increasing with the persistence time T .

The net stretching and separation of trajectories is formally proved in mat-hematics by considering random strain matrix σ(t) and the transfer matrix Wdefined by r(t) = W (t, t1)r(t1). It satisfies the equation dW/dt = σW . The Liou-ville theorem tr σ = 0 means that det W = 1. The modulus r(t) of the separationvector may be expressed via the positive symmetric matrix W T W . The main re-sult (Furstenberg and Kesten 1960; Oseledec, 1968) states that in almost everyrealization σ(t), the matrix 1

t ln W T (t, 0)W (t, 0) tends to a finite limit as t→∞.In particular, its eigenvectors tend to d fixed orthonormal eigenvectors fi. Geo-metrically, that precisely means than an initial sphere evolves into an elongatedellipsoid at later times. The limiting eigenvalues

λi = limt→∞

t−1 ln |W fi| (46)

define the so-called Lyapunov exponents, which can be thought of as the mean

stretching rates. The sum of the exponents is zero due to the Liouville theorem

so there exists at least one positive exponent which gives stretching. Therefore,

as time increases, the ellipsoid is more and more elongated and it is less and less

likely that the hierarchy of the ellipsoid axes will change. Mathematical lesson to

learn is that multiplying N random matrices with unit determinant (recall that

determinant is the product of eigenvalues), one generally gets some eigenvalues

growing and some decreasing exponentially with N . It is also worth remembering

that in a random flow there is always a probability for two trajectories to come

32

Page 33: Physical Nature of Information - Weizmann

closer. That probability decreases with time but it is finite for any finite time.

In other words, majority of trajectories separate but some approach. The separa-

ting ones provide for the exponential growth of positive moments of the distance:

E(a) = limt→∞ t−1 ln [〈ra(t)/ra(0)〉] > 0 for a > 0. However, approaching trajec-

tories have r(t) decreasing, which guarantees that the moments with sufficiently

negative a also grow. Mention without proof that E(a) is a concave function,

which evidently passes through zero, E(0) = 0. It must then have another zero

which for isotropic random flow in d-dimensional space can be shown to be a = −d,

see home exercise.

The probability to find a ball turning into an exponentially stretchingellipse thus goes to unity as time increases. The physical reason for it is thatsubstantial deformation appears sooner or later. To reverse it, one needs tocontract the long axis of the ellipse, that is the direction of contraction mustbe inside the narrow angle defined by the ellipse eccentricity, which is lesslikely than being outside the angle:

contracting direction

must be within the angle

To transform ellipse to circle,

This is similar to the argument about the irreversibility of the Boltz-mann equation in the previous subsection. Randomly oriented deformationson average continue to increase the eccentricity. Drop ink into a glass ofwater, gently stir (not shake) and enjoy the visualization of Furstenberg andOseledets theorems.

Armed with the understanding of the exponential stretching, we nowreturn to the dynamical foundation of the second law of thermodynamics.We assume that our finite resolution does not allow us to distinguish betweenthe states within some square in the phase space. That square is our ”grain”in coarse-graining. In the figure below, one can see how such black square ofinitial conditions (at the central box) is stretched in one (unstable) directionand contracted in another (stable) direction so that it turns into a long narrowstrip (left and right boxes). Later in time, our resolution is still restricted- rectangles in the right box show finite resolution (this is coarse-graining).Viewed with such resolution, our set of points occupies larger phase volumeat t = ±T than at t = 0. Larger phase volume corresponds to larger entropy.Time reversibility of any trajectory in the phase space does not contradict thetime-irreversible filling of the phase space by the set of trajectories consideredwith a finite resolution. By reversing time we exchange stable and unstable

33

Page 34: Physical Nature of Information - Weizmann

q

pp

q

p

qt=−T t=0 t=T

Figure 3: Increase of the phase volume upon stretching-contraction andcoarse-graining. Central panel shows the initial state and the velocity field.

directions (i.e. those of contraction and expansion), but the fact of spacefilling persists. We see from the figure that the volume and entropy increaseboth forward and backward in time. And yet our consideration does providefor time arrow: If we already observed an evolution that produces a narrowstrip then its time reversal is the contraction into a ball; but if we considera narrow strip as an initial condition, it is unlikely to observe a contractionbecause of the narrow angle mentioned above. Therefore, being shown twomovies, one with stretching, another with contraction we conclude that withprobability close (but not exactly equal!) to unity the first movie shows thetrue sequence of events, from the past to the future.

When the density spreads, entropy grows (as the logarithm of the volumeoccupied). If initially our system was within the phase-space volume ε6N ,then its density was ρ0 = ε−6N inside and zero outside. After stretching tosome larger volume eλtε6N the entropy S = −

∫ρ ln ρdx has increased by λt.

The positive Lyapunov exponent λ determines the rate of the entropy growth.If in a d-dimensional space there are k stretching and d − k contractingdirections, then contractions eventually stabilize at the resolution scale, whileexpansions continue. Therefore, the volume growth rate is determined by thesum of the positive Lyapunov exponents

∑ki=1 λi.

We shall formally define information later, here we use everyday intuitionabout it to briefly discuss our flow from this perspective. Consider an ensem-ble of systems having close initial positions within our finite resolution. Ina flow with positive Lyapunov exponents, with time we loose our ability topredict where it goes. This loss of information is determined by the growthof the available phase volume, that is of the entropy. But we can look bac-kwards in time and ask where the points come from. The two points along

34

Page 35: Physical Nature of Information - Weizmann

a stretching direction that were hidden inside the resolution circle separatewith time and can be distinguished:

Moreover, as time proceeds, we learn more and more about the initiallocations of the points. The acquisition rate of such information about thepast is again the sum of the positive Lyapunov exponents and is called theKolmogorov-Sinai entropy. As time lag from the present moment increases,we can say less and less where we shall be and more and more where we camefrom. It illustrates the Kierkegaard’s remark that the irony of life is that itis lived forward but understood backwards.

After the strip length reaches the scale of the velocity change (when onealready cannot approximate the phase-space flow by a linear profile σr),strip starts to fold because rotation (which we can neglect for a ball but notfor a long strip) is different at different parts of the strip. Still, howeverlong, the strip continues locally the exponential stretching. Eventually, onecan find the points from the initial ball everywhere which means that theflow is mixing, also called ergodic. Formal definition is that the flow iscalled ergodic in the domain if the trajectory of almost every point (exceptpossibly a set of zero volume) passes arbitrarily close to every other point. Anequivalent definition is that there are no finite-volume subsets of the domaininvariant with respect to the flow except the domain itself. Ergodic flow on anenergy surface in the phase space provides for a micro-canonical distribution(i.e. constant), since time averages are equivalent to the average over thesurface. While we can prove ergodicity only for relatively simple systems,like the gas of hard spheres, we believe that it holds for most systems ofsufficiently general nature (that vague notion can be make more precise bysaying that the qualitative systems behavior is insensitive to small variationsof its microscopic parameters).

At even larger time scales than the time of the velocity change for a trajec-tory, one can consider the motion as a series of uncorrelated random steps. Thatproduces random walk considered in detail in Sect 6.1 below, where we will showthat the spread of the probability density ρ(r, t) is described by a simple diffu-sion: ∂ρ/∂t = κ∆ρ. The total probability

∫ρ(r, t) dr is conserved but the entropy

35

Page 36: Physical Nature of Information - Weizmann

increases monotonically under diffusion:

dS

dt= − d

dt

∫ρ(r, t) ln ρ(r, t) dr = −κ

∫∆ρ ln ρ dr = κ

∫(∇ρ)2

ρdr ≥ 0 . (47)

Asymptotically in time the solution of the diffusion equation takes the universal

form ρ(r, t) = (4πκt)−d/2exp(−r2/4κt

), see (155) below; substituting it into (47)

we obtain a universal entropy production rate, dS/dt = 1/2t, independent of κ

(which is clear from dimensional reasoning).

Two concluding remarks are in order. First, the notion of an exponentialseparation of trajectories put an end to the old dream of Laplace to be ableto predict the future if only all coordinates and momenta are given. Evenif we were able to measure all relevant phase-space initial data, we can doit only with a finite precision ε. However small is the indeterminacy in thedata, it is amplified exponentially with time so that eventually ε exp(λT ) islarge and we cannot predict the outcome. Mathematically speaking, limitsε→ 0 and T → ∞ do not commute. Second, the above arguments did notuse the usual mantra of thermodynamic limit, which means that even thesystems with a small number of degrees of freedom need statistics for theirdescription at long times if their dynamics has a positive Lyapunov exponent(which is generic) - this is sometimes called dynamical chaos.11

2.4 Baker map

We consider a toy model of great illustrative value for the applications ofchaos theory to statistical mechanics. Take the phase-space to be a unitsquare in the (x, y)-plane, with 0 < x, y < 1. The measure-preserving trans-formation is an expansion in the x-direction and a contraction in the y-direction, arranged in such a way that the unit square is mapped onto itselfat each step. The transformation consists of two steps: First, the unit squareis contracted in the y-direction and stretched in the x-direction by a factorof 2. This doesn’t change the volume of any initial region. The unit square

11As a student, I’ve participated (as a messenger) in the discussion on irreversibilitybetween Zeldovich and Sinai. I remember Zeldovich asking why coarse-graining alone(already introduced by Boltzmann) is not enough to explain irreversibility. Why oneneeds dynamical chaos to justify what one gets by molecular chaos? I believe that Sinaiwas right promoting separation of trajectories. It replaces arbitrary assumptions by cleardemonstration from first principles, which is conceptually important, even though possiblein idealized cases only.

36

Page 37: Physical Nature of Information - Weizmann

becomes a rectangle occupying the region 0 < x < 2; 0 < y < 1/2. Next, therectangle is cut in the middle and the right half is put on top of the left halfto recover a square. This doesn’t change volume either. That way bakersprepare long thin strips of pasta. This transformation is reversible except onthe lines where the area was cut in two and glued back.

If we consider two initially closed points, then after n such steps thedistance along x and y will be multiplied respectively by 2n = en ln 2 and2−n = e−n ln 2. It is then easy to see, without a lot of formalities, that there aretwo Lyapunov exponents corresponding to the discrete time n. One of themis connected to the expanding direction and has the value λ+ = ln 2. Theother Lyapunov exponent is connected to the contracting direction and hasthe value λ− = − ln 2. For the forward time operation of the baker’s trans-formation, the expanding direction is along the x-axis, and the contractingdirection is along the y-axis. If one considers the time-reversed motion, theexpanding and contracting directions change places. Therefore, for the for-ward motion nearby points separated only in the y-direction approach eachother exponentially rapidly with the rate λ− = − ln 2. In the x-direction,points separate exponentially with λ+ = ln 2. The sum of the Lyapunovexponents is zero, which reflects the fact that the baker’s transformation isarea-preserving.

Let us argue now that the baker transformation is mixing, that is sprea-ding the measure uniformly over the whole phase space. Indeed, if a measureis initially concentrated in any domain, as in the grey area in the left panelof the Figure 4, after sufficiently long time the domain is transformed into alarge number of very thin horizontal strips of length unity, distributed moreand more uniformly in the vertical direction. Eventually any set in the unitsquare will have the same fraction of its area occupied by these little stripsof pasta as any other set. This is the indicator of a mixing system. Weconclude that a sufficiently smooth initial distribution function defined onthe unit square will approach a uniform (microcanonical) distribution on the

37

Page 38: Physical Nature of Information - Weizmann

Figure 4: Left panel: iterations of the baker map. Right panel: toral map.

square.To avoid impression that cutting and gluing of the baker map are ne-

cessary for mixing, consider a smooth model which has similar behavior.Namely, consider a unit two-dimensional torus, that is unit square with peri-odic boundary conditions, so that all distances are measured modulo 1 in thex- and y-direction. The transformation matrix T (an analog of the transfermatrix W from the previous section) maps unit torus into itself if a, b, c, dare all integers. The action of such map is shown in the right panel of the

Figure 4. The eigenvalues λ1,2 = (a+ d)/2±√

(a− d)2/4 + bc are real when

(a − d)2/4 + bc ≥ 0, in particular, when the matrix is symmetric. For thetransform to be area-preserving, the determinant of the matrix T , that isthe product of the eigenvalues must be unity: λ1λ2 = ad − bc = 1. In ageneral case, one eigenvalue is larger than unity and one is smaller, whichcorresponds respectively to positive and negative Lyapunov exponents lnλ1

and lnλ2.Baker map is area-preserving and does not change entropy, yet when we

allow for coarse-graining along with the evolutions, then the entropy growsand eventually reaches the maximum, which is the logarithm of the phasevolume, which corresponds the equilibrium microcanonical distribution.

38

Page 39: Physical Nature of Information - Weizmann

2.5 Entropy decrease and non-equilibrium fractal me-asures

As we have seen in the previous two sections, if we have indeterminacy inthe data or consider an ensemble of systems, then Hamiltonian dynamics (anincompressible flow) effectively mixes and makes distribution uniform in thephase space. Since we have considered isolated systems, they conserve theirintegrals of motion, so that the distribution is uniform over the respectivesurface. In particular, dynamical chaos justifies micro-canonical distribution,uniform over the energy surface.

But what if the dynamics is non-Hamiltonian, that is Liouville theoremis not valid? The flow in the phase space is then generally compressible.For example, we accelerate particles by external forces fi and damp theirmomenta with the dissipation rates γi, so that the equations of motion takethe form: pi = fi − γipi − ∂H/∂qi, qi = ∂H/∂pi, which gives generallydiv v =

∑i(∂fi/∂pi − γi) 6= 0. Let us show that such flows create quite

different distribution. Since div v 6= 0, then the probability density generallychanges along a flow: dρ/dt = −ρdiv v. That produces entropy,

dS

dt=∫ρ(r, t)div v(r, t) dr = 〈ρdiv v〉 . (48)

with the rate equal to the Lagrangian mean of the phase-volume local ex-pansion rate. If the system does not on average heats or cools (expands orcontracts), then the whole phase volume does not change. That means thatthe global average (over the whole volume) of the local expansion rate iszero: 〈div v〉 =

∫div v dr = 0. Yet for a non-uniform density, the entropy is

not the log of the phase volume but the minus mean log of the phase den-sity, S = −〈ρ ln ρ〉, whose derivative (48) is non-zero because of correlationsbetween ρ and div v. Since ρ is always smaller in the expanding regionswhere div v > 0, then the entropy production rate (48) is non-positive. Weconclude that the mean logarithm of the density (i.e. entropy) decreases.Since the uniform distribution has a maximal entropy under the condition offixed normalization, then the entropy decrease means that the distributionis getting more non-uniform.

What happens then to the density? Of course, if we integrate densityover all the phase space we obtain unity at any time: 〈ρ〉 =

∫ρ(r, t) dr = 1.

Let us now switch focus from space to time and consider the density of an

39

Page 40: Physical Nature of Information - Weizmann

arbitrary fluid element, which evolves as follows:

ρ(t)/ρ(0) = exp[−∫ t

0div v(t′) dt′

]= eC(t) . (49)

As we have seen in (45), if a mean is zero, the mean exponent generallyexceeds unity because of concavity of the exponential function. Now thecontraction factor averaged over the whole flow is zero at any time, 〈C〉 = 0,and its average exponent is larger than unity:

〈ρ(t)/ρ(0)〉 = limT→∞

1

T

∫ T

0dt exp

[−∫ t

0div v(t′) dt′

]= 〈eC〉 > 1 .

That concavity simply means that the parts of the flow with positive C givemore contribution into the exponent than the parts with negative C. More-over, for a generic random flow the density of most fluid elements must grownon-stop as they move. Indeed, if the Lagrangian quantity (taken in the flowreference frame) div v(r, t) is random function with a finite correlation time,then at longer times its integral

∫ t0 div v(t′) dt′ is Gaussian with zero mean

and variance linearly growing with time (see section 3.1). Since the total me-asure is conserved, growth of density at some places must be compensatedby its decrease in other places, so that the distribution is getting more andmore non-uniform, which decreases the entropy. Looking at the phase spaceone sees it more and more emptied with the density concentrated asymptoti-cally in time on a fractal set. That is opposite to the mixing by Hamiltonianincompressible flow.

In particular, for spatially smooth flow, the long-time Lagrangian average(along the flow)

limt→∞

1

t

∫ t

0div v(t′) dt′ =

∑i

λi

is a sum of the Lyapunov exponents, which is then non-positive (in dis-tinction from an instantaneous average over space, which is zero at any time:∫div v dr = 0). It is important that we allowed for a compressibility of a

phase-space flow v(r, t) but did not require its irreversibility. Indeed, evenif the system is invariant with respect to t → −t, v → −v, the entropyproduction rate is generally non-negative and the sum of the Lyapunov ex-ponents is non-positive for the same simple reason that contracting regionshave more measure and give higher contributions. Backwards in time themeasure also concentrates, only on a different set.

40

Page 41: Physical Nature of Information - Weizmann

This can be illustrated by a slight generalization of the baker map, ex-panding one region and contracting another, keeping the whole volume ofthe phase space unity:

The transformation has the form

x′ =x/l for 0 < x < l(x− l)/r for l < x < 1 ,

y′ =ry for 0 < x < lr + ly for l < x < 1

, (50)

where r + l = 1. The Jacobian of the transformation is not equal to unitywhen r 6= l:

J =

∣∣∣∣∣∂(x′, y′)

∂(x, y)

∣∣∣∣∣ =r/l for 0 < x < ll/r for l < x < 1 ,

. (51)

Like in the treatment of the incompressible baker map in the previous section,consider two initially closed points. If during n steps the points find themsel-ves n1 times in the region 0 < x < l and n2 = n− n1 times inside l < x < 1then the distances along x and y will be multiplied respectively by l−n1r−n2

and rn1ln2 . Taking the limit we obtain the Lyapunov exponents:

λ+ = limn→∞[n1

nln 1

l+ n2

nln 1

r

]= −l ln l − r ln r , (52)

λ− = limn→∞[n1

nln r + n2

nln l]

= r ln l + l ln r (53)

The sum of the Lyapunov exponents, λ+ + λ− = (l − r) ln(r/l) = ln J ,is non-positive and is zero only for l = r = 1/2. Long-time average volumecontraction of a fluid element and respective entropy production is the analogof the second law of thermodynamics. The volume contraction means thatthe expansion in the λ+-direction proceeds slower than the contraction inthe λ−-direction. Asymptotically our strips of pasta concentrate on a fractalset, that is one having non-integer dimensionality. Indeed, define the (box-counting) dimension of a set as follows

d = limε→0

lnN(ε)

ln(1/ε), (54)

41

Page 42: Physical Nature of Information - Weizmann

where N(ε) is the number of boxes of length ε on a side needed to coverthe set. After n iterations of the map, square having initial side δ 1will be stretched into a long thin rectangle of length δ exp(nλ+) and widthδ exp(nλ−). To cover contracting direction, we choose ε = δ exp(nλ−), thenN(ε) = exp[n(λ+ − λ−)], so that the dimension is

d = 1 +λ+

|λ−|, (55)

The dimension is between 1 and 2. The set is smooth in the x-directionand fractal in the y-direction, which respectively gives two terms in (55).We see the dramatic difference between equilibrium equipartition and non-equilibrium fractal distribution. Relation between compressibility and non-equilibrium is natural: to make system non-Hamiltonian one needs to act bysome external forces, which pump energy into some degrees of freedom and,to keep a steady state, absorb it from other degrees of freedom — expansionand contraction of the momentum part of the phase-space volume.

Adding any noise to the system or treating it with a finite resolutionmakes finite both the coarse-graining scale ε and the entropy S = lnN(ε).An equilibrium uniform (microcanonical) distribution in d-dimensional phasespace has the entropy S0 = d ln(L/ε); the non-equilibrium steady state(NESS) generally has a lower dimensionality df < d with a lower entropy,S0 ' df ln(L/ε).

The non-equilibrium distribution is singular in the limit ε→ 0, that is occupies

a zero-measure subset of the phase space. The singularity of the non-equilibrium

measure is probably related to non-analyticity of kinetic coefficients and other

quantities, mentioned at the end of the Section 2.2. In reality, approach to NESS

is asymptotic; not much is known what distribution modifications provide for a

permanent entropy decrease in the limit ε→ 0. Most likely, lower-order cumulants

stabilize first, while farther and farther tails of the distribution continue to evolve.

To conclude this Chapter, let us stress the difference between the entropygrowth described in the Sections 2.3-2.4 and the entropy decay described in

42

Page 43: Physical Nature of Information - Weizmann

the present Section. In the former, phase-space flows were area-preservingand the volume growth of an element was due to a finite resolution whichstabilized the size in the contracting direction, so that the mean rate ofthe volume growth was solely due to stretching directions and thus equalto the sum of the positive Lyapunov exponents, as described in Section 2.3.On the contrary, the present section deals with compressible flows whichdecrease entropy by creating more inhomogeneous distributions, so that themean rate of the entropy decay is the sum of all the Lyapunov exponents,which is non-positive since contracting regions contain more trajectories andcontribute the mean rate more than expanding regions.

Looking back at the previous Chapters, it is a good time to appreci-ate the complementarity of determinism and randomness expressed in terms”statistical mechanics” and ”dynamical chaos”.

3 Physics of information

This section presents an elementary introduction into the information theoryfrom the viewpoint of a natural scientist. It re-tells the story of statisticalphysics using a different language, which lets us to see the Boltzmann andGibbs entropies in a new light. What I personally like about the informationviewpoint is that it erases paradoxes and makes the second law of thermo-dynamics trivial. It also allows us to see generality and commonality in theapproaches (to partially known systems) of physicists, engineers, computerscientists, biologists, brain researchers, social scientists, market speculators,spies and flies. We shall treat communication, measurement and learning asessentially the same phenomena. We shall see how fast widens the regionof applications of the universal tool of entropy (and related notions of re-lative entropy and mutual information): from physics, communications andcomputations to artificial intelligence and quantum computing.

3.1 Central limit theorem and large deviations

The true logic of this world is to be found in the theory of probability.Maxwell

Here we switch from continuous thinking in terms of phase-space flowsto discrete combinatoric manipulations. As a bridge from statistical physicsto information theory, we start from a simple technical tool used in both.

43

Page 44: Physical Nature of Information - Weizmann

Mathematics, underlying most of the statistical physics in the thermodyna-mic limit, comes from universality, which appears upon adding independentrandom numbers. The weakest statement is the law of large numbers: thesum approaches the mean value exponentially fast. The next level is thecentral limit theorem, which states that majority of fluctuations around themean have Gaussian probability distribution. Consideration of large rarefluctuations requires the so-called large-deviation theory. Here we brieflypresent all three at the physical (not mathematical) level.

Consider the variable X which is a sum of many independent identicallydistributed (iid) random numbers X =

∑N1 yi. Its mean value 〈X〉 = N〈y〉

grows linearly with N . Here we show that its fluctuations X−〈X〉 not excee-ding O(N1/2) are governed by the Central Limit Theorem: (X − 〈X〉)/N1/2

becomes for large N a Gaussian random variable with variance 〈y2〉− 〈y〉2 ≡∆. The quantities yi that we sum can have quite arbitrary statistics, the onlyrequirements are that the first two moments, the mean 〈y〉 and the variance∆, are finite. Finally, the fluctuations X −〈X〉 on the larger scale O(N) aregoverned by the Large Deviation Theorem that states that the PDF of Xhas asymptotically the form

P(X) ∝ e−NH(X/N) . (56)

P(X) −lnP/N

N1

N2

<X>X /N X/N

To show this, we write

P(X) =∫δ(∑N

i=1 yi −X)P(y1)dy1 . . .P(yN) dyN

=∫∞−∞ dp

∫exp

[ıp(∑N

i=1 yi −X)]P(y1)dy1 . . .P(yN) dyN

=∫∞−∞ dpe

−ıpX ∏Ni=1

∫eıpyiP(yi)dyi =

∫∞−∞ dpe

−ıpX+NG(ıp) . (57)

Here we introduced the generating function 〈e zy〉 ≡ eG(z). The derivatives ofthe generating function with respect to z at zero are equal to the momentsof y, while the derivatives of its logarithm G(z) are equal to the moments of(y − 〈y〉) called cumulants (see exercise).

44

Page 45: Physical Nature of Information - Weizmann

For large N , the integral (57) is dominated by the saddle point z0 suchthat G′(z0) = X/N . This is similar to representing the sum (20) above byits largest term. If there are several saddle-points, the result is dominated bythe one giving the largest probability. We assume that contour of integrationcan be deformed in the complex plane z to pass through the saddle pintwithout hitting any singularity of G(z). We now substitute X = NG′(z0)into −zX +NG(z), and obtain the large deviation relation (56) with

H = −G(z0) + z0G′(z0) . (58)

We see that −H and G are related by the ubiquitous Legendre transform(which always appear in the saddle-point approximation of the integral Fou-rier or Laplace transformations). Note that NdH/dX = z0(X) and

N2d2H/dX2 = Ndz0/dX = 1/G′′(z0) .

The function H of the variable X/N − 〈y〉 is called Cramer or rate functionsince it measures the rate of probability decay with the growth of N for everyX/N . It is also sometimes called entropy function since it is a logarithm ofprobability.

Several important properties of H can be established independently ofthe distribution P(y) or G(z). It is a convex function as long as G(z) isa convex function since their second derivatives have the same sign. It isstraightforward to see that the logarithm of the generating function has apositive second derivative (at least for real z):

G′′(z) =d2

dz2ln∫ezyP(y) dy

=

∫y2ezyP(y) dy

∫ezyP(y) dy − [

∫yezyP(y) dy]2

[∫ezyP(y) dy]2

≥ 0 . (59)

This uses the Cauchy-Bunyakovsky-Schwarz inequality which is a generali-zation of 〈y2〉 ≥ 〈y〉2. Also, H(z0) takes its minimum at z0 = 0, i.e. for Xtaking its mean value 〈X〉 = N〈y〉 = NG′(0). The maximum of probabilitydoes not necessarily coincides with the mean value, but they approach eachother when N grows and maximum is getting very sharp — this is called thelaw of large numbers. Since G(0) = 0 then the minimal value of H is zero,that is the probability maximum saturates to a finite value when N → ∞.Any smooth function is quadratic around its minimum with H ′′(0) = ∆−1,

45

Page 46: Physical Nature of Information - Weizmann

where ∆ = G′′(0) is the variance of y. Quadratic entropy means Gaussianprobability near the maximum — this statement is (loosely speaking) theessence of the central limit theorem. In the particular case of Gaussian P(y),the PDF P(X) is Gaussian for any X. Non-Gaussianity of the y’s leads toa non-quadratic behavior of H when deviations of X/N from the mean arelarge, of the order of ∆/G′′′(0).

One can generalize the central limit theorem and the large-deviation approachin two directions: i) for non-identical variables yi, as long as all variances arefinite and none dominates the limit N → ∞, it still works with the mean andthe variance of X being given by the average of means and variances of yi; ii) ifyi is correlated with a finite number of neighboring variables, one can group such”correlated sums” into new variables which can be considered independent.

The above figure and (56,58) show how distribution changes upon adding more

numbers. Is there any distribution which does not change upon averaging, that

is upon passing from yi to∑Ni=1 yi/N? That can be achieved for H ≡ 0, that is

for G(z) = kz, which corresponds to the Cauchy distribution P(y) ∝ (y2 + k2)−1.

Since the averaging decreases the variance, it is no surprise that the invariant

distribution has infinite variance. We shall return to distributions invariant under

summation of variables considering Renormalization Group in Section 5.2.

Asymptotic equipartition. The above law of large numbers state thatthe sum of iid random numbers y1 + . . . + yN approaches N〈y〉 as N grows.One can also look at the given sequence y1, . . . , yN and ask: how probableit is? Can we answer this blatantly self-referential question without seeingother sequences? Yes, we can, if the sequence is long enough. Since thenumbers are independent, then the probability of any sequence is the productof probabilities, and the logarithm of the probability is the sum that satisfiesthe law of large numbers:

− 1

Nln p(y1, . . . , yN) = − 1

N

N∑i=1

lnP(yi) → −〈lnP(y)〉 = S(Y ) . (60)

We see that the log of probability converges to N times the entropy of y. Buthow we find S(Y )? For a sufficiently long sequence, we assume that the fre-quencies of different values of yi in our sequence give the probabilities of thesevalues; we thus estimate P(y) and compute S(Y ). In other words, we assumethat the sequence is typical. We then state that the probability of the typi-cal sequence decreases with N exponentially: p(y1, . . . , yN) = exp[−NS(y)].

46

Page 47: Physical Nature of Information - Weizmann

That probability is independent of the values y1, . . . , yN , that is the same forall typical sequences. Equivalently, the number of typical sequences growswith N exponentially with entropy setting the rate of growths. That focus ontypical sequences, which all have the same (maximal) probability, is knownas asymptotic equipartition and formulated as ”almost all events are almostequally probable”.

3.2 Information as a choice

”Nobody knows what entropy really is, so in adebate you will always have an advantage.”

von Neumann to Shannon

We want to know in which of n boxes a candy is hidden, that is weare faced with a choice among n equal possibilities. How much informa-tion we need to get the candy? Let us denote the missing information byI(n). Clearly, I(1) = 0, and we want the information to be a monotonicallyincreasing12 function of n. If we have several independent problems theninformation must be additive. For example, consider each box to have mcompartments. To know in which from mn compartments is the candy, weneed to know first in which box and then in which compartment inside thebox: I(nm) = I(n)+I(m). Now, we can write (Hartley 1927, Shannon 1948)

I(n) = I(e) lnn = k lnn (61)

That information must be a logarithm is clear also from obtaining the missinginformation by asking the sequence of questions in which half we find thebox with the candy, one then needs log2 n of such questions and respectiveone-bit answers. If we measure information in binary choices or bits thenI(n) = log2 n, that is k−1 = ln(2). We can easily generalize the definition (61)for non-integer rational numbers by I(n/l) = I(n)− I(l) and for all positivereal numbers by considering limits of the series and using monotonicity. Sothe message carrying the single number of the lucky box with the candybrings the information k lnn.

We used to think of information received through words and symbols.Essentially, it is always about in which box the candy is. Indeed, if we have

12The messages ”in box 2 out of 2” and ”in box 2 out of 22” bring the same candy butnot the same amount of information.

47

Page 48: Physical Nature of Information - Weizmann

an alphabet with n symbols then every symbol we receive is a choice out of nand brings the information k lnn. That is n symbols like n boxes. If symbolscome independently then the message of the length N can potentially be oneof nN possibilities so that it brings the information kN lnn. To convey thesame information by smaller alphabet, one needs longer message. If all the26 letters of the English alphabet were used with the same frequency thenthe word ”love” would bring the information equal to 4 log2 26 ≈ 4 · 4.7 =18.8 bits. Here and below we assume that the receiver has no other priorknowledge on subjects like correlations between letters (for instance, everyonewho knows English, can infer that there is only one four-letter word whichstarts with “lov...” so the last letter brings zero information for such people).

A

A

A

A

B

E

B

B

B

Z

Z

Z

Z

L

O

V

...

...

...

...

...

...

...

...

...

n

N

In reality, every letter brings on average even less information than log2 26since we know that letters are used with different frequencies. Indeed, con-sider the situation when there is a probability pi assigned to each letter (orbox) i = 1, . . . , n. It is then clear that different letters bring different in-formation. Let us evaluate the average information per symbol in a longmessage. As N →∞ we know that the i-th letter appears Npi times in a ty-pical sequence, that is we know that we receive the first alphabet symbol Np1

times, the second symbol Np2 times, etc. What we didn’t know and whatany message of the length N brings is the order in which different symbolsappear. Total number of orders (the number of different typical sequences)is equal to N !/Πi(Npi)!, and the information that we obtained from a stringof N symbols is the logarithm of that number:

IN = k ln(N !/Πi(Npi)!

)≈ −Nk

∑i

pi ln pi +O(lnN) . (62)

The mean information per symbol coincides with the entropy (29):

I(p1 . . . pn) = limN→∞

IN/N = −kn∑i=1

pi ln pi . (63)

48

Page 49: Physical Nature of Information - Weizmann

Note that when n→∞ then (61) diverges while (63) may well be finite.Alternatively, one can derive (63) without any mention of randomness.

Consider again n boxes and define pi = mi/∑ni=1mi = mi/M , where mi

is the number of compartments in the box number i. The information ona specific compartment is a choice out of M and bring information k lnM .That information must be a sum of the information about the box I plus theinformation about the compartment summed over the boxes, k

∑ni=1 pi lnmi.

That gives the information I about the box (letter) as the difference:

I = k lnM − kn∑i=1

pi lnmi = kn∑i=1

pi lnM − kn∑i=1

pi lnmi = −kn∑i=1

pi ln pi .

The mean information (63) is zero for delta-distribution pi = δij; it isgenerally less than the information (61) and coincides with it only for equalprobabilities, pi = 1/n, when the entropy is maximum. Indeed, equal pro-babilities we ascribe when there is no extra information, i.e. in a state ofmaximum ignorance. In this state, a message brings maximum informationper symbol; any prior knowledge can reduce the information. Mathemati-cally, the property

I(1/n, . . . , 1/n) ≥ I(p1 . . . pn) (64)

is called convexity. It follows from the fact that the function of a singlevariable s(p) = −p ln p is strictly concave since its second derivative, −1/p,is everywhere negative for positive p. For any concave function, the averageover the set of points pi is less or equal to the function at the average value(so-called Jensen inequality):

1

n

n∑i=1

s (pi) ≤ s

(1

n

n∑i=1

pi

). (65)

−Wln W

A (A+B)/2 B

S[(A+B)/2]>[S(A)+S(B)]/2

W

49

Page 50: Physical Nature of Information - Weizmann

From here one gets the entropy inequality:

I(p1 . . . pn) =n∑i=1

s (pi) ≤ ns

(1

n

n∑i=1

pi

)= ns

(1

n

)= I

(1

n, . . . ,

1

n

). (66)

The relations (65-66) can be proven for any concave function. Indeed, theconcavity condition states that the linear interpolation between two pointsa, b lies everywhere below the function graph: s(λa+ b− λb) ≥ λs(a) + (1−λ)s(b) for any λ ∈ [0, 1], see the Figure. For λ = 1/2 it corresponds to (65)for n = 2. To get from n = 2 to arbitrary n we use induction. For that end,we choose λ = (n− 1)/n, a = (n− 1)−1∑n−1

i=1 pi and b = pn to see that

s

(1

n

n∑i=1

pi

)= s

(n− 1

n(n− 1)−1

n−1∑i=1

pi +pnn

)

≥ n− 1

ns

((n− 1)−1

n−1∑i=1

pi

)+

1

ns (pn)

≥ 1

n

n−1∑i=1

s (pi) +1

ns (pn) =

1

n

n∑i=1

s (pi) . (67)

In the last line we used the truth of (65) for n− 1 to prove it for n.

You probably noticed that (61) corresponds to the microcanonical Boltz-mann entropy (19) giving information/entropy as a logarithm of the numberof states, while (63) corresponds to the canonical Gibbs entropy (29) givingit as an average. An advantage of Gibbs-Shannon entropy (63) is that isdefined for arbitrary distributions, not necessarily equilibrium.

3.3 Communication Theory

After we learnt, what information messages bring on average, we are ready todiscuss the best ways to transmit messages. That brings us to the Commu-nication Theory, which is interested in two key issues, speed and reliability:

i) How much can a message be compressed; i.e., how redundant is theinformation? In other words, what is the maximal rate of transmission inbits per symbol?

ii) At what rate can we communicate reliably over a noisy channel; i.e.,how much redundancy must be incorporated into a message to protect againsterrors?

50

Page 51: Physical Nature of Information - Weizmann

Both questions concern redundancy – how unexpected is every letter ofthe message, on the average. Entropy quantifies redundancy. We have seenthat a symbolic communication channel on average transmits one unit of theinformation (63) per symbol. Receiving letter (box) number i through a bi-nary channel (transmitting ones and zeros)13 brings information log2(1/pi) =log2M − log2mi bits. Indeed, the remaining choice (missing information) isbetween mi compartments. The entropy −∑z

i=a pi log2 pi is the mean infor-mation content per letter. Less probable symbols have larger informationcontent, but they happen more rarely.

So the entropy is the mean rate of the information transfer, since it isthe mean growth rate of the number of typical sequences. What about themaximal rate of the information transfer? Following Shannon, we answerthe question i) statistically, which makes sense in the limit of very longmessages, when one can focus on typical sequences, as we did at the endof the Section 3.1 and in deriving (62). Consider for simplicity a message ofN bits, where 0 comes with probability 1 − p and 1 with probability p. Tocompress the message to a shorter string of letters that conveys essentiallythe same information it suffices to choose a code that treats effectively thetypical strings — those that contain N(1 − p) zeroes and Np ones. Thenumber of such strings is given by the binomial CN

Np which for large N is

2NI(p), where I(p) = −p log2 p − (1 − p) log2(1 − p). The strings differ bythe order of appearance of 0 and 1. To distinguish between these 2NI(p)

messages, we encode any one using a binary string with lengths starting fromone and ending at NI(p) bits. For example, we encode by a one-bit wordthe message where all Np ones are at the beginning followed by all N(1− p)zeroes, then by two-bit word the message with one hole, etc. The maximalword length NI(p) is less than N , since 0 ≤ I(p) ≤ 1 for 0 ≤ p ≤ 1. In otherwords, to encode all 2N sequences we need words of N bits, but to encode alltypical sequences, we need only words up to NI(p) bits. We indeed achievecompression with the sole exception of the case of equal probability whereI(1/2) = 1. True, the code must include a bit more to represent atypicalmessages, but in the limit of large N the chance of their appearance and theircontribution to the rate of transmission are negligible. Therefore, entropysets both the mean and the maximal rate in the limit of long sequences. Theidea of typical messages in the limit N →∞ is an information-theory analogof ensemble equivalence in the thermodynamic limit. You may find it bizarre

13Binary code is natural both for signals (present-absent) and for logic (true-false).

51

Page 52: Physical Nature of Information - Weizmann

that one uses randomness in treating information communications, whereone usually transfers non-random meaningful messages. One of the reasonsis that encoding program does not bother to ”understand” the message, andtreats it as random. Draining the words of meaning is necessary for devisinguniversal communication systems.

Maximal rate of transmission correspond to the shortest mean length ofthe codeword. Consider an alphabet with q symbols and the source with theprobability distribution p(i), i = 1, . . . , q. Then Shannon proved that theshortest mean length of the codeword ` is bounded by

−∑i

p(i) logq p(i) ≤ ` < −∑i

p(i) logq p(i) + 1 . (68)

Of course, not any encoding guarantees the maximal rate of transmission.Designating sequences of the same length to objects with different probabi-lities is apparently sub-optimal. To make the mean word length shorter andachieve signal compression, one codes frequent objects by short sequences andinfrequent ones by more lengthy combinations - lossless compressions like zip,gz and gif work this way. Consider a fictional creature whose DNA containsfour bases A,T,C,G occurring with probabilities pi listed in the table:

Symbol pi Code 1 Code 2A 1/2 00 0T 1/4 01 10C 1/8 10 110G 1/8 11 111

We want a binary encoding for the four bases. Since there are exactly fourtwo-bit words, one can suggest the Code 1, which has exactly 4 words anduses 2 bits for every base. Here the word length is 2. However, it is straig-htforward to see that the entropy of the distribution I = −∑4

i=1 pi log2 pi islower that 2. One then may suggest a variable-length Code 2 (an example ofthe so-called Huffman code). It is built in the following way. We start fromthe least probable C and G, which we want to have the longest codewords ofthe same length differing by one (last) bit that distinguishes between the twoof them. We then can combine C and G into a single source symbol with theprobability 1/4, that is coinciding with the probability of T. To distinguishfrom C,G, we code T by two-bit word placing 0 in the second position. Now,we can code A by one-bit word 0 to distinguish it from T,C,G. Alternatively,one can start from the first bit ascribing 0 to A and 1 to T,C,G, then add

52

Page 53: Physical Nature of Information - Weizmann

the second bit to distinguish T from C,G and finish with adding the thirdbit to distinguish between C and G. Home exercise is to see which code, 1 or2, uses less bits per base on average. The most efficient code has the lengthof the mean codeword (the number of bits per base) equal to the entropy ofthe distribution, which determines the fastest mean transmission rate, thatis the shortest mean codeword length.

The inequality (64) tells us, in particular, that using an alphabet is notoptimal for the information transmission rate as long as the probabilitiesof the letters are different. We can use less symbols but variable codewordlength. Morse code uses just three symbols (dot, dash and space) to encodeany language14. In English, the probability of ”E” is 13% and of ”Q” is 0.1%,so Morse encodes ”E” by a single dot and ”Q” by ”− − ·−” (first Britishtelegraph managed to do without C,J,Q,U,X). One-letter probabilities givefor the written English language the information per symbol as follows:

−z∑i=a

pi log2 pi ≈ 4.11 bits ,

which is lower than log2 26 = 4.7 bits.

3.4 Correlations in the signals

Apart from one-letter probabilities, one can utilize more knowledge about thelanguage by accounting for two-letter correlation (say, that ”Q” is alwaysfollowed by ”U”, ”H” often follows ”T”, etc). That will further lower theentropy. A simple universal model with neighboring correlations is a Markovchain. It is specified by the conditional probability p(j|i) that the letter i isfollowed by j. For example p(U |Q) = 1. The probability is normalized forevery i:

∑j p(j|i) = 1. The matrix pij = p(j|i), which has positive elements

in every column whose sum is unity, is called stochastic. The probabilityof any N -string is then the product of N − 1 transition probabilities timesthe probability of the initial letter. As in (60), minus the logarithm of theprobability of a long N -string grows linearly with N :

log2 p(i1, . . . , iN) = log2 p(i1) +N∑k=2

log2 p(ik+1|ik) . (69)

14Great contributions of Morse were one-wire system and the simplest possible encoding(opening and closing the circuit), far more superior to multiple wires and magnetic needlesof Ampere, Weber, Gauss and many others.

53

Page 54: Physical Nature of Information - Weizmann

Therefore, the number of typical sequences grows with N exponentially, as2NS, where S is the conditional entropy −∑j p(j|i) log2 p(j|i), which mustbe averaged over different i with their probabilities p(i). The vector of pro-babilities p(i) and the transition matrix pij are not independent but are re-lated by the detailed balance: p(i)pij = p(j)pji. Summing over j, we obtainp(i) =

∑p(j)pji, that is p = p(a), . . . p(z) is an eigenvector with the unit

eigenvalue of the matrix pij. We now express the language entropy via p(i)and p(j|i) by averaging over i the entropy of the transition probability dis-tribution:

S = −∑i

pi∑j

p(j|i) log2 p(j|i) . (70)

That formula defines the information rate of the Markov source. We shallfurther discuss Markov chains describing Google PageRank algorithm inSection 6.4 below.

One can go beyond two-letter correlations and statistically calculate theentropy of the next letter when the previous N − 1 letters are known (Shan-non 1950). As N increases, the entropy approaches the limit which can becalled the entropy of the language. Long-range correlations and the fact thatwe cannot make up words further lower the entropy of English down to ap-proximately 1.4 bits per letter, if no other information given. Comparing1.4 and 4.7, we conclude that the letters in an English text are about 70%redundant. About the same value one finds asking people to guess the lettersin a text one by one, which they do correctly 70% of the time. This redun-dancy makes possible data compression, error correction and crosswords. Itis illustrated by the famous New York City subway poster of the 1970s:

”If u cn rd ths u cn gt a gd jb w hi pa!”

Note in passing that the human language encodes meaning not in separateletters but in words. An insight into the way we communicate is given by thefrequency distribution of words and their meanings (Zipf 1949). It was foundempirically that if one ranks words by the frequency of their appearance intexts, then the frequency decreases as an inverse rank. For example, the firstplace with 7% takes ”the”, followed by ”of” with 3.5%, ”and” with 1.7%,etc.

As we have seen, knowing the probability distribution one can computeentropy, which determines the most efficient rate of encoding. One can turntables and estimate the entropy of the data stream looking for its most com-pact lossless encoding. It can be done in a one-pass (online) way, that is

54

Page 55: Physical Nature of Information - Weizmann

not looking at the whole string of data, but optimizing encoding as one pro-cess the string from beginning to end. There are several such algorithmscalled adaptive codes (Lempel-Ziv, deep neural networks, etc). These codesare also called universal, since they do not require a priori knowledge of thedistribution.

In distinction from lossless coding, jpeg, mpeg, mp3 and telephone use lossycompression which removes information presumed to be unimportant for humans.Moreover, any transcript of spoken words has by orders of magnitude lower entropythan the acoustic waveform, which means that it is possible to compress furtherwithout losing any information about the words and their meaning, as will bediscussed in the next subsection. Linguists define the phoneme as the smallestacoustic unit that makes a difference in meaning. Their numbers in differentlanguages are subject to disagreements but generally are in tens. For example,most estimates for English give 45, that is comparable with the number of lettersin the alphabet. Incidentally, the great invention of alphabetic writing, whichdramatically improved handling of information (and irreversibly changed the wayswe speak, hear and remember), was done only once in history. All known alphabetsderive from that seminal (Semitic) script.

Another encoding of profound importance is a positional numeral system, ba-

sed on the fundamental discovery that number N can be encoded by logN symbols

instead of N times repeating the same mark. One cannot overestimate the im-

portance of encoding where the value depends on the position, since it already

implies algebraic operations. Indeed, reading (decoding) requires multiplying and

adding: 2021 = 2× 1000 + 2× 10 + 1. It then allowed simple automatic rules for

computations (formulated by Persian al-Khwarizmi, from whose name the word

algorithm appeared).

3.5 Mutual information as a universal tool

Answering the question i) in Sect. 3.2, we have found that the entropy ofthe set of symbols to be transferred determines the minimum mean numberof bits per symbol, that is the maximal rate of information transfer. In thissection, we turn to the question ii) and find out how this rate is lowered ifthe transmission channel can make errors. How much information then islost on the way? In this context one can treat measurements as messagesabout the value of the quantity we measure. One can also view storing andretrieving information as sending a message through time rather than space.

When the channel is noisy the statistics of inputs P (B) and outcomes

55

Page 56: Physical Nature of Information - Weizmann

P (A) are generally different, that is we need to deal with two probabilitydistributions and the relation between them. Treating inputs and outputs astaken out of distributions works for channels/measurements both with andwithout noise; in the limiting cases, the distribution can be uniform or peakedat a single value. Relating two distributions needs introducing conditionaland relative entropies and mutual information, which presently are the mostpowerful and universal tools of information theory.

The relation between the message (measurement) Ai and the event (quan-tity) Bj is characterized by the conditional probability (of Bj in the presenceofAi), denoted P (Bj|Ai). For everyAi, this is a normalized probability distri-bution, and one can define its entropy S(B|Ai) = −∑j P (Bj|Ai) log2 P (Bj|Ai).Since we are interested in the mean quality of transmission, we average thisentropy over all values of Aj, which defines the so-called conditional entropy:

S(B|A) =∑i

P (Ai)S(B|Ai) =∑ij

P (Ai)P (Bj|Ai) log2 P (Bj|Ai)

=∑ij

P (Ai, Bj) log2 P (Bj|Ai) . (71)

Here we related the conditional probability to the joint probability P (Ai, Bj)by the evident formula P (Ai, Bj) = P (Bj|Ai)P (Ai). The conditional entropymeasures what on average remains unknown after the value of A is known.The missing information was S(B) before the measurement and is equal tothe conditional entropy S(B|A) after it. Then what the measurements bringon average is their difference called the mutual information:

I(A,B) = S(B)− S(B|A) =∑ij

P (Ai, Bj) log2

[P (Bj|Ai)P (Bj)

]. (72)

Indeed, information is a decrease in uncertainty. Non-negativity of the mu-tual information means that on average measurements increase the conditio-nal probability: 〈log2[P (Bj|Ai)/P (Bj)]〉 ≥ 0. For example, let B be a choiceout of n equal possibilities: P (B) = 1/n and S(B) = log2 n. If A can happenonly in m cases out of those n, that is P (B|A) = 1/m and S(B|A) = log2m,then I(A,B) = log2(n/m) bits. If m = 1 then A tells us all we need to knowabout B.

We can now recognize in (70) the conditional entropy, which is lowerthan the one-letter entropy I = −∑i p(i) log p(i) by the mutual informationbetween subsequent letters. That lowering of the entropy was brought bythe knowledge of the conditional probabilities p(j|i) in addition to p(i).

56

Page 57: Physical Nature of Information - Weizmann

The formula P (Ai, Bj) = P (Bj|Ai)P (Ai) gives the chain rule,

S(A,B)=S(A)+S(B|A)=S(B)+S(A|B), (73)S(A) S(A,B) S(B)

S(A|B) I(A,B) S(B|A)and I(A,B) in a symmetric form:

I(A,B) =S(B)−S(B|A)=S(A)+S(B)−S(A,B) = S(A)− S(A|B) . (74)

When A and B are independent, the conditional probability is independentof A and the information is zero. When they are dependent, P (B,A) >P (A)P (B), so that that the information is indeed positive. When A,Bare related deterministically, S(A) = S(B) = S(A,B) = I(A,B), whereS(A) = −∑i P (Ai) log2 P (Ai), etc. And finally, since P (A|A) = 1 then themutual information of a random variable with itself is the entropy: I(A,A) =S(A). So one can call entropy self-information. Non-negativity of the mutualinformation also gives the so-called sub-additivity of entropy:

S(A) + S(B) > S(A,B) . (75)

One also uses P (A,B) = P (B|A)P (A) = P (A|B)P (B) for estimatingthe conditional probability of the event B given the marginal probability ofthe measurements A:

P (B|A) = P (B)P (A|B)

P (A). (76)

For example, experimentalists measure the sensory response of an animal tothe stimulus, which gives P (A|B)/P (A) or build a robot with the prescribedresponse. Then they go to the natural habitat of that animal/robot andmeasure the distribution of stimulus P (B) (see the example at the beginningof Section 4.3). After that one obtains the conditional probability (76) thatallows animal/robot to function effectively in that habitat.

It is intuitively clear that the mutual information must play for noisychannels the same role the entropy plays for ideal channels, in particularly,must set the maximal rate of reliable communication thus answering thequestion ii) from the Section 3.3. Indeed, the number of possible outputs fora given input is an inverse of the conditional probability P (Ai|Bj). For eachtypical input N -sequence, we have [P (A|B)]−N = 2NS(A|B) possible output

57

Page 58: Physical Nature of Information - Weizmann

sequences, all of them equally likely. To identify the input without error, weneed to divide the total number of typical outputs 2NS(A) into sets of size2NS(A|B) corresponding to different inputs. Therefore, we can distinguishat most 2NS(A)/2NS(A|B) = 2NI(A,B) sequences of the length N , which setsI(A,B) as the maximal rate of information transfer.

If one is interested in the channel as specified by P (B|A), then one max-imizes I(A,B) over all choices of the source statistics P (B) and call it theShannon’s channel capacity, which quantifies the quality of communicationsystems or measurements in bits per symbol:

C = maxP (B)

I(A,B) .

The channel capacity is the log of the maximal number of distinguishablesignals. For example, if our channel transmits the binary input exactly (zeroto zero, one to one), then the capacity is 1 bit, which is achieved by choosingP (B = 0) = P (B = 1) = 1/2, see the left panel in the Figure. Even ifthe channel has many outputs for every input out of n, the capacity is stilllog2 n, if those outputs are non-overlapping for different inputs, so that theinput can be determined without an error and P (B|A) = 1, as in the caseof m = 1 above, see the middle panel in the Figure. The capacity is loweredwhen the same outputs appear for different inputs, say, different groups of minputs each gives the same output, so that P (B|A) = 1/m. In this case, oneachieves error-free transition choosing only one input symbol from each ofn/m groups, that is using P (B) = m/n for the symbols chosen and P (B) = 0for the rest; the capacity is then indeed C = log2(n/m) bits (in the right panelof the Figure n = 6, m = 2). Lowered capacity means increased redundancy,that is a need to send more symbols to convey the same information. Asmentioned, shorter alphabet requires longer messages.

1

3456

2

1

234

5

1

23

4

1/2

1/2

6

B A

3/4

1/4B

1 1

A

0 0 0

1

C=1 C=1 C=S(B)−S(B|A)=log3

In most cases, however, noise does not allow to separate inputs into groupswith completely disjoint outputs, so errors always present. It was thoughtthat in such cases it is impossible to make probability of error arbitrarily

58

Page 59: Physical Nature of Information - Weizmann

small when sending information with a finite rate R. Shannon have shownthat it is possible, if there is any correlation between output A and inputB, that is C > 0. Then the probability of an error can be made 2−N(C−R),that is asymptotically small in the limit of N → ∞, if the rate is lowerthan the channel capacity. This (arguably the most important) result ofthe communication theory is rather counter-intuitive: if the channel makeserrors all the time, how one can decrease the error probability treating longmessages? Shannon’s argument is based on typical sequences and averageequipartition, that is on the law of large numbers (by now familiar to you).

For example, if in a binary channel the probability of every single bitgoing wrong is q, then A is binary random variable with equal probabilitiesof 0 and 1, so that S(A) = log2 2 = 1. Conditional probabilities are P (1|0) =P (0|1) = q and P (1|1) = P (0|0) = 1 − q, so that S(A|B) = S(B|A) =S(q) = −q log2 q − (1 − q) log2(1 − q). The mutual information I(A,B) =S(A)−S(A|B) = 1−S(q). This is actually the maximum, that is the channelcapacity: C = maxP (B)[S(B) − S(B|A)] = 1 − S(q), because the maximalentropy is unity for a binary variable B. Let us now see how the rate oftransmission is bounded from above by the capacity. In a message of lengthN , there are on average qN errors and there are N !/(qN)!(N−qN)! ≈ 2NS(q)

ways to distribute them. We then need to devote some m bits in the messagenot to data transmission but to error correction. Apparently, the numberof possibilities provided by these extra bits, 2m, must exceed 2NS(q), whichmeans that m > NS(q), and the transmission rate R = (N − m)/N <1 − S(q). The channel capacity is zero for q = 1/2 and is equal to 0.988bits per symbol for q = 10−3. The probability of errors is binomial with

the mean number of errors qN and the standard deviation σ =√Nq(1− q).

If we wish to bound the error probability from above, we must commit tocorrecting more than the mean number of errors, making the transmissionrate smaller than the capacity.

When the measurement/transmission noise ξ is additive, that is the out-put is A = g(B) + ξ with an invertible function g, we have S(A|B) = S(ξ)and

I(A,B) = S(A)− S(ξ) . (77)

The more choices of the output are recognizable despite the noise, the more isthe capacity of the channel. In general, when the conditional entropy S(A|B)is given, then to maximize the mutual information we need to choose themeasurement/coding procedure (for instance, g(B) above) that maximizes

59

Page 60: Physical Nature of Information - Weizmann

the entropy of the output S(A).

Gaussian Channel. As an illustration, consider a linear noisy channel:A = B+ξ, such that the noise is independent of B and Gaussian with 〈ξ〉 = 0and 〈ξ2〉 = N . Then P (A|B) = (2πN )−1/2 exp[−(A−B)2/2N ]. If in additionwe have a Gaussian input signal with P (B) = (2π)−1/2 exp(−B2/2S), thenP (A) = [2π(N + S)]−1/2 exp[−A2/2(N + S)]. Now, using (76) we can write

P (B|A) =

√N + S

2Nexp

[−S +N

2N

(B − A

S +N

)2].

In particular, the estimate of B is linearly related to the measurement A:

B =∫BP (B|A) dB =

A

S +N= A

SNR

1 + SNR, (78)

where signal to noise ratio is SNR = S/N . The rule (78) makes sense:To ”decode” the output of a linear detector we use the unity factor at highSNR, while at low SNR we scale down the output since most of what we areseeing must be noise. As is clear from this example, linear relation betweenthe measurement and the best estimate requires two things: linearity of theinput-output relation and Gaussianity of the statistics. Let us now find themutual information (77):

I(A,B)=S(A)−S(A|B)=S(A)−S(B+ξ|B)=S(A)−S(ξ|B)=S(A)−S(ξ)

= 12

[log2 2πe(S +N )− log2 2πeN ] = 12

log2(1 + SNR) . (79)

Here we used the formula for the entropy of the Gaussian distribution. Thecapacity of such a channel depends on the input statistics. One increasescapacity by increasing the input signal variance, that is the dynamic rangerelative to the noise. For a given input variance, the maximal mutual in-formation (channel capacity) is achieved by a Gaussian input, because theGaussian distribution has maximal entropy for a given variance. Indeed,varying

∫dxρ(x)(λx2 − ln ρ) with respect to ρ we obtain ρ(x) ∝ exp(−λx2).

Examples of redundancy. How redundant is the genetic code? Thereare four bases, which must encode twenty amino acids. There are 42 two-letter words, which is not enough. The designer then must use a triplet codewith 43 = 64 words, so that the redundancy factor is about 3. Number

60

Page 61: Physical Nature of Information - Weizmann

of ways to encode a given amino acid is approximately proportional to itsfrequency of appearance.

What are the error rates in the transmission of the genetic code? Typicalenergy cost of a mismatched DNA base pair is that of a hydrogen bond,which is about ten times the room temperature. If the DNA molecule was inthermal equilibrium with the environment, thermal noise would cause errorprobability e−10 ' 10−4 per base. This is deadly. A typical protein hasabout 300 amino acids, that is encoded by about 1000 bases; we cannothave mutations in every tenth protein. Moreover, synthesis of RNA fromDNA template and of proteins on the ribosome involve comparable energiesand could cause comparable errors. That means that Nature operates ahighly non-equilibrium state, so that bonding involves extra irreversible stepsand burning more energy. This way of sorting molecules is called kineticproofreading (Hopfield 1974, Ninio 1975) and is very much similar to theMaxwell demon discussed below in Section 4.2.

Another example of redundancy for error-protection is the NATO pho-netic alphabet used by the military and pilots. To communicate through anoisy acoustic channel, letters are encoded by full words: A is Alpha, B isBravo, C is Charlie, etc.

Mutual information also sets the limit on the data compression A → C,if coding has a random element so that its entropy S(C) is nonzero. In thiscase, the maximal data compression, that is the minimal coding length inbits, is min I(A,C).

Possible communication

schemes

transmission

limit

compression

limit

min I(A,C) max I(A,B)

Take-home lesson: entropy of the symbol set is the ultimate data com-pression rate; channel capacity is the ultimate transmission rate. Since wecannot compress below the entropy of the alphabet and cannot transfer fasterthan the capacity, then transmission is possible only if the former exceedsthe latter, which requires positivity of the mutual information.

3.6 Hypothesis testing and relative entropy

All empirical sciences need a quantitative tool for confronting data withhypothesis. One (rational) way to do that is statistical: update prior beliefsin light of the evidence. It is done using conditional probability. Indeed, for

61

Page 62: Physical Nature of Information - Weizmann

any e and h, we have P (e, h) = P (e|h)P (h) = P (h|e)P (e). If we now call hhypothesis and e evidence, we obtain the rule for updating the probabilityof hypothesis to be true, which is the Bayes’ rule:

P (h|e) = P (h)P (e|h)

P (e). (80)

That is the new (posterior) probability P (h|e) that the hypothesis is correctafter we receive the data e is the prior probability P (h) times the quotientP (e|h)/P (e) which presents the support e provides for h. Without exag-geration, one can say that most errors made by experimentalists in scienceand most wrong conclusions made by conspiracy theorists are connected tounfamiliarity with this simple formula. For example, your hypothesis is theexistence of a massive international conspiracy to increase the power of go-vernments and the evidence is COVID pandemic. In this case P (e|h) is high:a pandemic provoking increase of the state power is highly likely given sucha conspiracy exists. This is presumably why some people stop thinking hereand accept the hypothesis. But of course, absent such an event, the priorprobability P (h) could be vanishingly small. Only sequence of probability-increasing events may lead us to accept the hypothesis.

If choosing between two mutually exclusive hypotheses, h1 and h2, then

P (h1|e) = P (h1)P (e|h1)

P (h1)P (e|h1) + P (h2)P (e|h2). (81)

Suppose that a drug test is 99% sensitive and 99% specific. That is, thetest will produce 99% true positive results for drug users (hypothesis h1)and 99% true negative results for clean people (hypothesis h2). If we denotee the positive test result, then P (e|h1) = 0.99 and P (e|h2) = 1 − 0.99 =0.01. Suppose that 0.5% of people are users of the drug, that is P (h1) =0.005. The probability that a randomly selected individual with a positivetest is a drug user is 0.005 · 0.99/(0.99 · 0.005 + 0.01 · 0.995) ≈ 0.332 that isless that half. The result is more sensitive to specificity approaching unity,when P (e|h2) → 0, than to sensitivity. In particular, checking an a prioriimprobable hypothesis, P (h1) P (h2), it is better to design experimentwhich minimizes P (e|h2) rather than maximizes P (e|h1), that is rules outalternative rather than supports the hypothesis. To see the combination ofprobabilities which defines the posterior probability of the hypothesis being

62

Page 63: Physical Nature of Information - Weizmann

true, it is instructive to present the result in the following form:

P (h1|e) = 1 +P (h2)P (e|h2)

P (h1)P (e|h1). (82)

There is evidence that perception of our brain is inferential, that is based on the

prediction and hypothesis testing. Among other things, this is manifested by the

long known phenomenon of binocular rivalry and the recently established fact that

signals between brain and sensory organs travel in both directions simultaneously.

It is then likely that even our unconscious activity uses rational Bayes’ rule, where

e is sensory input. See e.g. ”The Predictive Mind” by J. Hohwy. Observing our

own mental processes gives us both the idea of logic and of statistical inference.

Note a shift in the interpretation of probability. Traditionally, mathe-maticians and gamblers treated probability as the frequency of outcomes inrepeating trials. Bayesian approach defines probability as a degree of belief;that definition allows wider applications, particularly when we cannot haverepeating identical trials. The approach may seem unscientific since it is de-pendent on the prior beliefs, which can be subjective. However, repeatedlysubjecting our hypothesis to variable enough testing, we hope that the re-sulting flow in the space of probabilities will eventually come close to a fixedpoint independent of the starting position.

Relative Entropy. If the true distribution is p but our hypothetical distri-bution is q, what number N of trials is sufficient to invalidate our hypothesis?For that we need to estimate the probability of the stream of data observed.We shall observe the result i number of times which is piN and judge theprobability of it happening as qpiNi times the number of sequences with thosefrequencies:

P =∏i

qpiNiN !∏

j(pjN)!. (83)

This is the probability of our hypothetical distribution being true. Conside-ring limit of large N we obtain a large-deviation-type relation like (56):

P ∝ exp

[−N

∑i

pi ln(pi/qi)

]. (84)

The probability of not-exactly-correct hypothesis to approximate the dataexponentially decreases with the number of trials. The rate of that decrease

63

Page 64: Physical Nature of Information - Weizmann

is the relative entropy (also called Kullback-Liebler divergence):

D(p|q) =∑i

pi ln(pi/qi) = 〈ln(p/q)〉 . (85)

The relative entropy determines how many trials we need: we prove ourhypothesis wrong when ND(p|q) becomes large. The closer is our hypothesisto the true distribution, the larger is the number of trials needed. On theother hand, when ND(p|q) is not large, our hypothetical distribution is justfine.

Relative entropy also quantifies how close to reality is the asymptoticequipartition estimate (60) of the probability of a given sequence. Assumethat we have an N -sequence where the values/letters appear with the fre-quencies qk, where k = 1, . . . , K. Then the asymptotic equipartition (thelaw of large numbers) advices us that the probability of that sequence is∏k q

Nqkk = exp(N

∑k qk ln qk) = exp[−NS(q)]. But if the true probabi-

lities of the letters are pk, then the probability of the sequence is ac-tually

∏k p

Nqkk = exp(N

∑k qk ln pk) = exp[N

∑k(qk ln qk + qk ln(pk/qk))] =

exp−N [S(q)−D(q|p)].How many different probability distributions qk (called types in infor-

mation theory) exist for an N -sequence? Since qk = n/N , where n can takeany of N+1 values 0, 1, . . . , N then the number of possible K-vectors qk isat most (N+1)K , which grows with N only polynomially, where the alphabetsize K sets the power. Since the number of sequences grows exponentiallywith N , then there is an exponential number of possible sequences for eachtype. The probability to observe a given type (empirical distribution) isdetermined by the relative entropy Pqk ∝ exp[−ND(q|p)].

The relative entropy measures how different is the hypothetical distribu-tion q from the true distribution p. Note that D(p|q) is not the differencebetween entropies (which just measures difference in uncertainties). The re-lative entropy is not a true geometrical distance since it does not satisfythe triangle inequality and is asymmetric, D(p|q) 6= D(q|p). Indeed, thereis no symmetry between reality and our version of it (no matter how somephilosophers want us to believe). Yet D(p|q) has important properties of adistance. Since the probability does not exceed unity, the relative entropyis non-negative, it turns into zero only when distributions coincide, that ispi = qi for all i. This can be readily demonstrated using the simple inequalitylnx ≤ x− 1, which turns into equality only for x = 1.

Mutual information is the particular case of the relative entropy when

64

Page 65: Physical Nature of Information - Weizmann

we compare the true joint probability p(xi, yj) with the distribution madeout of their separate measurements q(xi, yj) = p(xi)p(yj), where p(xi) =∑j p(xi, yj) and p(yj) =

∑i p(xi, yj): D(p|q) = S(X) + S(Y ) − S(X, Y ) =

I(X, Y ) ≥ 0. If i in pi runs from 1 to M we can introduce D(p|u) =log2M − S(p), where u is a uniform distribution. That allows one to showthat both relative entropy and mutual information inherit from entropy con-vexity properties. You are welcome to prove that D(p|q) is convex withrespect to both p and q, while I(X, Y ) is a concave function of p(x) for fixedp(y|x) and a convex function of p(y|x) for fixed p(x). In particular, convexityis important for making sure that the extremum we are looking for is uniqueand lies at the boundary of allowed states.

Relative entropy also measures the price of non-optimal coding. As we dis-

cussed before, a natural way to achieve an optimal coding would be to assign

the length to the codeword according to the probability of the object encoded:

li = − log2 pi. Indeed, the information in bits about the object, log2(1/pi), must

be exactly equal to the length of its binary encoding. For an alphabet with d

letters, li = − logd pi. The more frequent objects are then coded by shorter

words, and the mean length is the entropy. The problem is that li must all be

integers, while − logd pi are generally not. A set of integer li effectively corre-

sponds to another distribution with the probabilities qi = d−li/∑i d−li . Assume

for simplicity that we found encoding with∑i d−li = 1 (unity can be proved to

be an upper bound for the sum). Then li = − logd qi and the mean length is

l =∑i pili = −

∑i pi logd qi = −

∑i pi(logd pi − logd pi/qi) = S(p) + D(p|q), that

is larger than the optimal value S(p), so that the transmission rate is lower. In

particular, if one takes li = dlogd(1/pi)e, that is the integer part, then one can

show that S(p) ≤ l ≤ S(p) + 1, that is non-optimality is at most one bit.

Monotonicity and irreducible correlations. If we observe less varia-bles, then the relative entropy is less, property called monotonicity:

D[p(xi, yj)|q(xi, yj)] ≥ D[p(xi)|q(xj)] ,

where as usual p(xi) =∑j p(xi, yj) and q(xi) =

∑j q(xi, yj). When we ob-

serve less variables we need larger N to have the same confidence. In otherwords, information does not hurt (but only on average!). For three variables,one can define q(xi, yj, zk) = p(xi)p(yj, zk), which neglects correlations bet-ween X and the rest. What happens to D[p(xi, yj, zk)|q(xi, yj, zk)] if we donot observe Z at all? Integrating Z out turns q into a product. Monotonicity

65

Page 66: Physical Nature of Information - Weizmann

givesD[p(xi, yj, zk)|q(xi, yj, zk)] ≥ D[p(xi, yj)|q(xi, yj)] .

But when q is a product, D turns into I and we can use (74):

D[p(xi, yj, zk)|q(xi, yj, zk)] = S(X) + S(Y, Z)− S(X, Y, Z),

so we obtain S(X, Y ) + S(Y, Z) − S(Y ) − S(X, Y, Z) ≥ 0, which is calledstrong sub-additivity. It can be presented as the positivity of the conditionalmutual information

I(X,Z|Y ) = S(X|Y ) + S(Z|Y )− S(X,Z|Y ) = S(X, Y )− S(Y ) + S(Z, Y )

−S(Y )− S(X, Y, Z) + S(Y ) = S(X, Y ) + S(Y, Z)− S(Y )− S(X, Y, Z) .(86)

That allows one to make the next step in disentangling information encoding.The straightforward generalization of the mutual information for many ob-jects, I(X1, . . . , Xk) =

∑S(Xi) − S(X1, . . . , Xk), simply measures the total

correlation. We can introduce a more sophisticated measure of correlationscalled the interaction (or multivariate) information, which measures the ir-reducible information in a set of variables, beyond that which is present inany subset of those variables. For three variables it measures the differencebetween the total correlation and that encoded in all pairs and is defined asfollows (McGill 1954):

II = I(X,Z)− I(X,Z|Y ) = S(X) + S(Y ) + S(Z)− S(X, Y )− S(X,Z)

+S(X, Y, Z)− S(Y, Z) = I(X, Y ) + I(X,Z) + I(Y, Z)− I(X, Y, Z). (87)

Interaction information measures the influence of a third variable on theamount of information shared between the other two and could be of eithersign. When positive, it indicates that the third variable accounts for someof the correlation between the other two, that is its knowledge diminishesthe correlation. When negative, it indicates that the knowledge of the thirdvariable facilitates the correlation between the other two. Alternatively, onemay say that a positive II(X, Y, Z) measures redundance in the informationabout the third variable contained in the other two separately, while negativeone measures synergy which is the extra information about Y received byknowing X and Z together, instead of separately. For example, love trianglescan be either redundant or synergetic (information-wise). If Y dates eitherX, both X,Z or none, then the dating states of X and Z are correlated.

66

Page 67: Physical Nature of Information - Weizmann

Knowing one tells us more about another (chooses from more possibilities)when the state of Y is not known than when it is: I(X,Z) > I(X,Z|Y ).On the contrary, if Y can date with equal probability one, another, bothor none, the states of X and Z are uncorrelated, but the knowledge of Yinduces correlation between X,Z: if we know that Y presently dates, thenit is enough to know that X does not to conclude that Z does. Note thatII(X, Y, Z) is symmetric.

Home exercise. What sign is the interaction information between i) clouds,rain and darkness, ii) dead car battery, broken fuel pump and failure to start theengine?

Answer: i) Positive interaction information appears for common-cause phenomena.

Clouds cause rain and also block the sun; therefore, the correlation between rain

and darkness is partly accounted for by the presence of clouds: I(rain, dark|cloud)

< I(rain, dark) and I(rain, dark, cloud) > 0. ii) Negative interaction information

appear for common effect phenomena. Generally, malfunctions of the fuel pump

and the battery are uncorrelated. However, if we know that the engine failed to

start (fix common effect), then the dependency appears: if the check shows that

the battery is OK, we infer that the pump must be broken.

Capturing dependencies by using structured groupings can be generalizedfor arbitrary number of variables as follows:

In =∑ni=1 S(Xi)−

∑ij S(Xi, Xj) +

∑ijk S(Xi, Xj, Xk)

−∑ijkl S(Xi, Xj, Xk, Xl) + . . .+ (−1)n+1S(X1, . . . , Xn) . (88)

Entropy, mutual information and interaction information are the first threemembers of that hierarchy.

An important property of both relative entropy and all In for n > 1 isthat they are independent of the additive constants in the entropies that isof the choice of units or bin sizes.

Connections to Statistical Physics. The second law of thermodyna-mics is getting trivial from the perspective of mutual information. We haveseen in Section 2.2 that even when we follow the evolution with infiniteprecision, the full N -particle entropy is conserved, but one particle entropygrows. Now we see that there is no contradiction here: subsequent colli-sions impose more and more correlation between particles, so that mutualinformation growth compensates that of one-particle entropy. Indeed, the

67

Page 68: Physical Nature of Information - Weizmann

thermodynamic entropy of the gas is the sum of entropies of different par-ticles

∑S(pi, qi). In the thermodynamic limit we neglect inter-particle cor-

relations, which are measured by the generalized (multi-particle) mutual in-formation

∑i S(pi, qi)−S(p1 . . . pn, q1, . . . qn) = I(p1, q1; . . . ; pn, qn). Deriving

the Boltzmann kinetic equation (38) in Section 2.2, we replaced two-particleprobability by the product of one-particle probabilities. That gave the H-theorem, that is the growth of the thermodynamic (uncorrelated) entropy.Since the Liouville theorem guarantees that the phase volume and the trueentropy S(p1 . . . pn, q1, . . . qn) do not change upon evolution, then the increaseof the uncorrelated part must be compensated by the increase of the mutualinformation. In other words, one can replace the usual second law of thermo-dynamics by the law of conservation of the total entropy (or information): theincrease in the thermodynamic (uncorrelated) entropy is exactly compensa-ted by the increase in correlations between particles expressed by the mutualinformation. The usual second law then results simply from our renunciationof all correlation knowledge, and not from any intrinsic behavior of dynami-cal systems. Particular version of such renunciation has been presented inSection 2.3: the full N -particle entropy grows because of phase-space mixingand continuous coarse-graining.

Relative entropy allows also to generalize the second law for non-equilibriumprocesses. Entropy itself can either increase upon evolution towards thermalequilibrium or decrease upon evolution towards a non-equilibrium state, asseen in Section 2.5. However, the relative entropy between the distributionand the steady-state distribution monotonously decreases with time. Also,the conditional entropy between values of any quantity taken at different ti-mes, S(Xt+τ |Xt), grows with τ when the latter exceeds the correlation time.

4 Applications of Information Theory

My brothers are protons, my sisters are neuronsGogol Bordello ”Supertheory of Supereverything”

This Chapter puts some content into the general notions introducedabove. Choosing out of enormous variety of applications, I tried to balancethe desire to show beautiful original works and the need to touch diverse sub-jects to let you recognize the same ideas in different contexts. The Chapteris concerned with practicality no less than with optimality; we often sacrificethe latter for the former.

68

Page 69: Physical Nature of Information - Weizmann

4.1 Distribution from information

So far, we defined entropy and information via the distribution. In practicalapplications, however, the distribution is usually unknown and we need toguess it from some data. The use of information does that. Statisticalphysics is a systematic way of guessing, making use of partial information.How to get the best guess for the probability distribution ρ(x, t), based onthe information given as 〈Rj(x, t)〉 = rj, i.e. as the expectation (mean)values of some dynamical quantities? Here we also include normalization:R0 = r0 = 1. Our distribution must contain the whole truth (i.e. all thegiven information) and nothing but the truth that is it must maximize themissing information, which is the entropy S = −〈ln ρ〉. This is to providefor the widest set of possibilities for future use, compatible with the existinginformation. Looking for the extremum of

S +∑j

λj〈Rj(x, t)〉 =∫ρ(x, t)

− ln[ρ(x, t)] +

∑j

λjRj(x, t)dx ,

we differentiate it with respect to ρ(x, t) and obtain the equation ln[ρ(x, t)] =−1 +

∑j λjRj(x, t) which gives the distribution

ρ(x, t) = exp[−1 +

∑j

λjRj(x, t)]. (89)

The normalization factor

Z(λi) = e1−λ0 =∫

exp[∑j=1

λjRj(x, t)]dx ,

can be expressed via the measured quantities by using

∂ lnZ

∂λi= ri . (90)

The distribution (89) corresponds to the entropy extremum, but how we knowthat it is the maximum? Positivity of relative entropy proves that. Indeed,consider any other normalized distribution g which satisfies the constraints.Then

∫dx g ln ρ =

∫dx ρ ln ρ = −S(ρ) and

S(ρ)− S(g) = −∫dx(g ln ρ− g ln g) =

∫dx g ln(g/ρ) = D(g|ρ) ≥ 0 .

69

Page 70: Physical Nature of Information - Weizmann

Gibbs distribution is (89) with R1 being energy. When it is the kineticenergy of molecules, we have Maxwell distribution; when it is potential energyin an external field, we have Boltzmann distribution. For our initial ”candy-in-the-box” problem (think of an impurity atom in a lattice if you preferphysics), let us denote the number of the box with the candy j. Differentattempts give different j but on average after many attempts we find, say,the mean value 〈j〉 = r1. The distribution giving maximal entropy for afixed mean is exponential, which in this case is the geometric distribution:ρ(j) = (1− p)pj, where p = r1/(1 + r1). Similarly, if we scatter on the latticeX-ray with wavenumber k and find 〈cos(kj)〉 = 0.3, then

ρ(j) = Z−1(λ) exp[−λ cos(kj)]

Z(λ) =n∑j=1

exp[λ cos(kj)] , 〈cos(kj)〉 = d logZ/dλ = 0.3 .

We can explicitly solve this for k 1 kn when one can approximate thesum by the integral so that Z(λ) ≈ nI0(λ) where I0 is the modified Besselfunction. Equation I ′0(λ) = 0.3I0(λ) has an approximate solution λ ≈ 0.63.

Note in passing that the set of equations (90) may be self-contradictory orinsufficient so that the data do not allow to define the distribution or allowit non-uniquely. For example, consider Ri =

∫xiρ(x) dx for i = 0, 1, 2, 3.

Then (89) cannot be normalized if λ3 6= 0, but having only three constantsλ0, λ1, λ2 one generally cannot satisfy the four conditions. That means thatwe cannot reach the entropy maximum, yet one can prove that we can comearbitrarily close to the entropy of the Gaussian distribution ln[2πe(r2−r2

1)]1/2.If, however, the extremum is attainable, then (89) defines the missing

information Sri = −∑j ρ(j) ln ρ(j), which is analogous to thermodynamicentropy as a function of (measurable) macroscopic parameters. It is clearthat S have a tendency to decrease whenever we add a constraint by measu-ring more quantities Ri. Making an extra measurement R one changes thedistribution from ρ(x) to (generally non-equilibrium) ρ(x|R), which has itsown conditional entropy

S(x|R) = −∫dxdR ρ(R)ρ(x|R) ln ρ(x|R) = −

∫dxdRρ(x,R) ln ρ(x|R) .

(91)The conditional entropy quantifies my remaining ignorance about x once Iknow R. Measurement decreases the entropy of the system by the mutual

70

Page 71: Physical Nature of Information - Weizmann

information (72,74) — that how much information about x one gains:

S(x|R)− S(x) = −∫ρ(x|R) ln ρ(x|R) dxdR +

∫ρ(x) ln ρ(x) dx

=∫ρ(x,R) ln[ρ(x,R)/ρ(x)ρ(R)] dxdR = S(x,R)− S(R)− S(x) .(92)

But all our measurements happen in a real world at a finite temperature.Does it matter? Assume that our system is in contact with a thermostathaving temperature T , which by itself does not mean that it is in thermalequilibrium (as, for instance, a current-carrying conductor). We then candefine a free energy F (ρ) = E−TS(ρ). If the measurement does not changeenergy (like the knowledge in which half of the box the particles is), then theentropy decrease (92) increases the free energy, so that the minimal work toperform such a measurement is F (ρ(x|R)) − F (ρ(x)) = T [S(x) − S(x|R)].We shall consider the energy price of information processing in more detailin Section 4.2.

If we know the given information at some time t1 and want to makeguesses about some other time t2 then our information generally gets lessrelevant as the distance |t1 − t2| increases. In the particular case of guessingthe distribution in the phase space, the mechanism of loosing informationis due to separation of trajectories described in Sect. 2.2. Indeed, if weknow that at t1 the system was in some region of the phase space, the setof trajectories started at t1 from this region generally fills larger and largerregions as |t1 − t2| increases. Therefore, missing information (i.e. entropy)increases with |t1 − t2|. Note that it works both into the future and into thepast. Information approach allows one to see clearly that there is really nocontradiction between the reversibility of equations of motion and the growthof entropy.

Yet there is one class of quantities where information does not age. Theyare integrals of motion. A situation in which only integrals of motion areknown is called equilibrium. The distribution (89) takes the canonical form(32,33) in equilibrium. On the other hand, taking micro-canonical as constantover the constant-energy surface corresponds to the same approach of notadding any additional information to what is known (energy).

From the information point of view, the statement that systems approachthermal equilibrium is equivalent to saying that all information is forgottenexcept the integrals of motion. If, however, we possess the information aboutaverages of quantities that are not integrals of motion and those averages do

71

Page 72: Physical Nature of Information - Weizmann

not coincide with their equilibrium values then the distribution (89) deviatesfrom equilibrium. Examples are currents, velocity or temperature gradientslike considered in kinetics.

Traditional way of thinking is operational: if we leave the system alone,it is in equilibrium; we need to act on it to deviate it from equilibrium.Informational interpretation lets us to see it in a new light: If we leave thesystem alone, our ignorance about it is maximal and so is the entropy, sothat the system is in thermal equilibrium; when we act on a system in a waythat gives us more knowledge of it, the entropy is lowered, and the systemdeviates from equilibrium.

We see that looking for the distribution that realizes the entropy extre-mum under given constraints is a universal powerful tool whose applicabilitygoes far beyond equilibrium statistical physics. A beautiful example of usingthis approach is obtaining the statistical distribution of the ensemble of neu-rons (Schneidman, Berry, Segev and Bialek, 2006). In a small window oftime, a single neuron either generates an action potential or remains silent,and thus the states of a network of neurons are described naturally by bi-nary vectors σi = ±1. Most fundamental results of measurements are themean spike probability for each cell 〈σi〉 and the matrix of pairwise correla-tions among cells 〈σiσj〉. One can successfully approximate the probabilitydistribution of σi by maximum entropy distribution (89) that is consistentwith the two results of the measurement. The probability distribution of theneuron signals that maximizes entropy is as follows:

ρ(σ) = Z−1 exp

∑i

hiσi +1

2

∑i<j

Jijσiσj

, (93)

where the Lagrange multipliers hi, Jij have to be chosen so that the averages〈σi〉, 〈σiσj〉 in this distribution agree with the experiment. Such models bearthe name Ising in physics, where they were first used for describing systemsof spins. The distribution (93) corresponds to the thermal equilibrium in therespective Ising model, yet it describes the brain activity, which is apparentlyfar from thermal equilibrium (unless the person is brain dead).

One can also measure some multi-cell correlations and check how well theyagree with those computed from (93). Despite apparent patterns of collectivebehavior, that involve many neurons, it turns out to be enough to account forpairwise correlations to describe the statistical distribution remarkably well. Thisis also manifested by the entropy changes: measuring triple and multi-cell cor-relations imposes more restrictions and lowers the entropy maximum. One then

72

Page 73: Physical Nature of Information - Weizmann

checks that accounting for pairwise correlations changes entropy significantly whileaccount for further correlation changes entropy relatively little. The sufficiency ofpairwise interactions provides an enormous simplification, which may be impor-tant not only for our description, but also for the brain. The reason is that brainactually develops and constantly modifies its own predictive model of probabilityneeded in particular to accurately evaluate new events for their degree of surprise.The dominance of pairwise interactions means that learning rules based on pair-wise correlations could be sufficient to generate nearly optimal internal models forthe distribution of codewords in the brain vocabulary, thus allowing the brain toaccurately evaluate probabilities. Side remark: we should not think that what isencoded from sensors into an electrical neuron activity is then ”decoded” insidethe brain. Whatever it is, brain is not computer.

It is interesting how the entropy scales with the number of interacting neurons

N . The entropy of non-interacting (or nearest-neighbor interacting) neurons is

extensive that is proportional to N . The data show that Jij are non-zero for

distant neurons as well. That means that the entropy of an interacting set is lower

at least by the sum of the mutual information terms between all pairs of cells.

The negative contribution is thus proportional to the number of interacting pairs,

N(N − 1)/2, that is grows faster with N , at least when it is not too large. One

can estimate from low-N data a ”critical” N when entropy is expected to turn

into zero and it corresponds well to the empirically observed sizes of the clusters of

strongly correlated cells. The lesson is: even when pairwise correlations are weak,

sufficiently large clusters can be strongly correlated. It is also important that the

interactions Jij have different signs, so that frustration can prevent the freezing of

the system into a single state (like ferromagnetic or anti-ferromagnetic). Instead

there are multiple states that are local minima of the effective energy function, as

in spin glasses.

4.2 Exorcizing Maxwell demon

Demon died when a paper by Szilard appeared, but it continuesto haunt the castles of physics as a restless and lovable poltergeist.

P Landsberg, quoted from Gleick ”The Information”

Here we want to address the relation between information and energy, particu-larly, find out if information has any energy price. Since energy and entropy (in-formation) have different dimensionalities, we need something to relate them. Forexample, this can be temperature, which is the derivative of the energy with respectto the entropy. That makes it natural to consider a system in contact with a ther-mostat, but not necessarily in thermal equilibrium. The Gibbs-Shannon entropy

73

Page 74: Physical Nature of Information - Weizmann

(63) and the mutual information (198,72,92) can be defined for arbitrary distribu-tions. As we mentioned after (4.2), one can then define a free energy for any systemin a contact with a thermostat having temperature T as F (ρ) = E(ρ) − TS(ρ),even when the distribution of the system itself is not equilibrium. Thermody-namics interprets F as the energy we are free to use keeping the temperature.Information theory reinterprets that in the following way: If we knew everything,we can possibly use the whole energy (to do work); the less we know about thesystem, the more is the missing information S and the less work we are able toextract. In other words, the decrease of F = E−TS with the growth of S measu-res how available energy decreases with the loss of information about the system.Maxwell understood that already in 1878: ”Suppose our senses sharpened to sucha degree that we could trace molecules as we now trace large bodies, the distinctionbetween work and heat would vanish.”

The concept of entropy as missing information15 (Brillouin 1949) allows oneto understand that Maxwell demon or any other information-processing device donot really decrease entropy. Indeed, if at the beginning one has an informationon position or velocity of any molecule, then the entropy was less by this amountfrom the start; after using and processing the information the entropy can onlyincrease. Consider, for instance, a particle in the box at a temperature T . If weknow in which half it is, then the entropy (the logarithm of available states) isln(V/2). That teaches us that information has thermodynamic (energetic) value:by placing a piston at the half of the box and allowing particle to hit and move itwe can get the work T∆S = T ln 2 out of thermal energy of the particle:

On the other hand, the law of energy conservation tells that to get such aninformation one must make a measurement whose minimum energetic cost at fixedtemperature is Wmeas = T∆S = T ln 2 (that was realized by Szilard in 1929 whoalso introduced ”bit” as a unit of information). Such work needs to be done forany entropy change by a measurement (92). This is true for an ideal (or demonic)observer, which does not change its state upon measurements. In a general case,

15that entropy is not a property of the system but of our knowledge about the system

74

Page 75: Physical Nature of Information - Weizmann

the entropy change is the difference between the entropy of the system S(A) andthe entropy of the system interacting with the measuring device S(A,M). Whenthere is a change in the free energy ∆FM of the measuring device, the measurementwork is

Wmeas ≥ T∆S + ∆FM = T [S(A)− S(A,M)] + ∆FM . (94)

That guarantees that we cannot break the first law of thermodynamics. Butwe just turned thermal energy into work. Can we then break the second law byconstructing a perpetuum mobile of the second kind, regularly measuring particleposition and using its thermal energy to do work? To make a full thermodynamiccycle of our demonic engine, we need to return the demon’s memory to the initialstate. What is the energy price of erasing information? Such erasure involves com-pression of the phase space and is irreversible. For example, to erase information inwhich half of the box the particle is, we may compress the box to move the particleto one half irrespective of where it was. That compression decreases entropy andis accompanied by the heat T ln 2 released from the system to the environment. Ifwe want to keep the temperature of the system, we need to do exactly that amountof work compressing the box (Landauer 1961). In other words, demon cannot getmore work from using the information S(M) than we must spend on erasing it toreturn the system to the initial state (to make a full cycle). More generally, wecan lower the work at the price of cooling the measuring device:

Weras ≥ TS(M)−∆FM . (95)

Together, the energy price of the cycle,

Weras +Wmeas ≥ T [S(A) + S(M)− S(A,M)] = TI , (96)

can be recognized as the temperature times what was defined in the Section 3.5 asthe mutual information. Thermodynamic energy cost of measurement and infor-mation erasure depends neither on the information content nor on the free-energydifference; rather the bound depends only on the mutual correlation between themeasured system and the memory. Inequality (96) expresses the trade off betweenthe work required for erasure and that required for measurement: when one issmaller, the other one must be larger. The relations (94,95,96) are versions of thesecond law of thermodynamics, in which information content and thermodynamicvariables are treated on an equal footing.

Similarly, in the original Maxwell scheme, the demon observes the moleculesas they approach the shutter, allowing fast ones to pass from A to B and slow onesfrom B to A. Creation of the temperature difference with a negligible expenditureof work lowers the entropy precisely by the amount of information that the demoncollected. Erasing this information will require work.

75

Page 76: Physical Nature of Information - Weizmann

Landauers principle not only exorcizes Maxwells demon, but also imposes thefundamental physical limit on computations. Performing standard operations inde-pendent of their history requires irreversible acts (which do not have single-valuedinverse). Any Boolean function that maps several input states onto the same out-put state, such as AND, NAND, OR and XOR, is logically irreversible. When acomputer does logically irreversible operation the information is erased and heatmust be generated. It is worth stressing that one cannot make this heat arbitrarilysmall making the process adiabatically slow: T ln 2 per bit is the minimal amountof dissipation to erase a bit at a fixed temperature16.

Take-home lesson: information is physical. Processing information withoutstoring an ever-increasing amount of it must be accompanied by a finite heat releaseat a finite temperature. Of course, any real device dissipates heat just because itworks at a finite rate. Lowering that rate one lowers the dissipation rate too. Themessage is that no matter how slowly we process information, we cannot makethe dissipation rate lower than T ln 2 per bit. This is in distinction from usualthermodynamic processes where there is no information processing involved andwe can make heat release arbitrarily small making the process slower.

Experiment. Despite its fundamental importance for information theory andcomputer science, the erasure bound has not been verified experimentally untilrecently, the main obstacle being the difficulty of doing single-particle experimentsin the low-dissipation regime (dissipation in present-day silicon-based computersstill exceeds the Landauer limit by a factor 102 ÷ 103 but goes down fast). Theexperiment realized erasure of a bit by treating colloidal particle in a double-wellpotential as a generic model of a one-bit memory (Berut et al, Nature 2012; Jun,Gavrilov, Bechhoefer, PRL 2014). The initial entropy of the system is thus ln 2.The procedure is to put the particle into the right well irrespective of its initialposition, see Figure below. It is done by first lowering the barrier height (Fig. b)and then applying a tilting force that brings the particle into the right well (Figce). Finally, the barrier is increased to its initial value (Fig f). At the end of thisreset operation, the information initially contained in the memory has been erasedand the entropy is zero.

16In principle, any computation can be done using only reversible steps, thus eliminatingthe need to do work (Bennett 1973). That will require the computer to reverse all thesteps after printing the answer.

76

Page 77: Physical Nature of Information - Weizmann

c 4

3

2

1

0

0 10 20 30 40

(s) Q

(kT

)

The heat/work was determined by experimentally observing the particle tra-jectory x(t) and computing the integral of the power using the known potentialU(x, t):

W = Q(τ) = −∫ τ

0x(t)

∂U(x, t)

∂xdt . (97)

This heat was averaged over 600 realizations. According to the second law ofthermodynamics,

〈Q〉 ≥ −T∆S = T ln 2 . (98)

One can see in the right panel of the figure above how the limit is approached asthe duration of the process increases. We shall return to the Brownian particle ina potential in Section 6.3 where we present a generalization of (94,95).

4.3 Flies and spies

What lies at the heart of every living thing is not a fire,not warm breath, not a ’spark of life.’ It is information.

Richard DawkinsOne may be excused thinking that living beings consume energy to survive,

unless one is a physicist and knows that energy is conserved and cannot be con-sumed. All the energy, absorbed by plants from sunlight and by us from food, isemitted as heat. Instead, we consume information and generate entropy by inter-cepting entropy flows to high-entropy body heat from low-entropy energy sources

77

Page 78: Physical Nature of Information - Weizmann

— just think how much information was processed to squeeze 500 kkal into 100grams of a chocolate, and you enjoy it even more17.

If an elementary act of life as information processing (say, thought) generates∆S, we can now ask about its energy price. Similar to our treatment of thethermal engine efficiency (8), we assume that one takes Q from the reservoir withT1 and delivers Q − W to the environment with T2. Then ∆S = S2 − S1 =(Q−W )/T2 −Q/T1 and the energy price is as follows:

Q =T2∆S +W

1− T2/T1.

T

T2

1

QS =Q/T1

WW

− QS =(Q−W)/T2

1

2

When T1 → T2, the information processing is getting prohibitively ineffective,just like the thermal engine. In the other limit, T1 T2, one can neglect theentropy change on the source, and we have Q = T2∆S +W . Hot Sun is indeed alow-entropy source.

Let us now estimate our rate of information processing and entropy production.A human being dissipates about W = 200 watts of power at T = 300K. Sincethe Boltzmann constant is k = 1.38 × 10−23, that gives about W/kT ' 1023 bitsper second. The amount of information processed per unit of subjective time (perthought) is about the same, assuming that each moment of consciousness lastsabout a second (Dyson, 1979).

We now discuss how such beings actually process information.

Maximizing capacity. Imagine yourself on the day five of Creation designingthe response function for a sensory system of a living being:

Response

Input

state

1

4

2

3

For given value intervals of input and response, should we take the solid lineof linear proportionality between response and stimulus? Or choose the lowestcurve that treats all weak-intensity inputs the same and amplifies difference in

17Nor we consume matter, only make it more disordered: what we consume has muchlower entropy than what comes out of our excretory system

78

Page 79: Physical Nature of Information - Weizmann

strong signals? The choice depends on the goal. For example, the upper curve wasactually chosen (on the day six) for the auditory system of animals and humans:our ear amplifies differences in weak sounds and damp strong ones, sensing loudnessas the logarithm of the intensity. That way we better hear whisper of a close oneand aren’t that frightened by loud threats.

If, however, the goal is to maximize the mean information transfer rate (capa-city) at the level of a single neuron/channel, then the response curve (encoding)must be designed by the Creator together with the probability distribution of vi-sual stimuli. That it is indeed so was discovered in probably historically the firstapplication of information theory to the real data in biology (Laughlin 1981). Itwas conjectured that maximal-capacity encoding must use all response levels withthe same frequency, which requires that the response function is an integral of theprobability distribution of the input signals (see Figure). First-order interneuronsof the insect eye were found to code contrast rather than absolute light inten-sity. Subjecting the fly in the lab to different contrasts x, the response functiony = g(x) was measured from the fly neurons; the probability density of inputs,ρ(x), was measured across its natural habitat (woodlands and lakeside) using adetector which scanned horizontally, like a turning fly.

The coding strategy for maximizing information capacity by ensuring that all response levels are used with equal

frequency. Upper left curve: probability density function for stimulus intensities. Lower left curve: the response function,

which ensures that the interval between each response level encompasses an equal area under the distribution, so that

each state is used with equal frequency. In the limit where the states are vanishingly small this response function

corresponds to the cumulative probability function. Right panel: The contrast-response function of fly neuron compared to the cumulative probability function for natural contrasts. Simon Laughlin, Naturforsch. 36, 910-912 (1981)

We can now explain it noting that the representation with the maximal capa-city corresponds to the maximum of the mutual information between input andoutput: I(x, y) = S(y)− S(y|x). Originally, it was assumed that the transmissionis error-free, so that the conditional entropy S(y|x) is zero, but more realistically,

79

Page 80: Physical Nature of Information - Weizmann

we can assume that it is fixed anatomically and does not depend on the inputstatistics. Therefore, according to Section 3.5, we need to maximize the entropyof the output assuming that the input statistics ρ(x) is given. Absent any extraconstraints except normalization, the entropy is maximal when ρ(y) is constant.Indeed, since ρ(y)dy = ρ(x)dx = ρ(x)dydx/dy = ρ(x)dy/g′(x), then

S(y) = −∫ρ(x) ln[ρ(x)/g′(x)] dx = S(x) + 〈ln[g′(x)]〉 , (99)

δS

δg=

∂x

ρ

g′(x)= 0 ⇒ g′(x) = ρ(x) ,

as in the Figure. Since the probability is positive, the response function y = g(x)is always monotonic i.e. invertible. Note that we utilized only the probabilitydistribution of different signal levels, similar to language encoding which utili-zes different frequencies of letters (and not, say, their mutual correlations). Wehave also applied quasi-static approximation, neglecting dynamics. Let yourselfbe impressed by the agreement of theory and experiment — there were no fittingparameters. The same approach works well also for biochemical and genetic input-output relations. For example, the dependence of a gene expression on the level ofa transcription factor is dictated by the statistics of the latter. That works whenthe conditional entropy S(y|x) is independent of the form of the response functiony = g(x).

This approach turns out quite useful in image and speech recognition by com-puters using unsupervised learning. That can be done by considering the ”trainingset” of x-s to approximate the density ρ(x). We then choose some form of the re-sponse function y = g(x,w) characterized by the parameter w and find optimalvalue of w using an ”online” stochastic gradient ascent learning rule giving thechange of the parameter:

∆w ∝ ∂

∂wln

(∂g(x,w)

∂x

). (100)

For example, the form of the response function y(u) popular for its flexibility isthe two-parametric asymmetric generalized logistic function defined implicitly bydy/du = yp(1− y)r. Symmetric examples are given in the Figure, choosing p 6= rone can work with skewed distributions as well.

80

Page 81: Physical Nature of Information - Weizmann

0

1

u

y=g

(u)

p=1, r=1

dy/

du

u

0

u

y=g

(u)

p=5, r=5

u

0

1

y=g

(u) p=0.2, r=0.2

u

dy/

du

u

dy/

du

(a) (b) 1 (c)

After choosing the form of g, we need to properly center and normalize theoutput signal u = wx+w0. Using (100) one trains computer to optimize the dataprocessing with respect to all the parameters.

Of course, the eye of any living being provides not a single input signal, but thewhole picture. Let us now pass from a single channel to N inputs and outputs (neu-rons/channels). Consider a network with an input vector x = (x1, . . . , xN ) whichis transformed into the output vector y(x) monotonically, that is det[∂yi/∂xk] 6= 0.The multivariate probability density function of y is as follows:

ρ(y) =ρ(x)

det[∂yi/∂xk], (101)

Making it flat (distribute outputs uniformly) for maximal capacity is not straig-htforward now. In one dimension, it is enough to follow the gradient to arrive atan extremum, but there are many possible paths to the mountain summit. Max-imizing the total mutual information between input and output, which requiresmaximizing the output entropy, is often (but not always) achieved by minimizingfirst the mutual information between the output components. For two outputs wemay start by maximizing S(y1, y2) = S(y1) + S(y2) − I(y1, y2), that is minimizeI(y1, y2). If we are lucky and find encoding in terms of independent components,then we choose for each component the transformation (99), which maximizes itsentropy making the respective probability flat. For a good review and specificapplications to visual sensory processing see Atick 1992.

For particular types of signals, practicality may favor non-optimal but simpleschemes like amplitude and frequency modulation (both are generally non-optimalbut computationally feasible and practical). Even in such cases, the choice is dic-tated by the information-theory analysis of the efficiency. For example, neuroneither fires a standard pulse (action potential) or stays silent, which makes it na-

81

Page 82: Physical Nature of Information - Weizmann

tural to assume that the information is encoded as binary digits (zero or one) indiscrete equal time intervals. Yet one can imagine that the information is enco-ded by the time delays between subsequent pulses. On the engineer’s language,the former method of encoding is a limiting case of amplitude modulation, whilethe latter case is that of frequency modulation. The maximal rate of informationtransmission is the former case is only dependent on the minimal time delay bet-ween the pulses determined by the neuron recovery time. On the other hand, inthe latter case, the rate depends on both the minimal error of timing measurementand of admissible maximal time between pulses. In the home exercise, you shallcompare the two schemes. In reality, brain activity ”depends in one way or anotheron all the information-bearing parameters of an impulse — both on its presenceor absence as a binary digit and on its precise timing” (MacKay and McCulloch1952).

Minimizing correlation between components. Finding least correlatedcomponents can be a practical first step in maximizing capacity. Note how tomaximize the mutual information between input and output, we minimize themutual information between the components of the output. This is particularlytrue for natural signals where most redundancy comes from strong correlations(like that of the neighboring pixels in visuals). In addition, finding an encodingin terms of least dependent components is important by itself for its cognitiveadvantages. For example, such encoding generally facilitates pattern recognition.In addition, presenting and storing information in the form of independent (orminimally dependent) components is important for associative learning done bybrains and computers. Indeed, for an animal or computer to learn a new associationbetween two events, A and B, the brain should have knowledge of the prior jointprobability P (A,B). For correlated N -dimensional A and B one needs to storeN × N numbers, while only 2N numbers for quantities uncorrelated (until theassociation occurs).

Another cognitive task is the famous ”cocktail-party problem” posed by spies:N microphones (flies on the wall) record N people speaking simultaneously, andwe need the program to separate them — so-called blind separation problem. Herewe assume that uncorrelated sources s1, . . . , sN are mixed linearly by an unknownmatrix A. All we receive are the N superpositions of them x1, . . . , xN . Thetask is to recover the original sources by finding a square matrix W which is theinverse of the unknown A, up to permutations and re-scaling. Closely related isthe blind de-convolution problem also illustrated in the Figure below (from Belland Sejnowski, 1995): a single unknown signal s(t) is convolved with an unknownfilter giving a corrupted signal x(t) =

∫a(t− t′)s(t′) dt′, where a(t) is the impulse

response of the filter. The task is to recover s(t) by integrating x(t) with the inverse

82

Page 83: Physical Nature of Information - Weizmann

filter w(t), which we need to find by learning procedure. Upon discretization, s, xare turned into N -vectors and w into N × N matrix, which is lower triangularbecause of causality: wij = 0 for j > i and the diagonal values are all the samewii = w. The determinant in (101) is simplified in this case. For y = g(wx)we have det[∂y(ti)/∂x(tj)] = det w

∏Ni y′(ti) = wN

∏Ni y′(ti). One that applies

some variant of (100) to minimize mutual information. What was described is asingle-layer processing. More powerful are multi-layer nonlinear schemes, wherecomputing determinants and formulating learning rules is more complicated.

(b)

unknown mixing process

BLIND SEPARATION(learnt weights)

A Wx u

(a) s

a(t)

w(t)

u(t)BLIND DECONVOLUTION

x(t)unknown convolution

s(t)

Ideally, we wish to find the (generally stochastic) encoding y(x) that achievesthe absolute minimum of the mutual information

∑i S(yi) − S(y). One way to

do that is to minimize the first term while keeping the second one, that is undercondition of the fixed entropy S(y) = S(x). In general, one may not be able to findsuch encoding without any entropy change S(y)−S(x). In such cases, one definesa functional that grades different codings according to how well they minimize boththe sum of the entropies of the output components and the entropy change. Thesimplest energy functional for statistical independence is then

E =∑i

S(yi)− β[S(y)− S(x)] =∑i

S(yi)− β ln det[∂yi/∂xk] . (102)

A coding is considered to yield an improved representation if it possesses a smallervalue of E. The choice of the parameter β reflects our priorities — whetherstatistical independence or increase in indeterminacy is more important. Similarminimization procedures will be considered in the next Section.

Maximizing information transfer and reducing the redundancy between theunits in the output is applied practically in all disciplines that analyze and processdata, from physics and engineering to biology, psychology and economics. So-metimes it is called infomax principle, the specific technique is called independent

83

Page 84: Physical Nature of Information - Weizmann

component analysis (ICA). More sophisticated schemes employs not only mutualinformation, but also interaction information (87). Note that the redundancy re-duction is usually applied after some procedure of eliminating noise. Indeed, ourgain function provides equal responses for probable and improbable events, but thelatter can be mostly due to noise, which thus needs to be suppressed. Moreover,if input noises were uncorrelated, they can get correlated after coding. And moregenerally, it is better to keep some redundancy for corrections and checks whendealing with noisy data.

4.4 Rate Distortion and Information Bottleneck

When we transfer information, we look for maximal transfer rate and thus definechannel capacity as the maximal mutual information between input and output.But when we encode the information, we may be looking for the opposite: what isthe minimal number of bits, sufficient to encode the data with a given accuracy.

For example, description of a real number requires infinite number of bits.Representation of a continuous input B by a finite discrete rate of the outputencoding generally leads to some distortion, which we shall characterize by the realfunction d(A,B). How large is the mean distortion D =

∑ij P (Ai, Bj)d(Ai, Bj)

for a given encoding with R bits and 2R values? It depends on the choice of thedistortion function, which specifies what are the most important properties of thesignal B. For Gaussian statistics (which is completely determined by the variance),one chooses the squared error function d(A,B) = (A− B)2. We first learn to useit in the standard least squares approximations — now we can understand why— because minimizing variance minimizes the entropy of a Gaussian distribution.Consider a Gaussian B with 〈B〉 = 0 and 〈B2〉 = σ2. If we have one bit torepresent it, apparently, the only information we can convey is the sign of B.To minimize squared error, we encode positive/negative values by A = ±σ

√2/π,

which corresponds to

D(1) = (2π)−1/2∫ ∞

0

(B − σ

√2/π

)2exp[−B2/2σ2]

dB

σ= σ2/4 .

Let us now turn the tables and ask what is the minimal rate R sufficient to providefor distortion not exceeding D. This is called rate distortion function R(D). Weknow that the rate is the mutual information I(A,B), but now we are lookingnot for its maximum (as in channel capacity) but for the minimum over all theencodings defined by P (B|A), such that the distortion does not exceed D. It ishelpful to think of distortion as produced by the added noise ξ with the varianceD. For a fixed variance, maximal entropy S(B|A) corresponds to the Gaussiandistribution, so that we have a Gaussian input with 〈B2〉 = σ2 plus (imaginary)

84

Page 85: Physical Nature of Information - Weizmann

Gaussian channel with the variance 〈(B−A)2〉 = D, and the minimal rate is givenby (79):

R(D) = I(A,B) = S(B)− S(B|A) = S(B)− S(B −A|A) ≥ S(B)− S(B −A)

=1

2log2(2πeσ2)− 1

2log2(2πeD) =

1

2log2

σ2

D. (103)

It turns into zero for D = σ2 and goes to infinity for D → 0. Presenting it asD(R) = σ22−2R we see that every bit of description reduces distortion by a factorof 4.

One can show that the rate distortion function R(D) is monotonous and convexfor all systems. In solving practical problems, it is usually found as a solution of thevariational problem, where one needs to find normalized P (B|A) which minimizesthe functional

F = I + βD =∑ij

P (Bj |Ai)P (Ai)

lnP (Bj |Ai)P (Bj)

+ βd(Ai, Bj)

. (104)

After variation with respect to P (Bj |Ai) and accounting for the identity P (Bj) =∑i P (Bj |Ai)P (Ai) we obtain

P (Bj |Ai) =P (Bj)

Z(Ai, β)e−βd(Ai,Bj) , (105)

where the partition function Z(Ai, β) =∑j P (Bj)e

−βd(Ai,Bj) is the normaliza-tion factor. Immediate physical analogy is that this is a Gibbs distribution withthe ”energy” equal to the distortion function. Maximizing entropy for a givenenergy (Gibbs) is equivalent to minimizing mutual information for a given dis-tortion function. Choice of the value of the inverse temperature β reflects ourpriorities: at small β the conditional probability is close to the unconditional one,that is we minimize information without much regard to the distortion. On thecontrary, large β requires our conditional probability to be sharply peaked at theminima of the distortion function.

Similar, but more sophisticated optimization procedures are applied, in parti-cular, in image processing. Images contain enormous amount of information. Therate at which visual data is collected by the photoreceptor mosaic of animals andhumans is known to exceed 106 bits/sec. On the other hand, studies on the speedof visual perception and reading speeds give numbers around 40-50 bits/sec forthe perceptual capacity of the visual pathway in humans. The brain then haveto perform huge data compressions. This is possible because visual information ishighly redundant due to strong correlations between pixels. Mutual informationis the main tool in the theory of (image, voice, pattern) recognition and AI.

85

Page 86: Physical Nature of Information - Weizmann

B A Cevent measurement encoding

The measured quantity A thus contains too much data of low information value.We wish to compress A to C while keeping as much as possible information aboutB. Understanding the given signal A requires more than just predicting/inferringB, it also requires specifying which features of the set of possible signals A playa role in the prediction. Here meaning seeps back into the information theory.Indeed, information is not knowledge (and knowledge is not wisdom). We formalizethis problem as that of finding a short code for A that preserves the maximuminformation about the set B. That is, we squeeze the information thatA providesabout B through a ”bottleneck” formed by a limited set of codewords C. This isreached via the method called Information Bottleneck, targeted at characterizingthe tradeoff between information preservation (accuracy of relevant predictions)and compression. Here one looks for the minimum of the functional

I(C,A)− βI(C,B) . (106)

The coding A → C is also generally stochastic, characterized by P (C|A). Thequality of the coding is determined by the rate, that is by the average number ofbits per message needed to specify an element in the codebook without confusion.This number per element A of the source space A is bounded from below bythe mutual information I(C,A) which we thus want to minimize. Effective codingutilizes the fact that mutual information is usually sub-extensive in distinctionfrom entropy which is extensive. Note the difference from the Section 3.5, wherein characterizing the channel capacity (upper bound for the error-free rate) wemaximized I(A,B) over all choices of the source space B, while now we mini-mize I(C,A) over all choices of coding. To put it differently, there we wanted tomaximize the information transmitted, now we want to minimize the informationprocessed. This minimization, however, must be restricted by the need to retainin C the relevant information about B which we denote I(C,B). Having cho-sen what properties of B we wish to stay correlated with the encoded signal C,we add the mutual information I(C,B) with the Lagrange multiplier −β to thefunctional (106). The sign is naturally chosen such that β > 0 (analog of inversetemperature), indeed, we want minimal coding I(A,B) preserving maximal infor-mation I(C,B) (that is I(C,B) is treated similarly to the channel capacity in theprevious section). The single parameter β again represents the tradeoff betweenthe complexity of the representation measured by I(C,A), and the accuracy ofthis representation, measured by I(C,B). At β = 0 our quantization is the mostsketchy possible — everything is assigned to a single point. At β grows, we arepushed toward detailed quantization. By varying β one can explore the tradeoff

86

Page 87: Physical Nature of Information - Weizmann

between the preserved meaningful information and compression at various resolu-tions. Comparing with the rate distortion theory functional (105), we recognizethat we are looking for the conditional probability of the mapping P (C|A), thatis we explicitly want to treat some pixels as more relevant than the others.

However, the constraint on the meaningful information is now nonlinear inP (C|A), so this is a much harder variational problem. Indeed, variation of (106)with respect to the conditional probability now gives the equation (rather than anexplicit expression):

P (Cj |Ai) =P (Cj)

Z(Ai, β)exp

[−β

∑k

P (Bk|Ai) logP (Bk|Ai)P (Bk|Cj)

], (107)

The conditional probability in the exponent is given by the Bayes’ rule

P (Bk|Cj) =1

P (Cj)

∑i

P (Ai)P (Bk|Ai)P (Cj |Ai) , (108)

Technically, this system of equations is usually solved by iterations, for instance,via deep learning (which is one of the most successful paradigms for unsupervisedlearning to emerge over the last 15 years). Doing compression procedure manytimes, A→ C1 → C2 . . . is used in multi-layered Deep Learning Algorithms. Hereknowledge of statistical physics helps in several ways, particularly in identifyingphase transitions (with respect to β) and the relation between processing fromlayer to layer and the renormalization group: features along the layers becomemore and more statistically decoupled as the layers gets closer to the fixed point.

Practical problems of machine learning are closely related to fundamental pro-blems in understanding and describing the biological evolution. Here an importanttask is to identify classes of functions and mechanisms that are provably evolvable— can logically evolve into existence over realistic time periods and within rea-listic populations, without any need for combinatorially unlikely events to occur.Quantitative theories of evolution in particular aim to quantify the complexity ofthe mechanisms that evolved, which is done using information theory.

4.5 Information is money

This section is for those brave souls who decided to leave physics for gambling. Ifyou have read till this point, you must be well prepared for that.

Let us start from the simplest game: you can bet on a coin, doubling your bet ifyou are right or loosing it if you are wrong. Surely, an intelligent person would notbet money hard saved during graduate studies on a totally random process witha zero gain. You bet only when you have an information that sides have unequal

87

Page 88: Physical Nature of Information - Weizmann

probabilities: p > 1/2 and 1 − p. Betting all your money on the more probableside and retiring after one bet is not an option for two fundamental reasons: youhaven’t saved enough and you don’t like a nonzero probability to loose it. To havea steady income and an average growth you want to play the game many times. Toavoid loosing it all, you bet only a fraction f of your money on the more probablep-side. What to do with the remaining money, keep it as an insurance or bet on aless probable side? The first option just diminishes the effective amount of moneythat works. Moreover, the other side also wins sometimes, so we put 1− f on theside with 1 − p chance. If after N such bets the p-side came n times then yourmoney is multiplied by the factor (2f)n[2(1− f)]N−n = exp(NΛ), where the rateis

Λ(f) = ln 2 +n

Nln f +

(1− n

N

)ln(1− f) . (109)

As N →∞ we approach the mean rate, which is λ = ln 2 + p ln f + (1− p) ln(1−f). Note the similarity with the Lyapunov exponents from Sections 3.3–3.5 —we consider the logarithm of the exponentially growing factor since we knowlimN→∞(n/N) = p (it is called self-averaging quantity because it is again a sumof random numbers). Differentiating Λ(f) with respect to f you find that themaximal growth rate corresponds to f = p (proportional gambling) and equals to

λ(p) = ln 2 + p ln p+ (1− p) ln(1− p) = S(u)− S(p) , (110)

where we denoted the entropy of the uniform distribution S(u) = ln 2. We thussee that the maximal rate of money growth equals to the entropy decrease, thatis to the information you have (Kelly 1950). What is beautiful here is that theproof of optimality is constructive and gives us the best betting strategy. Animportant lesson is that we maximize not the averaged return but its logarithm,which is a geometric mean. The geometric mean is less than the arithmetic mean.Therefore, we may have a situation when the arithmetic growth rate is larger thanunity while the geometric mean is smaller than unity. That would be a disaster,since the probability to loose it all will tend to unity as N → ∞, even thoughthe mean returns grows unbounded. For example, the choice f = 1 gives youthe maximal mean arithmetic growth rate (2p)N , all coming from a single all-winrealization; the probability of that realization goes to zero as pN .

It is straightforward to generalize (110) for gambling on horse races, wheremany outcomes have different probabilities pi and payoffs gi. Maximizing

∑pi ln(figi)

we find fi = pi independent of gi, so that

λ(p, g) =∑i

pi ln(pigi) . (111)

Here you have a formidable opponent - the track operator, who actually sets thepayoffs. Knowing the probabilities, an ideal operator can set the payoffs, gi = 1/pi,

88

Page 89: Physical Nature of Information - Weizmann

to make the game fair and your rate zero. More likely is that the real operator hasbusiness sense to make the racecourse profitable by setting the payoffs a bit lowerto make your λ negative. Your only hope then is that your information is better.Indeed, if the operator assumes that the probabilities are qi and sets payoffs asgi = 1/Zqi with Z > 1, then

λ(p, q) = − lnZ +∑i

pi ln(pi/qi) = − lnZ +D(p|q) . (112)

That is if you know the true distribution but the operator uses the approximateone, the relative entropy D(p|q) determines the rate with which your winningscan grow. Nobody’s perfect so maybe you use the distribution q′, which is notthe true one. In this case, you still have a chance if your distribution is closer tothe true one: λ(p, q, q′) = − lnZ + D(p|q) − D(p|q′). Remind that the entropydetermines the optimal rate of coding. Using incorrect distribution incurs the costof non-optimal coding. Amazingly, (112) tells that if you can encode the datadescribing the sequence of track winners shorter than the operator, you get paidin proportion to that shortening.

To feel less smug, note that bacteria follow the same strategy without evertaking this or other course on statistical physics. Indeed, analogously to coinflipping, bacteria often face the choice between growing fast but being vulnerable toantibiotic or grow slow but being resistant. They then use proportional gamblingto allocate respective fractions of populations to different choices. There couldbe several lifestyle choices, which is analogous to horse racing problem (calledphenotype switching in this case). The same strategy is used by many plants,where the fraction of the seeds do not germinate in the same year they weredispersed; the fraction increases together with the environment variability.

Bacteria, plants and gamblers face the problem we haven’t mentioned yet:acquiring information, needed for proportional gambling, has its own cost. Onethen looks for a tradeoff between maximizing growth and minimizing informationcost. Assume that the environment is characterized by the parameter A, say, theconcentration of a nutrient. The internal state of the bacteria is characterized byanother parameterB, which can be the amount of enzyme needed to metabolize thenutrient. The growth rate is then the function of these two parameters r(A,B). Weare looking for the conditional probability P (B|A), which determines the mutualinformation between the external world and the internal state:

I(A,B) =

∫dAP (A)

∫dBP (B|A) log2

P (B|A)

P (B). (113)

To decrease the cost aI of acquiring this information, we wish to let P (B|A) closer

89

Page 90: Physical Nature of Information - Weizmann

to P (B). Yet we also wish to maintain the average growth rate

λ =

∫dAP (A)

∫dB P (B|A)r(A,B) . (114)

Therefore, we look for the maximum of the functional F = λ − aI, which givessimilarly to (104,105)

P (B|A) =P (B)

Z(A, β)eβr(A,B) , (115)

where β = a−1 ln 2 and the partition function Z(A, β) =∫dBP (B)eβr(A,B) is

the normalization factor. We now recognize the rate distortion theory from theprevious subsection; the only difference is that the energy now is minus the growthrate. The choice of β reflects relative costs of the information and the metabolism.If information is hard to get, one chooses small β, which makes P (B|A) weaklydependent of r(A,B) and close to unconditional probability. If information ischeaper, (115) tells us that we need to peak our conditional probability around themaxima of the growth rate. All the possible states in the plane r, I are below somemonotonic convex curve, much like in the energy-entropy plane in Section 1.1. Onecan reach optimal (Gibbs) state on the boundary either by increasing the growthrate at a fixed information of by decreasing the information at a fixed growth rate.

Economic activity of humans is not completely reducible to gambling and itsessence understood much less. When you earn enough money, it may be a goodtime to start thinking about the nature of money itself. Money appeared firstas a measure of value, it acquired probabilistic aspect with the development ofcredit. These days, when most of it is in bits, it is clear that this is not matter(coins, banknotes) but information. Moreover, the total amount of money growson average, but could experience sudden drops when the crisis arrives. Yet inpayments money behaves as energy, satisfying the conservation law. I have afeeling that we need a new concept for describing money, which has properties ofboth entropy and energy.

Concluding this Chapter, which is essentially a long list of representation ex-amples of various efficiency, mention briefly the standard problem of choosing howmany fitting parameters to choose. While it is intuitively clear that one should notuse too many parameters for too few data points, mutual information makes thischoice precise. If we work with a given class of functions (say, Fourier harmonics)then it is clear that increasing the number K of functions we can approximate ourN points better and better, so that the deviation D(K) is a decreasing function ofK. But we know that our data contain noise so it does not make much sense to ap-proximate every fluctuation. In other words, every fitting parameter introduces itsown entropy si (for Fourier harmonics and white noise, we have all si the same).Technically, we need to minimize the mutual information of the representation,

90

Page 91: Physical Nature of Information - Weizmann

which would consist of two parts: ND(K) +∑Ki=1 si. Here the first term comes

from an imperfect data fitting, so it contains the relative entropy D(K) betweenour hypothetical distribution and the true one, while the second term is the en-tropy related to our K degrees of freedom. Extremum comes from a competitionbetween two terms. When we obtain more data and N is getting large, the valueof K, which gives a minimum, usually saturates.

5 Renormalization group and entanglement

entropy

5.1 Analogy between quantum mechanics and statisti-cal physics

Many aspects of quantum world are bizarre and have no classical analog. And yetthere are certain technical similarities between descriptions of quantum random-ness and thermal noise, related to the necessity of summing over different possi-bilities. One can see one such similarity using the formalism of the path integral,where one sums over different trajectories, — that will be briefly discussed in theSection 6.1 below. Here we describe the similarity in using the formalism of thetransfer matrix for the systems with nearest neighbor interaction, which we shallalso need in the next subsection. Indeed, in a simplest setting, quantum mechanicsis done by specifying two sets of states |q〉 and 〈p|, which have ortho-normalityand completeness: 〈p|q〉 = δqp and

∑q |q〉〈q| = 1. Physical quantities are represen-

ted by operators, and measurement corresponds to taking a trace of the operatorover the set of states: traceP=

∑q〈q|P |q〉. One special operator, called Hamil-

tonian H, determines the temporal evolution of all other operators according toP (t) = exp(iHt)P (0) exp(−iHt). The operator T (t) = exp(iHt) is called timetranslation operator or evolution operator. The quantum-mechanical average ofany operator Q is calculated as a trace with the evolution operator normalized bythe trace of the evolution operator:

〈Q〉 =traceT (t)Q

Z(t), Z(t) = traceT (t) =

∑a

e−itEa . (116)

The normalization factor is naturally to call the partition function, all the more ifwe formally consider it for an imaginary time t = iβ

Z(β) = traceT (iβ) =∑a

e−βEa . (117)

91

Page 92: Physical Nature of Information - Weizmann

If the inverse ”temperature” β goes to infinity then all the sums are dominated bythe ground state, Z(β) ≈ exp(−βE0) and the average in (117) are just expectationvalues in the ground state.

That quantum mechanical description can be compared with the transfer-matrix description of the Ising model (which was formulated by Lenz in 1920and solved in one dimension by his student Ising in 1925). It deals with the dis-crete spin variable σi = ±1 at every lattice site. The energy includes interactionwith the external field and between nearest neighbors (n.n.):

H =J

2

N−1∑i=1

(1− σiσi+1) . (118)

It is better to think not about spins but about the links between spins. Startingfrom the first spin, the state of the chain can be defined by saying whether thenext one is parallel to the previous one or not. If the next spin is opposite it givesthe energy J and if it is parallel the energy is zero. There are N − 1 links. Thepartition function is that of the N − 1 two-level systems:

Z = 2[1 + exp(−βJ)]N−1 . (119)

Here 2 because there are two possible orientations of the first spin.To avoid considering the open ends of the chain (which must be irrelevant in

the thermodynamic limit), we consider it on a ring so that σN+1 = σ1 and writethe partition function as a simple sum over spin value at every cite:

Z =∑σi

exp

[−βJ

2

N−1∑i=1

(1−σiσi+1)

](120)

=∑σi

N−1∏i=1

exp

[−βJ

2(1−σiσi+1)

](121)

Every factor in the product can have four values, which correspond to four differentchoices of σi = ±1, σi+1 = ±1. Therefore, every factor can be written as a matrixelement of 2 × 2 matrix: 〈σj |T |σj+1〉 = Tσjσj+1 = exp[−βJ(1 − σiσi+1)/2]. It iscalled the transfer matrix because it transfers us from one cite to the next.

T =

(T1,1 T1,−1

T−1,1 T−1,−1

)(122)

where T11 = T−1,−1 = 1, T−1,1 = T1,−1 = e−βJ . For any matrices A, B thematrix elements of the product are [AB]ik = AijBjk. Therefore, when we sum

92

Page 93: Physical Nature of Information - Weizmann

over the values of the intermediate spin, we obtain the matrix elements of thematrix squared:

∑σi Tσi−1σiTσiσi+1 = [T 2]σi−1σi+1 . The sum over N −1 spins gives

TN−1. Because of periodicity we end up with summing over a single spin whichcorresponds to taking trace of the matrix:

Z =∑σi

Tσ1σ2Tσ2σ3 . . . TσNσ1 =∑

σ1=±1

〈σ1|TN−1|σ1〉 = traceTN−1 . (123)

The eigenvalues λ1, λ2 of T are given by

λ1,2 = 1± e−βJ . (124)

The trace is the sum of the eigenvalues

Z = λN−11 + λN−1

2 . (125)

Therefore

F = −T log(λN−11 + λN−1

2 ) = −T[(N − 1) log(λ1)

+ log

(1 +

(λ2

λ1

)N−1)]→ −NT log λ1 as N →∞ (126)

Note that the partition functions (125) and (119) give the same free energies onlyat the thermodynamics limit when a ring is indistinguishable from a chain withopen ends.

We thus see that taking the sum over two values of σ at every cite in the Isingmodel is the analog of taking trace in quantum-mechanical average. If there arem values on the cite, then T is m×m matrix. For a spin in n-dimensional space(described by so-called O(n) model), trace means integrating over orientations.The translations along the chain are analogous to quantum-mechanical translationsin (imaginary) time. This analogy is not restricted to 1d systems, one can consider2d strips that way too.

5.2 Renormalization group and information loss

Erase the features Chance installed,and you will see the world’s great beauty18.

A BlokStatistical physics in general is about lack of information. One of the most

fruitful ideas of the 20-th century is to look how one looses information step by

18Erase the features Chance installed. Watch by chance do not rub a hole. V Nekrasov

93

Page 94: Physical Nature of Information - Weizmann

step and what universal features appear in the process. Most often we looseinformation about microscopic properties. We can do that by averaging over small-scale fluctuations in a procedure called coarse-graining. A general formalism whichdescribes how to make a coarse-graining to keep only most salient features in thedescription is called the renormalization group (RG). It consists in subsequentlyeliminating degrees of freedom, renormalizing remaining ones and looking for fixedpoints of such a procedure. There is a dramatic shift of paradigm brought bythe renormalization group approach. Instead of being interested in this or thatprobability distribution, we are interested in different RG-flows in the space ofdistribution. Whole families (universality classes) of different systems describedby different distribution flow under RG transformation to the same fixed point i.e.have the same asymptotic behavior.

Let us first describe the simplest case of RG, where we consider a set of randomiid variables x1 . . . xN, each having the probability density ρ(x) with zero meanand unit variance. The two-step RG reduces the number of random variables byreplacing any two of them by their sum and re-scales the sum to keep the variance:zi = (x2i−1 + x2i)/

√2. Since summing doubles the variance we divided by

√2.

The new random variables each has the following distribution:

ρ′(z) =√

2

∫dxdyρ(x)ρ(y)δ(x+ y − z

√2) . (127)

There are three conservation laws of that transformation,∫xmρ(x) dx, which are

normalization (m = 0), zero mean (m = 1) and unit variance (m = 2). Thedistribution which does not change upon such procedure is called fixed point (eventhough it is not a point but rather a whole function) and satisfies the equation

ρ(x) =√

2

∫dyρ(y)ρ(

√2x− y) .

Since this is a convolution equation, the simplest is to solve it by the Fouriertransform, ρ(k) =

∫ρ(x)eikxdx, which gives

ρ(k√

2) = ρ2(k) . (128)

We may also say that ρ(k) is the generating function, which is multiplied uponsummation of independent variables. The solution of (128) is ρ0(k) ∼ e−k

2and

ρ0(x) = (2π)−1/2e−x2/2. We thus have shown that the Gaussian distribution is a

fixed point of repetitive summation and re-scaling of random variables. To turnthat into the central limit theorem, we need also to show that this distributionis linearly stable, that is RG indeed flows towards it. Near the fixed point, ρ =ρ0(1 + h), the transform can be linearized in h, giving h′(k) = 2h(k/

√2). The

eigenfunctions of the linearized transform are hm = km with eigenvalues 21−m/2.

94

Page 95: Physical Nature of Information - Weizmann

The three conservation laws mean that h(0) = h′(0) = h′′(0) = 0, so that onlym > 2 are admissible, which means stability, that is deviations from the fixedpoint decrease. To conclude, in the space of distributions with the same variance,the RG-flow eventually brings us to the distribution with the maximal entropy,forgetting all the information except the invariants - normalization, the mean andthe variance.

Another natural transformation is replacing a pair by their mean zi = (x2i−1 +x2i)/2. The fixed point of this distribution satisfies the equation

ρ(z) =

∫ρ(x)ρ(y)δ(z − x/2− y/2) dxdy ⇒ ρ(k) = ρ2(k/2) .

It has the solution ρ(k) = exp(−|k|) and ρ(x) = (1 + x2)−1, which is the Cauchydistribution mentioned in Section 3.1. In this case, the distribution has an infinitevariance, and RG preserves only the mean (which is zero). More generally, onecan consider zi = (x2i−1 +x2i)/2

µ and obtain the family of universal distributions,see the home exercise.

Let us now present a physical model in a discrete version - block spin transfor-mation. For Ising model, to eliminate small-scale degrees of freedom, we divide allthe spins into groups (blocks) with the side k so that there are kd spins in everyblock (d is space dimensionality). We then assign to any block a new variableσ′ which is ±1 when respectively the spins in the block are predominantly up ordown. We assume that the phenomena very near critical point can be describedequally well in terms of block spins with the energy of the same form as original,E′ = −h′

∑i σ′i − J ′/2

∑ij σ′iσ′j , but with different parameters J ′ and h′.

Let us demonstrate how it works using 1d Ising model with h = 0 and J/2T ≡K. The partition function is now Z(K) = 2(2 coshK)N−1 for the chain withopen ends and Z(K) = (2 coshK)N−1 + (2 sinhK)N−1 for the ring, yet the re-normalization group has the same action. Let us transform the partition function∑σ exp

[K∑i σiσi+1

]by the procedure (called decimation19) of eliminating de-

grees of freedom by ascribing (undemocratically) to every block of k = 3 spins thevalue of the central spin. Consider two neighboring blocks σ1, σ2, σ3 and σ4, σ5, σ6

and sum over all values of σ3, σ4 keeping σ′1 = σ2 and σ′2 = σ5 fixed. The re-spective factors in the partition function can be written as follows: exp[Kσ3σ4] =coshK + σ3σ4 sinhK, which is true for σ3σ4 = ±1. Denote x = tanhK. Thenonly the terms with even powers of σ3 and σ4 contribute the factors in the parti-tion function that involve these degrees of freedom (assuming a chain of spins, i.e

19the term initially meant putting to death every tenth soldier of a Roman army regimentthat run from a battlefield.

95

Page 96: Physical Nature of Information - Weizmann

σi+N = σi): ∑σ3,σ4=±1

exp[K(σ′1σ3 + σ3σ4 + σ4σ′2)]

= cosh3K∑

σ3,σ4=±1

(1 + xσ′1σ3)(1 + xσ4σ3)(1 + xσ′2σ4)

= 4 cosh3K(1 + x3σ′1σ′2) = e−g(K) coshK ′(1 + x′σ′1σ

′2) , (129)

g(K) = ln

(coshK ′

4 cosh3K

). (130)

The expression (129) has the form of the Boltzmann factor exp(K ′σ′1σ′2) with the

re-normalized constant K ′ = tanh−1(tanh3K) or x′ = x3 — this formula and(130) are called recursion relations. The partition function of the whole system inthe new variables can be written as∑

σ′exp

[−g(K)N/3 +K ′

∑i

σ′iσ′i+1

].

The term proportional to g(K) represents the contribution into the free energy ofthe short-scale degrees of freedom which have been averaged out. This term doesnot affect the calculation of any spin correlation function. Yet the renormalizationof the constant, K → K ′, influences the correlation functions. Let us discuss thisrenormalization. Since K ∝ 1/T then T →∞ correspond to x→ 0+ and T → 0 tox→ 1−. One is interested in the set of the parameters which does not change underthe RG, i.e. represents a fixed point of this transformation. Both x = 0 and x = 1are fixed points of the transformation x→ x3, the first one stable and the secondone unstable. Indeed, iterating the process for 0 < x < 1, we see that x approacheszero and effective temperature infinity. That means that large-scale degrees offreedom are described by the partition function where the effective temperatureis high so the system is in a paramagnetic state in agreement with the generalargument on impossibility of long-range order in one-dimensional systems withshort-range interaction. At this limit we have K,K ′ → 0 so that the contributionof the small-scale degrees of freedom is getting independent of the temperature:g(K) → − ln 4. We see that spatial re-scaling leads to the renormalization of

temperature. T=0 K=0 T=0 cT K=0

2d1d6σ5σ4σ3σ2σ1σ

Similarly, we may sum over every second spin:∑σ2k=±1

exp[K(σ2k−1σ2k + σ2kσ2k+1] = 2 cosh[K(σ2k−1 + σ2k+1)] , . (131)

96

Page 97: Physical Nature of Information - Weizmann

which gives the recursive relation e2K′ = cosh 2K.It is done similarly in a continuous case. Every stage of the renormalization

group consists now of three steps: coarse-grain, re-scale, re-normalize. The firststep is to decrease the resolution by increasing the minimum length scale fromthe microscopic scale a to ba where b > 1. This is achieved by integrating outsmall-scale fluctuations which we denote φ. The result is a renormalization of theprobability distribution, which is now expressed in terms of a coarse-grained fieldη, whose profile smoother than the original and fluctuate less. In other words, thecoarse-grained ”picture” is grainier than the original and has less contrast. Theoriginal resolution can be restored by the second step called re-scaling: decreasingall length scales by a factor b. To restore the contrast, renormalizes the field,multiplying by b-dependent factor.

How such transformation changes the statistical distribution? One starts fromthe Gibbs distribution that contains microscopic Hamiltonian: expβ[F −H(φ, η)].After averaging it over microscopic degrees of freedom at a given large-scale fieldη(r), we obtain macroscopic probability distribution expβ[F −F(η)]. As weaverage over the small-scale degrees of freedom φ we incur an information lossS(η, φ) − S(η) > 0, which was defined in (91) as the conditional entropy - theaverage entropy of the fast variables for a given configuration of slow variables:S(φ|η) = S(η, φ) − S(η). One can think of φ as an input of a channel while ηis the signal received. The entropy of the remaining variables decreases upon therenormalization group flow. That happens even when the RG flow is towardsa high-temperature fixed point, where the system is getting disordered. Thatmeans that this measure of the information loss is not very informative aboutthe flow in the space of coupling constants, it just measures the overall loss inthe degrees of freedom. We expect, however, the entropy per degree of freedom(or lattice site) to increase, reflecting irreversibility of the transformation, like itwas was was for the first example of RG with summing and re-scaling. Anot-her extensive (that is also not very useful) information characteristic is the rela-tive entropy between the distribution at any point in the RG flow and its fixedpoint value. Such relative entropy monotonously decreases upon the flow. Toget a quantity more intimately connected with RG is to subtract the informa-tion lost from the total information, which gives the familiar mutual informationI(η, φ) = S(φ) − S(φ|η) = S(η) + S(φ) − S(η, φ), compare with (92). It mea-sures how much information about the eliminated variables is still stored in therenormalized effective action. Such quantity is also of much interest for computermodeling of systems with many degrees of freedom, which is usually done for asubset of modes. The mutual information is identically zero without interaction;multi-mode mutual information I =

∑k Sk − Sηk is the relative entropy D(p|q)

between the true distribution pηk and the distribution of non-interactive modes

97

Page 98: Physical Nature of Information - Weizmann

q =∏k q(ηk). Eliminating short modes generally decreases I, but re-scaling and

renormalization may increase it. This is apparently because some of the informa-tion about eliminated degrees of freedom is stored in the renormalized values ofthe parameters of the distribution. Increase or decrease of I upon RG thus showswhether the large-scale behavior is respectively ordered or disordered.

For the decimation procedure, it is useful to define the mutual informationbetween two sub-lattices: eliminated and remaining. The positivity of the mutualinformation then implies the monotonic growth of the entropy per site h(K) =S(K,N)/N . Indeed, consider, for instance, the RG eliminating every second spin,N → N/2, and renormalizing the coupling constant by K → K ′. Subtractingthe entropy of the original lattice from the sum of the entropies of two identi-cal sub-lattices gives MI: I = 2S(N/2,K ′) − S(N,K) = N [h(K ′) − h(K)] ≥ 0.Unfortunately, this is still rather trivial in 1d, where RG moves systems towardsdisorder, so that K ′ < K and h(K ′) > h(K). In higher dimensions, on the otherhand, the next iteration of decimation cannot be performed.

In a finite system with short-range correlations, the entropy for large N isgenerally as follows:

S(N) = hN + C , I = N [h(K ′)− h(K)] + 2C ′ − C . (132)

In the critical point the extensive terms in MI cancel and I = C > 0. One canexplain positivity of C saying that a finite system appears more random than itis, since we haven’t seen all the possible correlations.

Mutual information also naturally appears in the description of the informationflows in the real space. For 1d N -chain, it makes sense to consider breakingit into two parts, M and N − M . The mutual information between two partsof the chain (or between the past and the future of a message) is as follows:I(M,N −M) = S(M) + S(N −M) − S(N). Here, the extensive parts (linear inM,N) cancel in the limit N,M → ∞. Therefore, such mutual information, alsocalled excess entropy, is equal to C from (132).

For example, the partition function of the 1d Ising chain is Z(K) = 2(2 coshK)N−1.Remind that the entropy is expressed via the partition function as follows:

S =E − FT

= T∂ lnZ

∂T+ lnZ .

So that we obtain for the chain: h = ln(2 coshK)−K tanhK and C = K tanhK−ln(coshK). Upon RG flow, these quantities monotonously change from h(K) ≈3e−2K , C ≈ ln 2 at K → ∞ to h(K) ≈ ln 2, C → 0 at K → 0. One can interpretthis, saying that C = ln q, where q is the degeneracy of the ground state, i.e. of thelargest eigenvalue of the transfer matrix Ts,s′ = exp(Kss′). Indeed, q = 2 at zero-temperature fixed point due to two ground states with opposite magnetization,

98

Page 99: Physical Nature of Information - Weizmann

while q = 1 in the fully disordered state. So this mutual information (and theexcess entropy) indeed measures how much information per one degree of freedomone needs to specify (for non-integer q, obtained mid-way of the RG flow, one canthink of it as viewing the system with finite resolution). Note that the past-futuremutual information also serves as a measure of the message complexity (that is thedifficulty of predicting the message). Without going into details, note also that Cdepends on the boundary conditions. For the ring, C = K tanhK − ln(2 coshK),but then the ring is not separated by one cut and I does not make sense.

It is important to stress the difference between the mutual information betweenpoints and regions in space and modes in Fourier space. For non-interacting modes,the MI is zero, but the MI between the points in physical space is not. Indeed,if the occupation numbers of the modes are nk, then the total entropy of therespective Gaussian distribution is

∑k ln(eπnk). The variance in every space point

is N−1∑Nk=1 nk, so that the respective entropy is ln

(eπN−1∑N

k=1 nk). The sum

of the entropies of the points is N ln(eπN−1∑N

k=1 nk), while the total entropy of

the system is the same for both representations. The difference between the sumof the point entropies and the total entropy is the N -point mutual information,

I(x1, . . . , xN ) = N ln(eπN−1∑N

k=1 nk)−∑k ln(eπnk), which is positive because

of convexity, whenever nk are not all the same. For example, equipartition inthermal equilibrium means that the energy density has a flat spectrum εk = T , ,so that the densities in different points are not correlated. Away from equilibriumthough there is generally nonzero mutual information between the energy densitiesin different points in space, even when the correlation between modes in weak.

In two dimensions, consideration is much more involved, and the reader de-serves a warning: the discussion is getting quite superficial. Breaking a singlebond now does not really separate, so in a plane H we consider a (finite) line Land break the direct interactions between the degrees of freedom on the differentsides of it. That is we make a cut in a plane and ascribe to its every point two(generally different) values on the opposite sides. The statistics of such set is nowcharacterized not by a scalar function - probability on the line - but by a matrix,similar to the density matrix in quantum statistics. As we shall describe in thenext subsection, in a quantum case, we take the whole system in a ground state,trace out (average over) all the degrees of freedom outside the line and obtain thedensity matrix ρL of the degrees of freedom on the line. For that density matrixone defines von Neumann entropy S(L) = −TrρL log ρL

For long lines in short-correlated systems, that information can be shown todepend only on the distance r between the end points (and not on the form of aline connecting them, that is information flows like an incompressible fluid). Mo-reover, this dependence is not linear but logarithmic at criticality (when we have

99

Page 100: Physical Nature of Information - Weizmann

fluctuations of all scales and the correlation radius is infinite). Then one definesthe function c(r) = rdSL(r)/dr (to cancel non-universal terms depending on themicroscopic details) and shows that this is a monotonic zero degree function usingpositivity of the mutual information (sub-additivity of the entropy) between li-nes with r and r + dr (Zamolodchikov 1986, Casini and Huerta 2006). The samefunction changes monotonically under RG flow and in a fixed point takes a finitevalue equal to the so-called zero charge of the respective conformal field theory,which is again some measure of relevant degrees of freedom which respond toboundary perturbations. Moment reflection is enough to understand how difficultit must be to introduce proper intensive measure of information flow in dimen-sions higher than two, and yet recently, such function was also found for d = 4(Komargodsky and Schwimmer 2011) and d = 3 (Klebanov).

In looking for fundamental characteristics of order in fluctuating systems inhigher dimensions, one can go even deeper20. For instance, one can consider forquantum system in 2+1 dimensions the relative entanglement of three finite planarregions, A,B,C, all having common boundaries. As a quantum analog of theinteraction information (87), one can introduce so-called topological entanglemententropy SA+SB+SC+SABC−SAB−SBC−SAC . For some classes of systems, onecan show that in the combination, the terms depending on the boundary lengthescancel out and what remains (if any) can be thus independent of the deformationsof the boundaries, that is characterizing the topological order, if it exists in thesystem (Kitaev, Preskill 2006).

5.3 Quantum information

this chapter should have been called”quantum information theory for the impatient”

J PreskillSince our world is fundamentally quantum mechanical, it is interesting what

that reveals about the nature of information. Communication (measurement) dra-matically differ in a quantum world: i) it irreversibly changes the system and thischange cannot be made arbitrarily small, and ii) the results are truly random (notbecause we did not bother to learn more on the system). Apart from that fun-damental difference, interest in quantum information is also pragmatic. Classicalsystems, including computers, are limited by locality (operations have only localeffects) and by the classical fact that systems can be in only one state at the time.However, a quantum system can be in a superposition of many different states

20Mention in passing that in dimensions d > 4, the block-spin renormalization of theIsing-class models leads to asymptotic Gaussian distribution ln ρ(η) ∝ −|∇η|2

100

Page 101: Physical Nature of Information - Weizmann

at the same time, which means that quantum evolution exhibit interference ef-fects and proceeds in the space of factorially more dimensions than the respectiveclassical system. This is a source of the parallelism of quantum computations.Moreover, spatially separated quantum systems may be entangled with each otherand operations may have non-local effects because of this. Those two basic factsmotivate an interest in quantum computation and then in quantum informationtheory. Non-surprisingly, it is also based on the notion of entropy, which is si-milar to classical entropy yet differs in some important ways. Uncertainly andprobability exist already in quantum mechanics where our knowledge of the stateof the system is complete. On top of that we shall consider quantum statisticsdue to incomplete knowledge, which is caused by considering subsystems. In thissubsection I’ll give a mini-introduction to the subject, focusing on entropy21. Thepresentation mostly follows Witten lectures with an occasional borrowing from thePreskill’s book at

https://arxiv.org/pdf/1805.11965.pdf

http://www.theory.caltech.edu/~preskill/ph219/chap10_6A.pdf

Quantum mechanics describes a system by presenting its stationary states andpossible jumps between those states. The fundamental statement is that any sy-stem can be in a pure state ψi (like, for instance, eigenstate of a Hamiltonian,which has fixed energy) or in a superposition of states: ψ =

∑i√piψi. This is

the total breakdown from classical physics, where those states (say, with differentenergies) are mutually exclusive. There are two things we can do with a quantumstate: let it evolve (unitarily) without touching or measure it. Measurement isclassical, it produces one and only pure state from the initial superposition; imme-diately repeated measurements will produce the same outcome. However repeatedmeasurement of the identically prepared initial superposition find different states,the state i appears with probability pi. A property that can be measured is calledan observable and is described a self-adjoint operator.

There is an uncertainty already in a pure state of an isolated quantum system.Quantum uncertainty principle states that if the operators of two observables arenon-commuting, then the values of the observables cannot be fixed at the sametime. In particular, momentum and coordinate are such pair, so that we cannotdescribe quantum states as points in the phase space. Indeed, quantum entang-lement is ultimately related to the fact that one cannot localize quantum statesin a finite region — if coordinates are zero somewhere, then the momenta arenot. The uncertainty restriction was formulated initially by Heisenberg in terms

21Recall that the entropy consideration by Planck is what started quantum physics inthe first place.

101

Page 102: Physical Nature of Information - Weizmann

of variances: for a free quantum particle, the variances of the coordinate and themomentum along the same direction satisfy the inequality

√σpσq ≥ h/2. However,

variances are sufficient characteristics of uncertainty only for Gaussian distributi-ons, generally entropy plays that role. Taking log of the Heisenberg relation, weobtain log(2σp/h) + log(2σq/h) = S(p) + S(q) ≥ 0, recasting it as requirementson the entropies of the Gaussian probability distributions of the momentum andthe coordinate of a free particle. In d dimensions, different components commute,so that

√σpσq ≥ dh/2 and S(p) + S(q) ≥ log d. When the respective probability

distributions of non-commuting variables are not Gaussian, formulation in termsof variances does not make sense; yet the entropic uncertainty relation remainsuniversally valid and is thus fundamental (Deutsch 1982).

More formally, if we measure a quantum state ρ by projecting onto the ortho-normal basis |x〉, the outcomes define a classical probability distribution p(x) =〈x|ρ|x〉, which is a probability vector whose entries are the diagonal elements of ρin the x-basis. The Shannon entropy S(X) quantifies how uncertain we are aboutthe outcome before we perform the measurement. If |z〉 is another orthonormalbasis, there is a corresponding classical probability distribution of outcomes whenwe measure the same state ρ in the z-basis. If the operators projecting on x, z donot commute, the two bases are incompatible, so that there is a tradeoff betweenour uncertainty about X and about Z, captured by the inequality

S(X) + S(Z) ≥ log(1/c) , c = maxx,z|〈x|z〉|2 .

We say that two different bases |x〉, |z〉 for a d-dimensional Hilbert spaceare mutually unbiased if for all |〈xi|zk〉|2 = 1/d. That means that if we measureany x-basis state in the z-basis, all d outcomes are equally probable and give thesame contribution into the total probability:

∑k |〈xi|zk〉|2 =

∑i |〈xi|zk〉|2 = 1. For

example, if a particle is in a definite point, all momentum directions are equallyprobable. For measurements in two mutually unbiased bases performed on a purestate, the entropic uncertainty relation becomes

S(X) + S(Z) ≥ log d .

Clearly this inequality is tight, as it is saturated by x-basis (or z-basis) states, forwhich S(X) = 0 and S(Z) = log d. In particular, in one dimension log d = 0.The uncertainty increases if ρ is not a pure state but a mixture, characterized byadditional S(ρ):

S(X) + S(Z) ≥ log(1/c) + S(ρ) .

In reality we always study subsystems, which requires us to pass from quantummechanics to quantum statistics and brings the fundamental notion of the density

102

Page 103: Physical Nature of Information - Weizmann

matrix. Consider a composite system AB, which is in a pure state ψAB. Denoteby x the coordinates on A and by y on B. The expectation value of any O(x) can bewritten as O =

∫dxdyψ∗AB(x, y)O(x)ψAB(x, y). For non-interacting sub-systems,

one has ψAB(x, y) = ψA(x)ψB(y) and O =∫dxψ∗A(x)O(x)ψA(x), so that one can

forget about B. But generally, dependencies on x and y are not factorized, so oneought to characterize A by the so-called density matrix

ρ(x, x′) =

∫dyψ∗AB(x′, y)ψAB(x, y),

so that O =∫dx[O(x)ρ(x, x′)]x′=x, where O(x) acts only on x and then we put

x′ = x.More formally, if the pure state ψAB is a (tensor) product22 of pure states

of A and B, ψAB = ψA⊗ψB, then any operator OA acting only in A has the

expectation value

〈ψAB|OA⊗

1B|ψAB〉 = 〈ψA|OA|ψA〉〈ψB|1B|ψB〉 = 〈ψA|OA|ψA〉 .However, a general pure state ψAB is not a (tensor) product of pure states, it iswhat is called entangled: some generic M×N matrix where N,M dimensionalitiesof A,B respectively: ψ1

Aψ1B . . . ψ1

AψMB

. . .ψNAψ

1B . . . ψNAψ

MB

(133)

We can make it N×N diagonal with N positive eigenvalues√pi and extra M−N

rows of zeroes by applying unitary transformation in A and B, that is makingψAB → UψABV with M ×M unitary V-matrix and N × N U-matrix. That iscalled Schmidt decomposition by orthonormal vectors ψiA, ψ

iB, which allows us to

present the state of AB as a sum of the products:

ψAB =∑i

√piψ

iA

⊗ψiB , (134)

If there is more than one term in this sum, we call subsystems A and B entangled.There is no factorization of the dependencies in such a state. We can alwaysmake ψAB a unit vector, so that

∑i pi = 1 and these numbers can be treated as

probabilities (to be in the state i). Now the operator acting only on A has theexpectation value

〈ψAB|OA⊗

1B|ψAB〉 =∑i,j

√pipj〈ψiA|OA|ψ

jA〉〈ψ

iB|1B|ψ

jB〉

=∑i,j

√pipj〈ψiA|OA|ψ

jA〉δij =

∑i

pi〈ψiA|OA|ψiA〉 = TrAρAOA ,

22You multiply every component of one N-vector by every component of another M-vector and you obtain a MN-vector called tensor product

103

Page 104: Physical Nature of Information - Weizmann

where the density matrix in such notations is written as follows:

ρA =∑i

pi|ψiA〉〈ψiA| . (135)

It is all we need to describe A. The matrix is hermitian, it has all non-negativeeigenvalues and a unit trace. Every matrix with those properties can be ”purified”that is presented (non-uniquely) as a density matrix of some pure state ψAB in theextended system AB. Possibility of purifications is quantum mechanical with noclassical analog: the classical analog of a density matrix is a probability distributionwhich cannot be purified.

Statistical density matrix describes the mixed state (or, in other words, anensemble of states), which must be distinguished from quantum-mechanical super-position. The superposition is in both states simultaneously; the ensemble is inperhaps one or perhaps the other, characterized by probabilities - that uncertaintyappears because we do not have any information of the state of the B-subsystem.

Let us illustrate this in the simplest case of a qubit — a quantum systemhaving only two states: |0〉 and |1〉. The most general state of a qubit A is asuperposition of two states, ψA = a |0〉+ b |1〉, then any observable is as follows:

〈ψA|OA|ψA〉 = |a|2〈0|OA|0〉+ |b|2〈1|OA|1〉+ (a∗b+ ab∗)〈0|OA|1〉 . (136)

Normalization requires |a|2 + |b|2 = 1 and if the overall phase does not matter,then a qubit is characterized by a complex number. The qubit represents the unitof quantum information, while the bit represents the unit of classical information.Here we see that quantum systems operate with much more information - oneneeds many bits to record a complex number with a reasonable precision, and thedifference grows exponentially when we compare the states of N classical bits withthe possible states of N qubits. Moreover, qubit is not a classical bit because itcan be in a superposition, nor can it be considered a random ensemble of classicalbits with the probability |a|2 in the state |0〉, because the phase difference of thecomplex numbers a, b matter.

Consider now a pure quantum state of the two-qubit system A,B of the form

ψAB = a |00〉+ b |11〉 . (137)

A and B are correlated in that state, one can predict one by knowing another.Now any operator acting on A gives

〈ψAB|OA⊗

1B|ψAB〉 = (a∗〈00|+ b∗〈11|)OA⊗

1B|(a|00〉+ b|11〉)

= |a|2〈0|OA|0〉+ |b|2〈1|OA|1〉 , (138)

ρA = |a|2 |0〉 〈0|+ |b|2 |1〉 〈1| =[|a|2 00 |b|2

]. (139)

104

Page 105: Physical Nature of Information - Weizmann

We can interpret this as saying that the system A is in a mixed state, that iswith probability |a|2 in the quantum state |0〉, and with probability |b|2 it is in thestate |1〉. Due to the orthogonality of B-states, the same result (138) is obtainedif 〈0|OA|1〉 6= 0, in distinction from (136). Being in a superposition is not thesame as being in mixed state, where the relative phases of the states |0〉 , |1〉 areexperimentally inaccessible.

General 2 × 2 density matrix is characterized by three real numbers, P =(P1, P2, P3):

ρ(P) =1

2

[1− P1 P2 + ıP3

P2 − ıP3 1 + P1

]. (140)

Non-negativity of the eigenvalues requires P 2 = P 21 + P 2

2 + P 23 ≤ 1, that is the

space of such density matrices is a ball. The boundary P 2 = 1 corresponds tozero determinant, that is to eigenvalues 0 and 1 — that means that ρ is a one-dimensional projector and the subsystem is in a pure state.

We call the system entangled if its density matrix has more than one nonzeroeigenvalue, so that there is more than one term in the sum (135).

The von Neumann entropy of a density matrix ρA is defined by the formulaanalogous to the Gibbs-Shannon entropy of a probability distribution:

S(ρA) = SA = −Tr ρA log ρA . (141)

It is manifestly invariant under a unitary transformation ρA → UρAU−1, which

is an analog of the Liouville theorem on the conservation of distribution by Ha-miltonian evolution. Just like the classical entropy, it is non-negative, equals tozero only for a pure state and reaches its maximum log d for equipartition (whenall d non-zero eigenvalues are equal). It also satisfies concavity (66). What doesnot have a classical analog is that the purifying system B has the same entropyas A (since the same pi appears in its density matrix). Moreover, von Neumannentropy of a part SA can be larger than that of the whole system SAB. WhenAB is pure, SAB = 0, but SA could be nonzero (Landau 1927). Information canbe encoded in the correlations among the parts, yet be invisible when we lookat one part of a quantum system, but not a classical one. That purely quantumcorrelation between different parts is called entanglement, and the von Neumannentropy of a subsystem of pure state is called entanglement entropy. Classically,we measured the nonlocality of information encoding by the mutual informationI(A,B) = S(A) + S(B) − S(A,B), which never exceeds the sum of two entro-pies. Nonlocality of information encoding is raised to the whole new level in thequantum world.

The Von Neumann entropy of a density matrix is the Shannon entropy S(p) =−∑i pi log pi of its vector of eigenvalues, which is the probability distribution pi

105

Page 106: Physical Nature of Information - Weizmann

of its orthonormal eigenstates. In particular, S(ρA) = −|a|2 log2 |a|2− |b|2 log2 |b|2for (137,138). The maximum S(ρA) = 1 is reached when a = b = 1/

√2, which

is called a state of maximal entanglement. Pure state of a qubit can store onebit. The four maximally entangled states of the qubit pair, (|00〉 ± |11〉)/

√2 and

(|01〉 ± |10〉)/√

2 can store two bits, yet when we trace out B (or A) we wipe outthis information: any measurement on A or B cannot tell us anything about thestate of the pair, since both outcomes are equally probable. On the contrary, wheneither b→ 0 or a→ 1, the entropy S(ρA) goes to zero and measurements (of eitherA or B) give us definite information on the state of the system.

If a pure state AB was built from non-orthogonal states, taken with proba-bilities qi, then the density matrix is non-diagonal. There is then the differencebetween the Shannon entropy of the mixture and the von Neumann entropy of thematrix, Sqi−S(ρA). It is non-negative and quantifies how much distinguishabi-lity is lost when we mix nonorthogonal pure states. Measuring ρA we can receiveS(ρA) bits, which is less than Sqi bits that was encoded mixing the states withprobabilities qi.

On the other hand, we may measure by projecting ρA on the orthogonalset |ψiA〉 which is different from the basis of eigenvectors |φkA〉, where ρA =∑k pk|φkA〉〈φkA| is diagonal. In this case, the outcome i happens with the pro-

bability q′(i) = 〈ψiA|ρA|ψiA〉 =∑k pkDij , where Dik = |〈ψiA|φkA〉|2 is so-called

double stochastic matrix, that is∑iDik =

∑kDik = 1. The Shannon en-

tropy of that probability distribution is larger than the von Neumann entropy,S(q′) = S(p) +

∑ikDik logDik ≥ S(p) = S(ρA), that is such measurements are

less predictable.

Similar to the classical conditional entropy (71) and the mutual information(72) we can define

S(A|B) = SAB − SB , I(A,B) = SA + SB − SAB . (142)

Quantum I is non-negative as well as classical, but generally is different. Forexample, when AB is in an entangled pure state then SAB = 0, A and B togetherare perfectly correlated, but separately each one is in a mixed state, so that SA =SB > 0. Classically, the mutual information of perfectly correlated quantities isequal to each of their entropies, but quantum mutual information is their sumthat is twice more: I(A,B) = SA + SB − SAB = 2SA. Quantum correlations arestronger than classical.

Quantum S(A|B) is even more bizarre; this is not an entropy conditional onsomething known, moreover it is not zero for correlated quantities but negative!Indeed, for pure AB, one has S(A|B) = −SB < 0. Classical conditional entropymeasures how many classical bits we need to add to B to fully determine A.

106

Page 107: Physical Nature of Information - Weizmann

Similarly, we would expect quantum conditional entropy to measure how manyqubits Alice needs to send to Bob to reveal herself. But what does it mean whenS(A|B) is negative?

That negativity allows for a quantum trick of teleportation of a quantum state.Imagine that Alice has in her possession a qubit A0. Alice would like to helpBob create in his lab a qubit in a state identical to A0. However, she wants tocommunicate by sending a classical message (sat, over the telephone). If Aliceknows the state of her qubit, there is no problem: she tells Bob the state of herqubit and he creates one like it in his lab. If, however, Alice does not know thestate of her qubit, all she can do is make a measurement, which will give someinformation about the prior state of qubit A0. She can tell Bob what she learns,but the measurement will destroy the remaining information about A0 and it willnever be possible for Bob to recreate it.

Suppose, however, that Alice and Bob have previously shared a qubit pairA1, B1 in a known entangled state, for example,

ψA1B1 =1√2

(|00〉+ |11〉)A1B1 . (143)

Bob then took B1 with him, leaving A1 with Alice. In this case, Alice can solvethe problem by making a joint measurement of her system A0A1 in a basis, that ischosen so that no matter what the answer is, Alice learns nothing about the priorstate of A0. In that case, she also loses no information about A0. But after gettingher measurement outcome, she knows the full state of the system and she can tellBob what to do to recreate A0. To see how this works, let us describe a specificmeasurement that Alice can make on A0A1 that will shed no light on the state ofA0. The measurement must be a projection on a state where the probability ofA0 to be in the state |0〉 is exactly equal to the probability to be in the state |1〉.The following four states of A0A1 satisfy that property:

1√2

(|00〉 ± |11〉)A0A1 ,1√2

(|01〉 ± |10〉)A0A1 . (144)

The states are chosen to be entangled, that is having A0, A1 correlated. We don’tuse the state with |00〉 ± |10〉, which has equal probability of zero and one for A,but no correlation between the values of A0, A1.

Denote the unknown initial state of the qubit A0 as α |0〉 + β |1〉, then theinitial state of A0A1B1 is

1√2

(α |000〉+ α |011〉+ β |100〉+ β |111〉)A0A1B1 . (145)

107

Page 108: Physical Nature of Information - Weizmann

Let’s say that Alice’s measurement, that is the projection on the states (144),reveals that A0A1 is in the state

1√2

(|00〉 − |11〉)A0A1 . (146)

That means that only the first and the last terms in (145) contribute (with equalweights but opposite signs). After that measurement, B1 will be in the state(α |0〉−β |1〉)B1 , whatever the (unknown) values of α, β. Appreciate the weirdnessof the fact that B1 was uncorrelated with A0 initially, but instantaneously acquiredcorrelation after Alice performed her measurement thousand miles away. Knowingthe state of B1, Alice can send two bits of classical information, telling Bob thathe can recreate the initial state α |0〉+ β |1〉 of A0 by multiplying the vector of his

qubit B1 by the matrix

[1 00 −1

], that switches the sign of the second vector of

the basis. The beauty of it is that Alice learnt and communicated not what wasthe state A0, but how to recreate it.

To understand the role of the quantum conditional entropy (142), we symme-trize and purify our problem. Notice that A1 and B1 are maximally entangled(come with the same weights), so that SB = log2 2 = 1. On the other hand,A1B1 is in a pure state so SA1B1 = 0. Let us now add another system R which ismaximally entangled with A0 in a pure state, say

ψA0R =1√2

(|00〉+ |11〉)A0R . (147)

Neither Alice nor Bob have access to R. From this viewpoint, the combined systemRAB = RA0A1B1 starts in a pure state which is a direct product ψRA0

⊗ψA1B1 .

Since A0 is maximally entangled with R then also SA0 = log2 2 = 1 and the sameis the entropy of the AB system S(A0A1B1) = S(A0) = 1 since A1B1 is a purestate. Therefore, S(A|B) = S(A0A1|B) = S(A0A1B1)−S(B1) = 0. One can showthat teleportation only possible when S(A|B) is non-positive.

Recall that classically S(A|B) measures how many additional bits of informa-tion Alice has to send to Bob after he has already received B, so that he will havefull knowledge of A. Quantum analog of this involves qubits rather than classi-cal bits. Suppose that S(A|B) > 0 and Alice nevertheless wants Bob to recreateher states. She can simply send her states. Alternative is to do teleportation,which requires sharing with Bob an entangled pair for every qubit of her state.Either way, Alice must be capable of quantum communication, that is of sending aquantum system while maintaining its quantum state. For teleportation, she firstcreates some maximally entangled qubit pairs and sends half of each pair to Bob.Each time she sends Bob half of a pair, SAB is unchanged but SB goes up by 1,

108

Page 109: Physical Nature of Information - Weizmann

so S(A|B) = SAB − SB goes down by 1. So S(A|B), if positive, is the numberof such qubits that Alice must send to Bob to make S(A|B) non-positive and somake teleportation possible without any further quantum communication. Nega-tive quantum conditional entropy measures the number of possible future qubitteleportations.

In quantum communications, we pose the same natural question: i) How muchcan a message be compressed, that is what is the maximum information one cantransmit per quantum state - is it given by von Neumann entropy? Now the lettersof our message are quantum states picked with their respective probabilities pk,that is each letter is described by the density matrix and the message is a tensorproduct of N matrices. If the states are mutually orthogonal then the densitymatrix is diagonal and it is essentially the classical case, that is the answer isgiven by the Shannon entropy S(p) = −

∑k pk log pk, which is the same as von

Neumann entropy in this case. The central issue in quantum information theoryis that nonorthogonal quantum states cannot be perfectly distinguished, a featurewith no classical analog. For non-orthogonal states sending classical informationabout the probabilities of different states is not optimal since non-orthogonal statesare not completely distinguished. For pure states the answer is given by S(ρ) — vonNeumann entropy is the number of qubits of quantum information carried per letterof a long message, and it is the best possible rate of quantum information transfer(the only difference from the classical case is that instead of typical sequences weconsider typical subspaces). It is easy to see that S(ρ) wont be the answer formixed states. To give a trivial example, suppose that a particular mixed state ρ0

with S(ρ0) > 0 is chosen with probability p0 = 1. Then the message ρ0⊗ρ0⊗. . .

carries no information. If, however, our alphabet is made of mixed yet mutuallyorthogonal states, then the states are distinguishable and the problem is classical,since we can just send the probabilities of the states, so the maximal rate is theShannon entropy. However, it is now different from von Neumann entropy becauseevery mixed state ρk has a nonzero entropy:

S(ρ)=−∑k

Tr(pkρk) log(pkρk)

=−∑k

(pk log pk + pkTr ρk log ρk)=S(p) +∑k

pkS(ρk) .

Now Shannon entropy is less than von Neumann entropy:

S(p) = S(ρ)−∑k

pkS(ρk) = S

(∑k

pkρk

)−∑k

pkS(ρk) ≡ χ(ρk, pk) . (148)

It is believed that this quantity χ (called in quantum communications Holevo infor-mation) defines the limiting compression rate in all cases including non-orthogonal

109

Page 110: Physical Nature of Information - Weizmann

mixed states when it does not coincide with S(p). The reason is that χ is mono-tonic (i.e. decreases when we take partial traces), but S(ρ) is not - indeed one canincrease von Neumann entropy by going from a pure to a mixed state. It followsfrom concavity that χ is always non-negative. We see that it depends on the pro-babilities pk that is on the way we prepare the states. Of course, (148) is a kindof mutual information, it tells us how much, on the average, the von Neumannentropy of an ensemble is reduced when we know which preparation was chosen,exactly like classical mutual information I(A,B) = S(A) − S(A|B) tells us howmuch the Shannon entropy of A is reduced once we get the value of B. So we seethat classical Shannon information is a mutual von Neumann information.

Early idea of entanglement was conjured in 17 century precisely for communi-cations: it was claimed that if two magnetic needles were magnetized at the sameplace and time, they would stay ”in sympathy” forever at however large distancesand the motion of one is reflected on another. One con man tried to sell this ideato Galileo, but the latter naturally wanted to see an experiment first.

5.4 Thermalization, Area law and Bekenstein bound

Coming to equilibrium. Are thermalization and entropy growth possible fora quantum system which as a whole remains in a pure quantum state? Yes theyare! Thermalization and entropy growth take place for any subsystem of a largesystem if the dynamics is ergodic. Then the system as a whole acts as a ther-mal reservoir for its subsystems, provided those are small enough. As a quantumsubsystem approaches equilibrium, the von Neumann entropy reaches its maxi-mum. Information, which was initially encoded locally in an out-of-equilibriumstate, becomes encoded more and more nonlocally as the system evolves, eventu-ally becoming invisible to an observer confined to the subsystem.

Is there any quantum analog of chaos which underlies thermalization the sameway that dynamical chaos underlies mixing and thermalization in the classicalstatistics, as described in Sect. 2.3? Writing the classical formula of exponen-tial separation, δx(t) = δx(0)eλt as ∂x(t)/∂x(0) = eλt and replacing quantum-mechanically the space derivative by the momentum operator, one naturally comesto consider the commutator of x(t) and p(0). Indeed, ∂x(t)/∂x(0) = x(t), p(0) →h−1[x(t), p(0)]. The commutator measures the effect of having at t = 0 the valueof p on the later measurement of x(t). The average value of this commutatorover Gibbs distribution with a finite temperature T is zero. Averaging the square,C(t) = 〈[x(t)p(0)]2)〉, brings the concept of a so-called out-of-time-order correlationfunction like 〈x(t)p(0)x(t)p(0)〉. Such quantities are found to grow exponentiallyin time in some quantum systems (complicated enough to allow chaos and simpleenough to allow for analytic solvability): C(t) = h2eλt, where uncertainty relation

110

Page 111: Physical Nature of Information - Weizmann

gives the starting value at t = 0. Respective Lyapunov exponent dimensionallymust be temperature divided by h and indeed λ = 2πT/h was shown to be auniversal upper limit (reached by black holes, for instance).

Area law. In a thermodynamic limit, the total entropy is extensive, that isproportional to the system volume or total number of degrees of freedom. Wehave already noticed that in systems with local interaction extensive (volume)terms cancel in the entanglement entropy and mutual information, which are thussub-extensive. This is a manifestation of the so-called area law. If we define,according to (135,141), the von Neumann entropy SA of the region A or considerits classical analog, the mutual information between A and its complement H/A,they both are proportional to the boundary area rather than to the volume of A.

In particular, the quantum entanglement entropy (between interior and ex-terior) is thought to be responsible for the entropy of black holes, which is pro-portional to the boundary area and not to the volume. That area law can beargued as follows. Black hole is an object, whose size is smaller than its horizonrh = 2GM/c2, where M is the mass and G is the gravitational constant. Onecannot escape from within rh, since the speed needed for that exceeds c. Theenergy of the hole is thus EBH = Mc2 = c4rh/2G. The temperature of the holeis determined by its radiation, which is a purely quantum phenomenon: one partof a particle-antiparticle pair appearing from the vacuum fluctuations can be ab-sorbed by the hole, then another part will be registered as radiation (this is howthe entanglement appears). The typical wavelength of such radiation is rh and itsenergy/temperature is then T = hc/4πrh. Now we can obtain entropy integratingthe equation of state T = dE/dS:

T =hc

4πrh=dEBHdSBH

=c4

2G

drhdSBH

⇒ SBH =πr2

hc3

Gh=πr2

h

l2p=

4πGM2

hc. (149)

The Planck length lp ' 10−17 cm is the only combination with that dimensiona-lity of the three fundamental physical constants, c, h, G; it is the scale wherequantization of gravity is expected to be important23.

The area law behavior of the entanglement entropy in microscopic theoriescould be related to the holographic principle — the conjecture that the informationcontained in a volume of space can be completely encoded by the degrees of freedomwhich live on the boundary of that region.

23One-parameter theories: G-theory, 17th century; c-theory, 19-20 centuries; h-theory,early 20th century. Two-parameter theories: c,G - general relativity, c, h -quantum elec-trodynamics, 20th century. Hopefully, c,G, h-theory will appear in the 21st century.

111

Page 112: Physical Nature of Information - Weizmann

Bekenstein bound. One defines the information capacity (not a channel ca-pacity!) as the maximal amount of information that can be encoded in a systemby exploiting all of its degrees of freedom. Conversely, it corresponds to the max-imal amount of information required to perfectly describe a given physical systemdown to the quantum level. Is there a universal limit on how large the entropy ofa physical system can be? The answer is given by the so-called Bekenstein bound(and its generalizations): If R is the radius of a sphere that can enclose a givensystem, while E is its total energy including any rest masses, then its entropy isbounded by

S ≤ 2πRE

hc. (150)

(Bekenstein 1981, 2004; Casini 2008). That bound was argued by exploiting theonly known way to eliminate entropy from the observable world — to drop it intoa black hole. If we drop a body of the energy E and entropy S into a black hole oflarge mass M E/c2, then the black hole’s mass will grow by E/c2. According to(149), the entropy of the hole will then grow by 8πGME/hc3 plus a negligible termof order E2. Meanwhile the entropy S has gone forever out of this world24. Thesecond law then requires that S < 8πGME/hc3 = 4πrhE/hc. The smallest blackhole, that can absorb the body, must have the horizon exceeding the body size,which gives the estimate for the bound (150) (up to a factor 1/2 which requiresactual consideration of the process of adiabatic lowering of the body into the hole).

We assumed that the body is itself not a black hole, that is its size exceeds itshorizon, R > rh(E) = 2GE/c4, so that the entropy restriction can be formulatedsolely in terms of the radius:

S ≤ πR2c3

Gh. (151)

Comparing (151) and (149), we conclude that a system must be a black hole torealize the capacity limit.

While the gravitational constant G does not enter (150), the appearance ofgravity in the argument and in (151) deserves reflection. Via black holes, gravityprovides the boundaries of the observable world, which is a source of the bound.I cannot resist mentioning in this connection the hypothesis that the gravity isitself an entropic force caused by changes in the information associated with thepositions of material bodies (Verlinde 2011). Entropic force is the derivative of the

24”When I put my hot tea next to a cool tea they come to a common temperature. Ihave committed a crime. I have increased the entropy of the universe. Jakob Bekensteincame by and I said to him if a black hole comes by and I drop the cups behind it I canconceal my crime. He was upset by this and came back in a few months to tell me thatblack holes have entropy.” John Wheeler.

112

Page 113: Physical Nature of Information - Weizmann

free energy due to its entropy part; such is the elastic stress of a stretched polymerwhich has much more available configurations when it coils.

6 Stochastic processes

In this Section we further exploit formal similarities between quantum fluctuati-ons and thermal noise and present modern generalizations of the second law andfluctuation-dissipation relations. This is best done using the fundamental processof a random walk in different environments. It is interesting both for fundamen-tals of science and for numerous modern applications related to fluctuations innano-particles, macro-molecules, stock market prices etc.

6.1 Random walk and diffusion

Consider a particle that can hop randomly to a neighboring cite of d-dimensionalcubic lattice, starting from the origin at t = 0. We denote a the lattice spacing, τthe time between hops and ei the orthogonal lattice vectors that satisfy ei · ej =a2δij . The probability to be in a given cite x satisfies the equation

P (x, t+ τ) =1

2d

d∑i=1

[P (x + ei, t) + P (x− ei, t)] . (152)

The first (painful) way to solve this equation is to use a generating function andaverage exponents, as customarily in statistical physics. This is done using theFourier transform, P (x) = (a/2π)d

∫eikxP (k) ddk, which gives

P (k, t+ τ) =1

d

d∑i=1

cos aki P (k, t) . (153)

The initial condition for (152) is P (x, 0) = δ(x), which gives P (k, 0) = 1 and

P (k, t) =(d−1∑d

i=1 cos aki)t/τ

. That gives the solution in space as an integral

P (x, t) = (a/2π)d∫eikx

(1

d

d∑i=1

cos aki

)t/τddk . (154)

We are naturally interested in the continuous limit a → 0, τ → 0. If we takeτ → 0 first, the integral tends to zero. If we take a→ 0 first, the answer remainsdelta-function. A non-trivial evolution appears when the lattice constant and the

113

Page 114: Physical Nature of Information - Weizmann

jump time go to zero simultaneously. Consider the cosine expansion,(1

d

d∑i=1

cos aki

)t/τ=(1− a2k2/2d+ . . .

)t/τ,

where k2 =∑di=1 k

2i . The finite answer exp(−κtk2) appears only if one takes the

limit keeping constant the ratio κ = a2/2dτ . In this limit, the space density of theprobability stays finite and is given by the integral:

ρ(x, t) = P (x, t)a−d ≈ (2π)−d∫eikx−tκk

2ddk = (4πκt)−d/2exp

(− x2

4κt

). (155)

The second (painless) way to get this answer is to pass to the continuum limitalready in the equation (152):

P (x, t+ τ)− P (x, t) =1

2d

d∑i=1

[P (x + ei, t) + P (x− ei, t)− 2P (x, t)] . (156)

This is a finite difference approximation to the diffusion equation

(∂t − κ∆)P (x, t) = 0 . (157)

Of course, ρ satisfies the same equation, and (155) is its solution. Note that(155,157) are isotropic and translation invariant while the discrete version re-spected only cubic symmetries. Also, the diffusion equation conserves the to-tal probability,

∫ρ(x, t) dx, because it has the form of a continuity equation,

∂tρ(x, t) = −div j with the current j = −κ∇ρ.Another way to describe a random walk is to treat ei as a random variable

with 〈ei〉 = 0 and 〈eiej〉 = a2δij , so that x =∑t/τi=1 ei. The probability of the sum

is (155), that is the product of Gaussian distributions of the components, with thevariance growing linearly with t.

A path of a random walker behaves rather like a surface than a line. Two-dimensionality of the random walk is a reflection of the square-root diffusion law:〈x〉 ∝

√t. Indeed, we defined in Section 2.5 the box-counting dimension (54)

looking how the number of boxes N(a) needed to cover the a geometric objectgrow as the box size a decreases. For a line, N ∝ 1/a, generally N ∝ a−d. As wediscussed above, diffusion requires the time step to shrink with the lattice spacingaccording to τ ∝ a2. The number of elements is the number of steps and growsfor a given t as N(a) = t/τ ∝ a−2, so that d = 2. One can also obtain dimensionof a set as the relation between its size x and the number N ∝ xd of standard-sizeelements needed to cover it. For a random walk, the number of elements is of order

114

Page 115: Physical Nature of Information - Weizmann

of the number of steps, N ∝ t ∝ x2. Surfaces generally intersect along curves in3d, they meet at isolated points in 4d and do not meet at d > 4. That is reflectedin special properties of critical phenomena in 2d (where random walker fills thesurface) and 4d (where random walkers do not meet and hence do not interact). Isthe mean time spent on a given site,

∑∞t=0 P (x, t), finite or infinite? It follows from

(155) that the answer depends on the space dimensionality:∫∞ t−d/2dt diverges

for d ≤ 2. In other words, the walker in 1d and 2d returns to any point infinitenumber of times.

The properties of random walks can be expressed alternatively in terms of sumsover different paths. Let us write the transition probability indicating explicitlythe origin: ρ(x, t; 0, 0). It is a conditional probability of x under the condition thatthe time −t before the walker was at 0. Then we can write the convolution identitywhich simply states that the walker was certainly somewhere at an intermediatetime t1:

ρ(x, t; 0, 0) =

∫ρ(x, t;x1, t1)ρ(x1, t1; 0, 0) dx1 . (158)

We now divide the time interval t into an arbitrary large number of intervals andusing (155) we write

ρ(x, t; 0, 0) =

∫Πni=0

dxi+1

[4πκ(ti+1 − ti)]d/2exp

[− (xi+1 − xi)

2

4κ(ti+1 − ti)

]

→∫Dx(t′) exp

[− 1

∫ t

0dt′x2(t′)

]. (159)

The last expression is an integral over paths that start at zero and end up at xat t. Notation Dx(t′) implies integration over the positions at intermediate timesnormalized by square roots of the time differences. The exponential it gives theweight of every trajectory.

Looking at the transition probability (159), one can also see the analogy bet-ween the statistical physics of a random walk and quantum mechanics. Accordingto Feynman, for a quantum non-relativistic particle with a mass M , the transitionamplitude T (x, t; 0, 0) from zero to x during t is given by the sum over all possiblepaths connecting the points. Every path is weighted by the factor exp(iS/h) whereS is the classical action:

T (x, t; 0, 0) =

∫Dx(t′) exp

[i

h

∫ t

0dt′Mx2

2

]. (160)

Comparing with (159), we see that the transition probability of a random walkis given by the transition amplitude of a free quantum particle during an imagi-nary time. In quantum theory, one averages over quantum rather than thermalfluctuations, yet the formal structure of the theory is similar.

115

Page 116: Physical Nature of Information - Weizmann

6.2 Brownian motion

Let us see how the properties of the random walk and diffusion appear a physicalsystem. We consider the motion of a small particle in a fluid. The momentum ofthe particle, p = Mv, changes because of collisions with the molecules. When theparticle M is much larger than the molecular mass m then the particle velocity v issmall comparing to the typical velocities of the molecules vT '

√T/m. Then one

can write the force f(p) acting on it as Taylor expansion in p, keeping the first twoterms, independent of p and linear in p: fi(p, t) = fi(0, t) + pj(t)∂fi(0, t)/∂pj(t)(note that we neglected the dependence of the force of the momentum at earliertimes). Such expansion makes sense if the second term is much smaller than thefirst one; then one may ask what is the reason to keep both. The answer is thatmolecules hitting standing particle produce force whose average is zero. The meanmomentum of the particle is zero as well. However, random force by itself wouldmake the squared momentum grow with time exactly like the squared displacementof a random walker in the previous section. To describe the particle in equilibriumwith the medium, the force must be balanced by resistance which is also providedby the medium. That resistance must be described by the second term, whichthen may be approximated as ∂fi/∂pj = −λδij . If the particle radius R is largerthan the mean free path `, in calculating resistance, we can consider fluid as acontinuous medium and characterize it by the viscosity η ' mnvT `, where n isthe concentration of the molecules. For a slow moving particle, v vT `/R, theresistance is given by the Stokes formula

λ = 6πηR/M . (161)

We then obtainp = f − λp . (162)

The solution of the linear equation (162) is

p(t) =

∫ t

−∞f(t′)eλ(t′−t)dt′ . (163)

We must treat the force f(t) as a random function since we do not track mo-lecules hitting the particle, which makes (162) Langevin equation. We assumethat 〈f〉 = 0 and that 〈f(t′) · f(t′ + t)〉 = 3C(t) decays with t during the cor-relation time τ which is much smaller than λ−1. Since the integration time in(163) is of order λ−1 then the condition λτ 1 means that the momentum of aBrownian particle can be considered as a sum of many independent random num-bers (integrals over intervals of order τ) and so it must have a Gaussian statistics

116

Page 117: Physical Nature of Information - Weizmann

ρ(p) = (2πσ2)−3/2 exp(−p2/2σ2) where

σ2 = 〈p2x〉=〈p2

y〉=〈p2z〉=

∫ ∞0C(t1 − t2)e−λ(t1+t2)dt1dt2

≈∫ ∞

0e−2λt dt

∫ 2t

−2tC(t′) dt′≈ 1

∫ ∞−∞

C(t′) dt′ . (164)

On the other hand, equipartition guarantees that 〈p2x〉 = MT so that we can

express the friction coefficient via the correlation function of the force fluctuations(a particular case of the fluctuation-dissipation theorem):

λ =1

2TM

∫ ∞−∞

C(t′) dt′ . (165)

Displacement

∆r(t′) = r(t+ t′)− r(t) =

∫ t′

0v(t′′) dt′′

is also Gaussian with a zero mean. To get its second moment we need the different-time correlation function of the velocities

〈v(t) · v(0)〉 = (3T/M) exp(−λ|t|) (166)

which can be obtained from (163). Note that the friction makes velocity correlatedon a longer timescale than the force. That gives

〈|∆r|2(t′)〉 =

∫ t′

0dt1

∫ t′

0dt2〈v(t1)v(t2)〉 =

6T

Mλ2(λt′ + e−λt

′ − 1) .

The mean squared distance initially grows quadratically (so-called ballistic regimeat λt′ 1). In the limit of a long time (comparing to the relaxation time λ−1

rather than to the force correlation time τ) we have the diffusive growth 〈(∆r)2〉 ≈6Tt′/Mλ. Generally 〈(∆r)2〉 = 2dκt where d is space dimensionality and κ -diffusivity. In our case d = 3 and then the diffusivity is as follows: κ = T/Mλ —the Einstein relation. Using (161) one can rewrite it as follows

κ =T

Mλ=

T

6πηR. (167)

Note that the diffusivity depends on particle radius, but not mass. Heavier parti-cles are slower both to start and to stop moving. Measuring diffusion of particleswith a known size one can determine the temperature25.

25With temperature in degrees, (167) contains the Boltzmann constant, k = κMλ/T ,which was actually determined by this relation and found constant indeed, i.e. independentof the medium and the type of particle. That proved the reality of atoms - after all, kT isthe kinetic energy of a single atom.

117

Page 118: Physical Nature of Information - Weizmann

The probability distribution of displacement at λt′ 1,

ρ(∆r, t′) = (4πκt′)−3/2 exp[−|∆r|2/4κt′] ,

satisfies the diffusion equation ∂ρ/∂t′ = κ∇2ρ. If we have many particles ini-tially distributed according to n(r, 0) then their distribution n(r, t) =

∫ρ(r −

r′, t)n(r′, 0) dr′, also satisfies the diffusion equation: ∂n/∂t′ = κ∇2n.

In the external field V (q), the particle satisfies the equations

p = −λp + f − ∂qV , q = p/M . (168)

Note that these equations characterize the system with the Hamiltonian H =p2/2M + V (q), that interact with the thermostat, which provides friction −λpand agitation f - the balance between these two terms expressed by (165) meansthat the thermostat is in equilibrium.

We now pass from considering individual trajectories to the description of the”cloud” of trajectories. Consider the over-damped limit λ2M ∂2

qqV , where wecan neglect the acceleration term on timescales exceeding the force correlationtime τ : λp p. For example, if we apply to a charged particle an electric fieldE = −∂qV constant in space, then ∂2

qqV = 0; averaging (coarse-graining) overtimes exceeding τ , we can neglect acceleration, since the particle move on averagewith a constant velocity E/λM . In this limit our second-order equation (168) onq is reduced to the first-order equation:

λp = λM q = f − ∂qV . (169)

We can now derive the equation on the probability distribution ρ(q, t), whichchanges with time due to random noise and evolution in the potential, the twomechanisms can be considered additively. We know that without V ,

q(t)− q(0) = (λM)−1∫ t

0f(t′)dt′ , 〈|q(t)− q(0)|2〉 = 2κt ,

and the density ρ(q, t) satisfies the diffusion equation. The dynamical equationwithout any randomness, λM q = −∂qV , corresponds to a flow in q-space with thevelocity w = −∂qV/λM . In that flow, density satisfies the continuity equation∂tρ = −div ρw = −∂qiwiρ. Together, diffusion and advection give the so-calledFokker-Planck equation

∂ρ

∂t= κ∇2ρ+

1

λM

∂qiρ∂V

∂qi= −div J . (170)

More formally, one can derive this equation by writing the Langevin equation(169) as qi − wi = ηi and taking the random force Gaussian delta-correlated:

118

Page 119: Physical Nature of Information - Weizmann

〈ηi(0)ηj(t)〉 = 2κδijδ(t). Since it is the quantity q − w which is Gaussian now,then the path integral representation (159) changes into

ρ(q, t; 0, 0) =

∫Dq(t′) exp

[− 1

∫ t

0dt′|q−w|2

], (171)

To describe the time change, consider the convolution identity (158) for an infini-tesimal time shift ε, then instead of the path integral we get simply the integralover the initial position q′. We substitute q = (q− q′)/ε into (171) and obtain

ρ(q, t) =

∫dq′(4πκε)−d/2 exp

[− [q− q′ − εw(q′)]2

4κε

]ρ(q′, t− ε) . (172)

What is written here is simply that the transition probability is the Gaussianprobability of finding the noise η with the right magnitude to provide for thetransition from q′ to q. We now change integration variable, y = q′+ εw(q′)− q,and keep only the first term in ε: dq′ = dy[1− ε∂q ·w(q)]. Here ∂q ·w = ∂iwi =divw. In the resulting expression, we expand the last factor ρ(q′, t− ε):

ρ(q, t) ≈ (1− ε∂q ·w)

∫dy(4πκε)−d/2e−y

2/4κερ(q + y − εw, t− ε)

≈ (1− ε∂q ·w)

∫dy(4πκε)−d/2e−y

2/4κε[ρ(q, t) + (y − εw) · ∂qρ(q, t)

+(yiyj − 2εyiwj + ε2wiwj)∂i∂jρ(q, t)/2− ε∂tρ(q, t)]

= (1− ε∂q ·w)[ρ− εw · ∂qρ+ εκ∆ρ− ε∂tρ+O(ε2)] , (173)

and obtain (170) collecting terms linear in ε. Note that it was necessary to expanduntil the quadratic terms in y, which gave the contribution linear in ε, namely theLaplacian, i.e. the diffusion operator.

The Fokker-Planck equation has a stationary solution which corresponds tothe particle in an external field and in thermal equilibrium with the surroundingmolecules:

ρ(q) ∝ exp[−V (q)/λMκ] = exp[−V (q)/T ] . (174)

Apparently it has a Boltzmann-Gibbs form, and it turns into zero the probabilitycurrent: J = −ρ∂V/∂q − κ∂ρ/∂q = e−V/T∂(ρeV/T )/∂q = 0. We shall use theFokker-Planck equation in the next section for the consideration of the detailedbalance and fluctuation-dissipation relations.

6.3 General fluctuation-dissipation relation

Fluctuation-dissipation theorem and Onsager reciprocity relations treated smalldeviations from equilibrium. Recently, a significant generalization of equilibrium

119

Page 120: Physical Nature of Information - Weizmann

statistical physics appeared for systems with one or few degrees of freedom devi-ated arbitrary far from equilibrium. This is under the assumption that the restof the degrees of freedom is in equilibrium and can be represented by a ther-mostat generating thermal noise. This new approach also allows one to treatnon-thermodynamic fluctuations, like the negative entropy change.

Consider again the over-damped Brownian particle with the coordinate x(t) ina time-dependent potential V (x, t):

x = −∂xV + η . (175)

Here the random function η(t) can be thought of as representing interaction witha thermostat with the temperature T so that 〈η(0)η(t)〉 = 2Tδ(t). This equation(used very often in different applications) can be applied to any macroscopic ob-servable, where one can distinguish a systematic and random part of the evolution.

The Fokker-Planck equation for the probability ρ(x, t) has the form (170):

∂tρ = T∂2xρ+ ∂x(ρ∂xV ) = −HFPρ . (176)

We have introduced the Fokker-Planck operator,

HFP = − ∂

∂x

(∂V

∂x+ T

∂x

),

which allows one to exploit another instance of the analogy between quantummechanics and statistical physics. We may say that the probability density is theψ-function is the x-representation, ρ(x, t) = 〈x|ψ(t)〉. In other words, we considerevolution in the Hilbert space of functions so that we may rewrite (176) in aSchrodinger representation as d|ψ〉/dt = −HFP |ψ〉, which has a formal solution|ψ(t)〉 = exp(−tHFP )|ψ(0)〉. The only difference with quantum mechanics is thattheir time is imaginary (of course, they think that our time is imaginary). Thetransition probability is given by the matrix element:

ρ(x′, t′;x, t) = 〈x′| exp[(t− t′)HFP )|x〉 . (177)

Without the coordinate-dependent field V (x), the transition probability is sym-metric, ρ(x′, t;x, 0) = ρ(x, t;x′, 0), which is formally manifested by the fact thatthe respective Fokker-Planck operator ∂2

x is Hermitian. This property is called thedetailed balance.

How the detailed balance is modified in an external field? If the potential Vis time independent, then we have a Gibbs steady state ρ(x) = Z−1

0 exp[−βV (x)],where Z0 =

∫exp[−βV (x, 0)] dx. That state satisfies a modified detailed balance:

the probability current is the (Gibbs) probability density at the starting point

120

Page 121: Physical Nature of Information - Weizmann

times the transition probability; forward and backward currents must be equal inequilibrium:

ρ(x′, t;x, 0)e−V (x)/T = ρ(x, t;x′, 0)e−V (x′)/T . (178)

〈x′|e−tHFP−V/T |x〉 = 〈x|e−tHFP−V/T |x′〉 = 〈x′|e−V/T−tH†FP |x〉 .

Since this must be true for any x, x′ then e−tH†FP = eV/T e−tHFP e−V/T and

H†FP ≡(∂V

∂x− T ∂

∂x

)∂

∂x= eV/THFP e

−V/T , (179)

i.e. eV/2THFP e−V/2T is hermitian, which can be checked directly. The quantum-

mechanical notations thus allowed us to translate the detailed balance from theproperty of transition probabilities to that of the evolution operator.

If we now allow the potential to change in time then the system goes awayfrom equilibrium. Consider an ensemble of trajectories starting from the initialpositions taken with the equilibrium Gibbs distribution corresponding to the initialpotential: ρ(x, 0) = Z−1

0 exp[−βV (x, 0)]. As time proceeds and the potentialcontinuously changes, the system is never in equilibrium, so that ρ(x, t) does notgenerally have a Gibbs form. Indeed, even though one can define a time-dependentGibbs state Z−1

t exp[−βV (x, t)] with Zt =∫

exp[−βV (x, t)]dx, one can directlycheck that it is not any longer a solution of the Fokker-Planck equation (176)because of the extra term: ∂tρ = −βρ∂tV . The distribution needs some time toadjust to the potential changes and is generally dependent on the history of these.For example, if we suddenly broaden the potential well, it will take diffusion (withdiffusivity T ) to broaden the distribution. Still, can we find some use of theGibbs factor and also have anything generalizing the detailed balance relation(178) we had in equilibrium? Such relation was found surprisingly recently despiteits generality and relative technical simplicity of derivation.

To find the quantity that has a Gibbs form (i.e. have its probability determinedby the instantaneous partition function Zt), we need to find an equation whichgeneralizes (176) by having an extra term that will cancel the time derivative ofthe potential. It is achieved by considering, apart from a position x, anotherrandom quantity defined as the potential energy change (or the external workdone) during the time t:

Wt =

∫ t

0dt′∂V (x(t′), t′)

∂t′. (180)

The time derivative is partial i.e. taken only with respect to the second argu-ment. The work is a fluctuating quantity depending on the trajectory x(t′), whichdepends on the initial point and noise.

121

Page 122: Physical Nature of Information - Weizmann

Let us now take many different realizations of the noise η(t), choose initial x(0)with the Gibbs probability ρ0 and run (175) many times with every initial data andevery noise realization. It will give us many trajectories having different endpointsx(t) and different energy changes W accumulated along the way. Now considerthe joint probability ρ(x,W, t) to come to x acquiring energy change W . Thistwo-dimensional probability distribution satisfies the generalized Fokker-Planckequation, which can be derived as follows: Similar to the argument preceding(170), we note that the flow along W in x −W space proceeds with the velocitydW/dt = ∂tV so that the respective component of the current is ρ∂tV and theequation takes the form

∂tρ = β−1∂2xρ+ ∂x(ρ∂xV )− ∂W ρ∂tV , (181)

Since W0 = 0 then the initial condition for (181) is

ρ(x,W, 0) = Z−10 exp[−βV (x, 0)]δ(W ) . (182)

While we cannot find ρ(x,W, t) for arbitrary V (t) we can multiply (181) by exp(−βW )and integrate over dW . Since V (x, t) does not depend on W , we get the closedequation for f(x, t) =

∫dWρ(x,W, t) exp(−βW ):

∂tf = β−1∂2xf + ∂x(f∂xV )− βf∂tV , (183)

Now, this equation does have an exact time-dependent solution

f(x, t) = Z−10 exp[−βV (x, t)] ,

where the factor Z−10 is chosen to satisfy the initial condition (182). Note that

f(x, t) is instantaneously defined by V (x, t), no history dependence as we havegenerally in ρ(x, t). In other words, the distribution weighted by exp(−βWt) lookslike Gibbs state, adjusted to the time-dependent potential at every moment oftime. Remark that the phase volume defines probability only in equilibrium, yetthe work divided by temperature is an analog of the entropy change (production),and the exponent of it is an analog of the phase volume change. Let us stressthat f(x, t) is not a probability distribution. In particular, its integral over x isnot unity but the mean phase volume change, which remarkably is expressed viaequilibrium partition functions at the ends (Jarzynski 1997):∫

f(x, t)dx =

∫ρ(x,W, t)e−βWdxdW =

⟨e−βW

⟩=ZtZ0

=

∫e−βV (x,t)dx∫e−βV (x,0)dx

. (184)

Here the bracket means double averaging, over the initial distribution ρ(x, 0) andover the different realizations of the Gaussian noise η(t) during the time interval

122

Page 123: Physical Nature of Information - Weizmann

(0, t). We can also obtain all weighted moments of x like 〈xn exp(−βWt)〉 26. Onecan introduce the free energy Ft = −T lnZt, so that Zt/Z0 = exp[β(F0 − Ft)].

Let us reflect. We started from a Gibbs distribution but considered arbitrarytemporal evolution of the potential. Therefore, our distribution was arbitrarilyfar from equilibrium during the evolution. And yet, to obtain the mean exponentof the work done, it is enough to know the partition functions of the equilibriumGibbs distributions corresponding to the potential at the beginning and at the end(even though the system is not in equilibrium at the end). This is, of course, be-cause the further relaxation to the equilibrium at the end value of the potential isnot accompanied by doing any work. Remarkable that there is no dependence onthe intermediate times. One can also look at it from the opposite perspective: noless remarkable is that one can determine the truly equilibrium property, the freeenergy difference, from non-equilibrium measurements (which could be arbitraryfast rather than adiabatically slow as we used to do in traditional thermodyna-mics).

We can write for the dissipation Wd = W − Ft + F0 (the work minus the freeenergy change) the following identity:

〈e−βWd〉 =

∫dWdρ(Wd) exp(−βWd) = 1 , (185)

which is a generalization of the second law of thermodynamics. Indeed, the meandissipation divided by temperature is the thermodynamic entropy change. Usingthe Jensen inequality 〈eA〉 ≥ e〈A〉, one can obtain the usual second law of thermo-dynamics in the following form:

〈βWd〉 = 〈∆S〉 ≥ 0 .

Moreover, the Jarzynski relation is a generalization of the fluctuation-dissipationtheorem, which can be derived from it for small deviations from equilibrium. Na-mely, we can consider V (x, t) = V0(x) − f(t)x, consider limit of f → 0, expand(184) up to the second-order terms in f and express the response to the field asthe time derivative of the second moment.

When information processing is involved, it must be treated on equal footing,which allows one to decrease the work and the dissipation below the free energydifference:

〈e−βWd−I〉 = 〈e−∆S〉 = 1 . (186)

(Sagawa and Uedo, 2012; Sagawa 2012). We have considered such a case inSection 4.2, where we denoted Wd = Q and used 〈Wd〉 ≥ −IT = −T∆S. Theexponential equality (186) is a generalization of this inequality and (96).

26I thank R. Chetrite for this derivation.

123

Page 124: Physical Nature of Information - Weizmann

So the modern form of the second law of thermodynamics is an equality ratherthan an inequality. The latter is just a partial consequence of the former. Compareit with the re-formulation of the second law in Section 3.3 as a conservation lawrather than a law of increase. And yet (186) is not the most general form. Thefurther generalization is achieved by relating the entropy production to irreversibi-lity, stating that the probability to have a change −∆S in a time-reversed processis as follows (Crooks 1999):

ρ†(−∆S) = ρ(∆S)e−∆S . (187)

Integrating (187) one obtains (184,185,186). That remarkable relation allows alsoone to express the mean entropy production via the relative entropy (85) betweenprobabilities of the forward and backward evolution:

〈∆S〉 =⟨

ln[ρ(∆S)/ρ†(−∆S)]⟩. (188)

Let us derive the relation (187) for our toy model of the generalized baker mapfrom the Section 2.5. Remind that we derived there a long-time average rate ofthe entropy production, which corresponded to the volume contraction of a fluidelement. However, during a finite time n there is always a finite probability toobserve an expansion of an element. This probability must decay exponentiallywith n, and there is a universal law relating relative probabilities of the extractionand contraction. If during n steps a small rectangular element finds themselvesn1 times in the region 0 < x < l and n2 = n − n1 times inside l < x < 1 thenits sides along x and y will be multiplied respectively by l−n1r−n2 and rn1 ln2 .The volume contraction factor for such n-sequence is (l/r)n2−n1 and its log is∆S = n ln J = n1 ln r

l+n2 ln lr . The probability of the sequence is P (ln J) = ln1rn2 .

Opposite sign of ln J will takes place, for instance, in a time-reversed sequence.Time reversal corresponds to the replacement x → 1 − y, y → 1 − x, that is theprobability of such sequence is P (− ln J) = rn1 ln2 . Therefore, denoting the entropyproduction rate σ = − ln J , we obtain the universal probability independent of r, l:

P (∆S)

P (−∆S)=

P (σ)

P (−σ)=

(l

r

)n2−n1

= enσ = e∆S . (189)

In a multi-dimensional case, apart from making the potential time-dependent,there is another way to deviate the system from equilibrium: to add to the randomthermal force f(t) and the potential force −∂qV (q) another coordinate-dependentforce F(q) which is non-potential (not a gradient of any scalar):

p = −λp+ f − ∂qV + F , q = p/M .

124

Page 125: Physical Nature of Information - Weizmann

The dynamical equations (with λ = 0 and f = 0) have the form p = −∂qH + F,q = ∂pH, where H = p2/2M +V . Since the non-potential force makes the systemnon-Hamiltonian, contact with the thermostat generally does not lead to thermalequilibrium, as we discussed in Section 2.5. The equation on the full phase-spacedistribution ρ(p,q, t) has the form

∂tρ = H, ρ+ T∆pρ+ ∂pρF = HKρ . (190)

It is called the Kramers equation (Fokker-Planck equation follows from it in theoverdamped limit). Only without F, the Gibbs distribution exp(−H/T ) is a steadysolution of (190) and one can formulate the detailed balance

H†K = ΠeβHHKe−βHΠ−1 , (191)

where we added the operator inverting momenta: ΠpΠ−1 = −p. A non-potentialforce violates the the detailed balance in the following way:

H†K = ΠeβHHKe−βHΠ−1 + β(F · q) . (192)

The last (symmetry-breaking) term is again the power (F · q) divided by tempe-rature i.e. the entropy production rate. The work done by that force depends onthe trajectory in distinction from the case of a time-independent potential force.That dependence of the work on the trajectory precludes thermal equilibrium andis common for non-potential forces and for time-dependent potential forces. Aclose analog of the Jarzynski relation can be formulated for the production rateaveraged during the time t:

σt =1

tT

∫ t

0(F · q) dt . (193)

If F = dU/dq, that is a gradient of a scalar, then (F · q) = dU(q(t))/dt, andthe integral turns into zero. The quantity (193) fluctuates from realization torealization. The probabilities P (σt) satisfy the relation, analogous to (187), whichwe give without general derivation

P (σt)

P (−σt)∝ etσt . (194)

The second law of thermodynamics states that to keep the system away fromequilibrium, the external force F must on average do a positive work. Over along time we thus expect σt to be overwhelmingly positive, yet fluctuations dohappen. The relations (194,189) show how low is the probability to observe anegative entropy production rate - this probability decays exponentially with the

125

Page 126: Physical Nature of Information - Weizmann

time of observation. Such fluctuations were unobservable in classical macroscopicthermodynamics, but they are often very important in modern applications to nanoand bio objects. In the limit t → ∞, when the probability of the integral (193)must have a large-deviation form, P (σt) ∝ exp[−tH(σt)], so that (194) means thatH(σt)−H(−σt) = −σt, as if P (σt) was Gaussian.

One calls (187,194) detailed fluctuation-dissipation relations since they arestronger than integral relations of the type (184,185). Indeed, it is straightforwardto derive 〈exp(−tσt)〉 = 1 from (194).

The relation similar to (194) can be derived for any system symmetric withrespect to some transformation, to which we add perturbation anti-symmetric withrespect to that transformation. Consider a system with the variables s1, . . . , sNand the even energy: E0(s) = E0(−s). Consider the energy perturbed by an oddterm, E = E0 − hM/2, where M(s) =

∑si = −M(−s). The probability of the

perturbation P [M(s)] satisfies the direct analog of (194), which is obtained bychanging the integration variable s→ −s:

P (a)=

∫dsδ[M(s)−a]eβ(ha−E0) =

∫dsδ[M(s)+a]e−β(ha+E0) =P (−a)e−2βha .

The validity condition for the results in this Section is that the interaction with thethermostat is represented by noise independent of the the evolution of the degreesof freedom under consideration.

6.4 Stochastic Web surfing and Google’s PageRank

When it was proclaimed that the Library contained all books, the firstimpression was one of extravagant happiness... As was natural, this was

followed by an excessive depression. The certitude that some ... preciousbooks were inaccessible seemed almost intolerable.

J L Borges ”The Library of Babel”Approaching the conclusion of the course, we cannot any more avoid the que-

stion: can we find an objective and quantitative measure not only of the amountof information, but also of its importance? We need to know which are the mostprecious books in the Library. By this time, it should come as no surprise for thereader that such measures can also be found using the statistical approach.

For an efficient information retrieval from the Web Library, webpages needto be ranked by their importance to order search results. A reasonable way tomeasure the importance of a page is to count the number of links that refer to it.Not all links are equal though — those from a more important page must bringmore importance. On the other hand, a link from a page with many links mustbring less importance (here probability starts creeping in). One then comes to

126

Page 127: Physical Nature of Information - Weizmann

the two rules: i) every page relays its importance score to the pages it links to,dividing it equally between them, ii) the importance score of a page is the sum ofall scores obtained by links. For Internet with n pages, we organize all their scoresinto a vector p = p1, . . . , pn which we normalize:

∑ni=1 pi = 1. According to the

above rules, pi =∑j pj/nj where nj is the number of outgoing links on page j,

which links to the page i. In other words, we are looking for the eigenvector ofthe hyperlink matrix, pA = p, where the matrix elements aij = 1/nj if j links toi and aij = 0 otherwise. Does a unique eigenvector with all non-negative entriesand a unit eigenvalue always exist? If yes, how to find it?

The iterative algorithm to find the score eigenvector is called PageRank27 (Brinand Page 1998). The algorithm equates the score pi of a page with the probabi-lity that a person randomly clicking on links will arrive at this page. It startsfrom ascribing equal probability to all pages, pi(0) = 1/n, and generates the newprobability distribution by applying the above rules of the score relay:

p(t+ 1) = p(t)A . (195)

We recognize that this stochastic process is a Markov chain, which means that thefuture is determined by the present state and the transition probability, but notby the past. We thus interpret A as the matrix of transition probabilities betweenpages for our stochastic surfer. In later modifications, one fills the elements of Anot uniformly as 1/nj but use information about actual frequencies of linking thatcan be obtained from access logs. Could our self-referential rules lead to a viciouscircle or the iterations converge at t→∞? It better be convergent fast to be of anyuse for the instant-gratification generation. It is clear that if the largest eigenvalueλ1 of A was larger than unity, than the iterations would diverge; if λ1 < 1, thenthe iterations would converge to zero. We need the largest eigenvalue to be unityand correspond to a single eigenvector, so that the iterations converge. How fastit converges then will be determined by the second largest eigenvalue λ2 (whichmust be less than unity).

Moment reflection is enough to identify the problem: some pages do not link toany other page, which corresponds to rows of zeroes in A. Such pages accumulatethe score without sharing it. Another problem is caused by loops. The figurepresents a simple example illustrating both problems:

27”Page” relates both to webpage and to Larry Page, who with Sergei Brin invented thealgorithm and created Google.

127

Page 128: Physical Nature of Information - Weizmann

If all transition probabilities are nonzero, the probability vector with time tendsto (0, 0, 1), that is the surfer is stuck at the page 3. When the probabilities a13, a23

are very small, the surfer tend to be caught for long times in the loop 1←→ 2.To release our random surfer from being stuck at a sink or caught in a loop,

the original PageRank algorithm allowed it to jump randomly to any other pagewith equal probability. To be fair with pages that are not sinks, these randomteleportations are added to all nodes in the Web: surfer either clicks on a link onthe current page with probability d or opens up a random page with probability1 − d. To quote the original: ”We assume there is a ”random surfer” who isgiven a web page at random and keeps clicking on links, never hitting ”back” buteventually gets bored and starts on another random page. The probability thatthe random surfer visits a page is its PageRank. And, the damping factor is theprobability at each page the ”random surfer” will get bored and request anotherrandom page.” This is equivalent to replacing A by G = dA + (1 − d)E. Herethe teleportation matrix E has all entries 1/n, that is E = eeT /n, where e is thecolumn vector with ei = 1 for i = 1, . . . , n. After that, all matrix entries aij arestrictly positive and the graph is fully connected.

It is important that now our matrix has positive elements in every columnwhose sum is unity. Such matrices are called stochastic, since every column canbe thought of as a probability distribution. Every stochastic matrix has unityas the largest eigenvalue. Indeed, since

∑j aij = 1, then e is an eigenvector of

the transposed matrix: ATe = e. Therefore, 1 is an eigenvalue for AT , and alsofor A, which has the same eigenvalues. We can now use convexity to prove thatthis is the largest eigenvalue. For any vector p, every element of pA is a convexcombination of the elements,

∑j pjaij , which cannot exceed the largest element

of p since∑j aij = 1. For an eigenvector with an eigenvalue exceeding unity,

at least one element of pA must exceed the largest element of p, therefore sucheigenvector cannot exist. This is a particular case of the theorem: The eigenvaluewith the largest absolute value of a positive square matrix is positive, and belongsto a positive eigenvector, where all of the vector’s elements are positive. All othereigenvectors are smaller in absolute value (Markov 1906, Perron 1907).

The great achievement of PageRank algorithm is the replacement of the itera-tive process (195) by

p(t+ 1) = Gp(t) . (196)

That process cannot be caught into a loop and converges, which follows from thefact that Gii 6= 0 for all i; that is there is always a probability to stay on the pagebreaking any loop. The eigenvalues of G are 1, dλ2 . . . dλn (prove it), so the choiceof d affects convergence, the smaller the faster. On the other hand, it is somewhatartificial to use teleportation to an arbitrary page, so larger values of d give more

128

Page 129: Physical Nature of Information - Weizmann

weight to the true link structure of the Web. As in other optimization problemswe encountered in this course, one needs a workable compromise. The standardGoogle choice d = 0.85 comes from estimating how often an average surfer usesbookmarks. As a result, the process usually converges after about 50 iterations.

One can design a personalized ranking by replacing the teleportation matrixby E = evT , where the probability vector v has all nonzero entries and allows forpersonalization, that is can be chosen according to the individual user’s history ofsearches and visits. That means that it is possible in principle to have our personalrankings of the webpages and make searches custom-made.

As mentioned, the sequence of the probability vectors defined by the relationsof the type (195,196) is a Markov chain. In particular, the three random quantitiesX → Y → Z is a Markov triplet if Y is completely determined by X,Z, whileX,Z are independent conditional on Y , that is I(X,Z|Y ) = 0. Such chains havean extremely wide domain of applications.

7 Conclusion

This Chapter attempts to compress the course to its most essential elements.

7.1 Take-home lessons

1. Thermodynamics studies restrictions imposed by hidden on observable.It deals with two fundamental extensive quantities. The first one (energy) Eis conserved for a closed system, and its changes are divided into work (due toobservable degrees of freedom) and heat (due to hidden ones). The second quantity(entropy) S can only increase for a closed system and reaches its maximum inthermal equilibrium, where the system entropy is a convex function of the energy.All available states lies below this convex curve in S − E plane.

2. Convexity of the dependence E(S) allows us to introduce temperatureas the derivative of the energy with respect to the entropy. Extremum of theentropy means that the temperatures of the connected subsystems are equal inequilibrium. The same is true for the energy derivatives with respect to volumeand other extensive variables. The entropy increase (called the second law ofthermodynamics) imposes restrictions on thermal engine efficiency, that is thefraction of heat used for work:

W

Q1=Q1 −Q2

Q1= 1− T2∆S2

T1∆S1≤ 1− T2

T1.

T

Q

W

T2

1

1

2Q

129

Page 130: Physical Nature of Information - Weizmann

Similarly, if information processing generates ∆S, its energy price is as follows:

Q =T2∆S +W

1− T2/T1.

T

T2

1

QS =Q/T1

WW

− QS =(Q−W)/T2

1

2

3. Need in statistics appear due to incomplete knowledge: We can measure onlypart of the degrees of freedom; even if we measure them all, we do it with a finiteprecision. Statistical physics defines the (Boltzmann) entropy of a closed system asthe log of the phase volume, S = log Γ and assumes (for the lack of any knowledge)the uniform distribution w = 1/Γ called microcanonical. For a subsystem, the(Gibbs) entropy is defined as the mean phase volume: S = −

∑iwi logwi; the

probability distribution is then obtained requiring maximal entropy for a givenmean energy: logwi ∝ −Ei. Information theory generalizes this approach, see 13below.

4. Irreversibility of the entropy growth seems to contradict Hamiltonian dyna-mics, which is time-reversible and preserves the N -particle phase-space probabilitydensity. However, one can obtain the equation on a one-particle density for a dilutegas. If we then assume that before every collision the particles were independent,then we obtain the Boltzmann kinetic equation, which, in particular, describes theirreversible growth of the one-particle entropy. Since the total entropy is concer-ned, while the sum of one-particle entropies grow, we conclude that their differencemust grow too. Later, we call it the mutual information. The lesson is that if wefollow precisely all the degrees of freedom, the entropy is conserved and no infor-mation is lost. But if we follow only part of them, the entropy of that part willgenerally grow — whatever information we had is getting less relevant with time.

5. The total entropy growth can appear even if we follow all the degrees offreedom, but do it with a finite precision. In this case, we essentially consider evo-lution of finite phase-space regions. Instability leads to separation of trajectories,which spread over the whole phase space under a generic reversible Hamiltoniandynamics, very much like flows of an incompressible liquid are mixing (we maysay metaphorically, that for unstable systems, any extra digit in precision adds anew degree of freedom). Spreading and mixing in phase space correspond to theapproach to equilibrium and entropy growth. On the contrary, to deviate a systemfrom equilibrium, one adds external forcing and dissipation, which makes its phaseflow compressible and distribution non-uniform.

6. Basic mathematical object we use in our discrete thinking is the sum of inde-pendent random numbers X =

∑Ni=1 yi. Three concentric statements were made.

The weakest statement is that X approaches its mean value X = N〈y〉 exponen-tially fast in N . The next statement is that the distribution P(X) is Gaussian in

130

Page 131: Physical Nature of Information - Weizmann

the vicinity N−1/2 of the maximum. For even larger deviations, the distributionis very sharp: P(X) ∝ e−NH(X/N) where H ≥ 0 and H(〈y〉) = 0. Applying thisto the log of the probability of a given sequence, limN→∞ p(y1 . . . yN ) = −NS(Y ),we learn two lessons: i) the probability is independent of a sequence for most ofthem (almost all events are almost equally probable), ii) the number of typicalsequences grows exponentially and the entropy is the rate.

7. Another simple mathematical property we use throughout is convexity.We first use it in the thermodynamics to make sure that the extremum is onthe boundary of the region and to make Legendre transform of thermodynamicpotentials. We then use convexity of the exponential function in establishing thateven when the mean of a random quantity is zero, its mean exponent is positive,which provides for an exponential separation of trajectories in an incompressibleflow and exponential growth of the density of an element in a compressible flow.

8. Since uncertainty or the lack of knowledge play such a prominent role, wewish to quantify it. The measure of uncertainly is the amount of informationneeded to remove it. This is consistently done in a discrete case, for instance,by counting the number of bits, that is answers to ”yes-no” questions. Thatway we realize that the information is log2 of the number of equally probablepossibilities (Boltzmann entropy) or the mean logarithm if the probabilities aredifferent (Shannon-Gibbs entropy). Here convexity of the function −w logw helpsus to prove that the information/entropy has its maximum for equal probabilities(when our ignorance is maximal).

9. The point 6 above states that the number of typical sequences grows withthe rate equal to the entropy I. The number of typical binary sequences of lengthN is then 2NI , which is smaller than 2N . The efficient encoding of the typicalsequences thus involves words with lengths from unity to NI, which is less thanN if the probabilities of 0 and 1 are not equal. That means that the entropy isboth the mean and the fastest rate of the reception of information brought bylong messages/measurements. To squeeze out all the unnecessary bits, encodingis used both in industry and in nature where sources often bring highly redundantinformation, like in visual signals.

10. If the transmission channel B → A makes errors, then the message doesnot completely eliminate uncertainty; what remains is the conditional entropyS(B|A) = S(A,B) − S(A), which is the mean rate of growth of the number ofpossible errors. Sending extra bits to correct these errors lowers the transmissionrate from S(B) to the mutual information I(A,B) = S(B) − S(B|A), which isthe mean difference of the uncertainties before and after the message. The greatnews is that one can still achieve an asymptotically error-free transmission if thetransmission rate is lower than I. The maximum of I over all source statisticsis the channel capacity, which is the maximal rate of asymptotically error-free

131

Page 132: Physical Nature of Information - Weizmann

transmission. In particular, to maximize the capacity of sensory processing, theresponse function of a living beings or a robot must be a cumulative probabilityof stimuli.

11. Very often our goal is not to transmit as much information as possible,but to compress it and process as little as possible, looking for an encoding witha minimum of the mutual information. For example, the rate distortion theorylooks for the minimal rate I of information transfer under the restriction that thesignal distortion does not exceed the threshold D. This is done by minimizing thefunctional I+βD. Another minimization task could be to separate the signal intoindependent components with as little as possible (ideally zero) mutual informationbetween them.

12. The conditional probability allows for hypothesis testing by the Bayes’ rule:P (h|e) = P (h)P (e|h)/P (e). That is the probability P (h|e) that the hypothesis iscorrect after we receive the data e is the prior probability P (h) times the supportP (e|h)/P (e) that e provide for h. Taking a log and averaging we obtain familiarS(h|e) = S(h) − I(e, h). If our hypothesis concerns the probability distributionitself, then the difference between the true distribution p and the hypotheticaldistribution q is measured by the relative entropy D(p|q) = 〈log2(p/q)〉. This is yetanother rate — with which the error probability grows with the number of trials.D also measures the decrease of the transmission rate due to non-optimal encoding:the mean length of the codeword is not S(p) but bounded by S(p) +D(p|q).

13. Since so much hangs on getting the right distribution, how best to guess itfrom the data? This is achieved by maximizing the entropy under the given data— ”the truth and nothing but the truth”. That explains and makes universal theapproach from the point 3. It also sheds new light on physics, telling us that onsome basic level all states are constrained equilibria.

14. Information is physical: to learn ∆S = S(A)−S(A,M) one does the workT∆S, where A is the system and M is the measuring device. To erase information,one needs to convert TS(M) into heat. Both acts require a finite temperature. Theenergetic price of a cycle is T times the mutual information: TI(A,M). Anotherside of the physical nature of information is that there is the (Bekenstein) limit onhow much entropy one can squeeze inside a given a radius; surprisingly, the limitis proportional to the area rather than the volume and is realized by black holes— our gates outside of this world.

15. The Renormalization Group is a best so far known way to forget infor-mation. Apart from forgetting in the first step, it involves the second step ofrenormalization. The focus is on the change of the probability distribution andthe appearance of an asymptotic distribution after many steps. We find that theentropy of the partially averaged and renormalized distribution is the proper me-asure of forgetting in simple cases, like adding random numbers on the way to

132

Page 133: Physical Nature of Information - Weizmann

the central limit theorem. In physical systems with many degrees of freedom, itcan be the mutual information defined in two ways: either between remaining andeliminated degrees of freedom or between different parts of the same system. Inparticular, it shows us examples of the area law, when I is sub-extensive.

16. That area law was the initial motivation for the quantum information the-ory, since the entropy of a black hole is proportional to its area rather than volume.Nowadays, of course, it is also driven by the quest for a quantum computer. Al-ready quantum mechanics has a natural entropic formulation. Quantum statisticsappears when we treat subsystems and must deal with the density matrix and itsvon Neumann entropy. The quantum entropy of the whole can be less than theentropy of a part. In particular, the whole system can be in a pure state with zeroentropy, then all the entropy of a subsystem comes from entanglement — such isa quantum sibling of the mutual information, since it also measures the degree ofcorrelation.

17. The last lesson is two progressively more powerful forms of the secondlaw of thermodynamics, which originally was 〈∆S〉 ≥ 0. The first new form,〈e−∆S〉 = 1, is the analog of a Liouville theorem. The second form relates theprobabilities of forward and backward process: ρ†(−∆S) = ρ(∆S)e−∆S .

7.2 Epilogue

The central idea of this course is that learning about the world means building amodel, which is essentially finding an efficient representation of the data. Opti-mizing information transmission or encoding may seem like a technical problem,but it is actually the most important task of science, engineering and survival.Science works on more and more compact encoding of the strings of data, whichculminates in formulating a law of nature, potentially describing infinity of phe-nomena. The mathematical tool we learnt here is an ensemble equivalence in thethermodynamic limit, its analog is the use of typical sequences in communicationtheory. The result is two roles of entropy: it defines maximum transmission andminimum compression.

Another central idea is that entropy is not a property of the physical world,but is an information we lack about it. And yet the information is physical — ithas an energetic value and a monetary price. Indeed, the difference between workand heat is that we have information about the former but not the later. Thatmeans that one can turn information into work and one needs to release heat toerase information. We also have learnt that one not only pays for information butcan turn information into money as well. The physical nature of information ismanifested in the universal limit on how much of it we can squeeze into a spacerestricted by a given area.

133

Page 134: Physical Nature of Information - Weizmann

Reader surely recognized that no rigorous proofs were given, replaced insteadby plausible hand-waving argument or even a particular example. Those interestedin proofs for Chapter 2 can find them in Dorfman ”An Introduction to Chaos inNonequilibrium Statistical Mechanics”. Detailed information theory with proofscan be found in Cowen & Thomas ”Elements of Information Theory”, whose Chap-ter 1 gives a concise overview. Nor the examples given are representative of theever-widening avalanche of applications; more biological applications can be foundin ”Biophysics” by Bialek, others in original articles and reviews. Numerous re-ferences scattered through the text, like (Zipf 1949), give you the most compactencoding of what is to google to find details.

Mention briefly one important subject left out of this course. Our focus wasmainly on finding a data description that is good on average. Yet there existsa closely related approach that focuses on finding the shortest description andultimate data compression for a given string of data. The Kolmogorov complexityis defined as the shortest binary computer program able to compute the string. Itallows us to quantify how much order and randomness is in a given sequence —truly random sequence cannot be described by an algorithm shorter than itself,while any order allows for compression. Complexity is (approximately) equal to theentropy if the string is drawn from a random distribution, but is actually a moregeneral concept, treated in courses on Computer Science. Another fundamentalissue not treated here is the dramatic difference between the classical and quantumclassifications of computational complexity.

Taking a wider view, I invite you to reflect on the history of our attempts torealize limits of possible, from heat engines through communication channels tocomputations.

Looking back one may wonder why accepting the natural language of informa-tion took so much time and was so difficult for physicists and engineers. Generati-ons of students (myself including) were tortured by ”paradoxes” in the statisticalphysics, which disappear when information language is used. I suspect that theresistance was to a large extent caused by the misplaced desire to keep scientistout of science. A dogma that science must be something ”objective” and onlyrelated to the things independent of our interest in them obscures the simple factthat science is a form of human language. True, we expect it to be objectivelyindependent of personality of this or that scientist as opposite, say, to literature,where we celebrate the difference between languages (and worlds) of Tolstoy andChekhov. However, science is the language designed by and for humans, so thatit necessarily reflects both the way body and mind operate and the restrictions onour ability to obtain and process the data. Presumably, omnipresent and omnis-cient being would have no need in the statistical information approach describedhere. One may also wonder to what extent essential presence of scientist in science

134

Page 135: Physical Nature of Information - Weizmann

may help us understand the special status of measurement in quantum mechanics.I wish to thank my colleagues whose advices were invaluable in preparing the

course: Yakov Sinai, Alexander Zamolodchikov, Massimo Vergassola, Eric Siggia,Leonid Levitov, Alexei Kitaev. Special thanks to Michal Shavit for preparingtutorials and catching numerous errors (the remaining ones are my responsibilityalone).

As we learnt here, better understanding must lead to a more compact presen-tation; hopefully, the next version of these lecture notes will be shorter.

8 Appendix 1: Continuous distributions

For the sake of completeness, we generalize (63) for a continuous distribution bydividing into cells (that is considering a limit of discrete points). Different choicesof variables to define equal cells give different definitions of information. It is insuch a choice that physics (or other specific knowledge) enters. Physics (quantummechanics) requires that for Hamiltonian system the equal volumes in phase spacecontain equal number of states, so the measure is uniform in canonical coordinates;we then write entropy in terms of the phase space density ρ(P,Q, t), which mayalso depend on time:

S(P,Q) = −∫ρ log ρ dPdQ . (197)

It is maximal for the uniform distribution ρ = 1/Γ = 1/∫dPdQ, I = ln Γ. To

avoid trivial errors, always consider Γ ≥ 0. Everywhere in this section we assumethat the integrals are absolutely convergent. The entropy defined in such a waycan be both positive, and negative, and can even be −∞. In distinction from adiscrete case, it is invariant with respect to shifts but not re-scaling of the variables:S(ax+ b) = S(x) + log a.

If the density of the discrete points in the continuous limit is inhomogeneous,say m(x), then the proper generalization is

S = −∫ρ(x) ln[ρ(x)/m(x)] dx .

It is invariant with respect to an arbitrary change of variables x → y(x) sinceρ(y)dy = ρ(x)dx and m(y)dy = m(x)dx while (197) was invariant only withrespect to canonical transformations (including a time evolution according to aHamiltonian dynamics) that conserve the element of the phase-space volume. Ifwe introduce the normalized distribution of points ρ′(x) = m(x)/Γ, then

S = ln Γ−∫ρ(x) ln[ρ(x)/ρ′(x)] dx . (198)

135

Page 136: Physical Nature of Information - Weizmann

The last term in (198) is the relative entropy between the distributions.Since statistics and information is ultimately about counting, it must be dis-

crete. Continuous treatment is just an approximation often convenient for analytictreatment; to avoid infinities and divergencies (like log of zero) it is convenient towork with differentials.

9 Appendix 2: On Zipf law

Probably the simplest model that gives such a distribution is random typing: allletters plus the space are taken with equal probability (Wentian Li 1992). Thenany word with the length L is flanked by two spaces and has the probabilityPi(L) = (M + 1)−L−2/Z, where i = 1, 2, . . . ,ML and M is the alphabet size. Thenormalization factor is Z =

∑LM

L(M + 1)−L−2 = (M + 1)2/M . On the otherhand, the rank r(L) of any L-word satisfies the inequality

M(ML−1 − 1)/(M − 1) =L−1∑i=1

M i < r(L) ≤L∑i=1

M i = M(ML − 1)/(M − 1) ,

which can be written as Pi(L) < C[r(L)+B]−α ≤ Pi(L−1) with α = logM (M+1),B = M/(M − 1) and C = Bα/M . In the limit of large alphabet size, M 1, weobtain

P (r) = (r + 1)−1 . (199)

This asymptotic actually takes place for wide classes of letter distributions, notnecessarily equiprobable. Closely related way of interpreting statistical distribu-tions is to look for variational principle it satisfies. One may require maximalinformation transferred with the least effort. The rate of information transferis S = −

∑r P (r) logP (r). The effort must be higher for less common words,

that is to grow with the rank. Such growth can be logarithmic (for instance,when the effort is proportional to the word length). The mean effort is thenW =

∑r P (r) log r. Looking for the minimum of S − λW , we obtain P (r) ∝ r−λ.

Zipf law corresponds to λ = 1, when goals and mean are balanced, In what followswe shall be minimizing a lot of two-term functionals and looking for a conditionalentropy maximum (which is what most of statisticians do most of the time).

Does then the Zipf law trivially appear because both number of words (inverseprobability) and rank increase exponentially with the word size? The answer isnegative because the number of distinct words of the same length in real languageis not exponential in length and is not even monotonic. It is reassuring that ourtexts are statistically distinguishable from those produced by an imaginary monkeywith a typewriter. Moreover, words have meaning. The number of meanings

136

Page 137: Physical Nature of Information - Weizmann

(counted, for instance, from the number of dictionary entries for a word) growsapproximately as the square root of the word frequency: mi ∝

√Pi. Meanings

correspond to objects of reference having their own probabilities, and it seems thatthe language combines these objects into groups whose sizes are proportional tothe mean probability of the group pi, so that Pi = mipi ∝ m2

i . It is temptingto suggest that the distributions appeared due to the balance between minimizingefforts of writers and readers, speakers and listeners. Writers and speakers wouldminimize their effort by having one word meaning everything and appearing withthe probability one. On the other end, difficulty of perception is proportional to thedepth of the memory keeping the context, needed, in particular, for choosing theright meaning. Readers and listeners then prefer a lot of single-meaning words. Sofar, no convincing optimization scheme giving different features of word statisticswas found28.

28Maybe, the statistical laws of language can be better understood by viewing conver-sation not as an exchange of information but as a mating game with multiple synonymsand meanings as a verbal plumage.

137