Foundations of Entropy I
Why Entropy?
Lecture series at the
School on Information, Noise, and Physics of Life
Nis 19.-30. September 2022
by Jan Korbel
all slides can be found at: slides.com/jankorbel
Warning!
(Korbel = Tankard = Bierkrug)
No questions are stupid
Please ask anytime!
Activity I
You have 3 minutes to write down on a piece of paper:
a) Your name
b) What do you study
c) What is entropy to you? (Formula/Concept/Definition/...)
\( S = k \cdot \log W\)
My take on what is entropy
Located at Vienna central cemetery
(Wien Zentralfriedhof)
We will get back to this formula
Why so many definions?
There is more than one face of entropy!
What measures entropy?
Randomness?
Disorder?
Energy dispersion?
Maximum data compression?
'Distance' from equilibrium?
Uncertainty?
Heat over temperature?
Information content?
Part of the internal energy unavailable for useful work?
Or is it just a tool? (Entropy = thermodynamic action?)
MaxEnt
MaxCal
SoftMax
MaxEP
Prigogine
My courses on entropy
a.k.a. evolution of how powerful entropy is
Field: mathematical physics
warning: Personal opinion!
SS 1st year Bc. - Thermodynamics
\(\mathrm{d} S = \frac{\delta Q}{T} \)
\(C_v = T \left( \frac{\partial S}{\partial T}\right)_V\)
\(\left ( \frac{\partial S}{\partial V}\right)_T = \left(\frac{\partial p}{\partial T} \right)_V \)
SS 2nd year Bc. - Statistical physics
\( S = - \sum_k p_k \log p_k\)
\(Z = \sum_k e^{-\beta \epsilon_k}\)
\(\ln Z = S - U/T\)
SS 3rd year Bc. Quantum mechanics 2
\( S = -Tr (\rho \log \rho) \)
\( Z = Tr (\exp(-\beta \hat{H}))\)
Bachelor's studies
differential forms?
probability theory?
Master's studies
Erasmus exchange
@ FU Berlin
WS - 2nd year MS - Advanced StatPhys
Fermi-Dirac & Bose-Einstein statistics
Ising spin model and transfer matrix theory
Real gas and virial expansion
WS - 2nd year MS - Noneq. StatPhys
Onsager relations
Molecular motors
Fluctuation theorems
Historical intermezzo
by
Motivations for introducing entropy
1. relation between energy, heat, work and temperature
Thermodynamics (should be rather thermoSTATICS)
2. relation between microscopic and macroscopic
R. Clausius
Lord Kelvin
H. von Helmholtz
S. Carnot
J. C. Maxwell
L. Boltzmann
M. Planck
J. W. Gibbs
Statistical mechanics/physics
Why statistical physics?
Microscopic to Macroscopic
Statistical Physics = Physics + Statistics
Role of statistics in physics
Classical mechanics (quantum mechanics)
- position & momenta given by equations of motion
- 1 body problem: solvable
- 2 body problem: center of mass transform
- 3 body problem: generally not solvable
...
- N body problem: ???
Do we need to know trajectories of all particles?
Liouville theorem
Let's have canonical coordinates \(\mathbf{q}(t)\), \(\mathbf{p}(t)\) evolving by Hamiltonian dynamics
$$\dot{\mathbf{q}} = \frac{\partial H}{\partial \mathbf{p}}\qquad \dot{\mathbf{p}} = - \frac{\partial H}{\partial \mathbf{q}}$$
Let \(\rho(p,q,t)\) be a probability distribution in the phase space. Then, \(\frac{\mathrm{d} \rho}{\mathrm{d} t} = 0.\)
Consequence: \( \frac{\mathrm{d} S(\rho)}{\mathrm{d} t}= - \frac{\mathrm{d}}{\mathrm{d} t} \left(\int \rho(t) \ln \rho(t)\right) = 0.\)
Useful results from statistics
1. Law of large numbers (LLN)
\( \sum_{i=1}^n X_i \rightarrow n \bar{X} \quad \mathrm{for} \ N \gg 1\)
2. Central limit theorem (CLT)
\( (\frac{1}{n} \sum_{i=1}^n X_i - \bar{X}) \rightarrow \frac{1}{\sqrt{n}} \mathcal{N}(0,\sigma^2)\)
Consequence: a large number of i.i.d. subsystems can be described by very few parameters for \(N \gg 1\)
\(\Rightarrow\) e.g., a box with 1 mol of gas particles
Useful results from combinatorics
Bars & Stars theorems (|*)
Emergence of statistical physics:
Coarse-graining
\(\bar{X}\)
Coarse-graining or Ignorance?
Microscopic systems
Classical mechanics (QM,...)
Mesoscopic systems
Stochastic thermodynamics
Macroscopic systems
Thermodynamics
Trajectory TD
Ensemble TD
Statistical mechanics
Coarse-graining in thermodynamics
Microstates, Mesostates and Macrostate
Consider again a dice with 6 states
Let us throw a dice 5 times. The resulting sequence is
Microstate
The histogram of this sequence is
0
0
2
1
1
1
Mesostate
The average value is 3,8 Macrostate
Coarse-graining
Coarse-graining
# micro: \(6^5 =7776\)
# meso: \(\binom{6+5-1}{5} =252\)
# macro: \( 5\cdot 6-5\cdot 1 =25\)
Multiplicity W (sometimes \(\Omega\)):
# of microstates with the same mesostate/macrostate
Now we come back to the formula on Boltzmann's grave
Question: how do we calculate multiplicity W for mesostate
Answer: see combinatorics lecture.
Full answer: 1.) permute all states, 2.) take care of overcounting
1.) Permuation of all states: 5! = 120
2.) Overcounting - permutation of 2! = 2
Together: \(W(0,2,0,1,1,1) = \frac{5!}{2!} =60\)
0
0
2
1
1
1
"General" formula - multinomials
$$W(n_1,\dots,n_k) = \left(\frac{\sum_{i=1}^k n_i}{n_1, \ \dots \ ,n_k}\right) = \frac{(\sum_{i=1}^k n_k)!}{\prod_{i=1}^k n_i!}$$
The question at stake: WHY \(\log \)?
Succint reason: \(\log\) transforms \(\prod\) to \(\sum\)
(similar to log-likelihood funciton)
Physical reason: multiplicity of \(X \times Y\)
is \(W(X)W(Y)\)
(extensivity/intensivity of thermodynamic variables)
Boltzmann entropy = Gibbs entropy?
\(\log W(n_1,\dots,n_k) = n \log n - \cancel{n} - \sum_{i=1}^k n_i \log n_i + \cancel{\sum_{i=1}^k n_i} \)
\(= \sum_{i=1}^k n_i (\log n - \log n_i) = - \sum_{i=1}^k n_i \log \frac{n_i}{n}\)
Stirling's approximation: \( \log(n!) \approx n \log n - n + \mathcal{O}(\log n) \)
Denote: \(\sum_{i=1}^k n_k = n\).
Denote: \(n_i/n = p_i\).
$$\log W(n_1,\dots,n_k) = - n \sum_{i=1}^k p_i \log p_i $$
What is actually \(p_i\)?
Frequentist vs Bayesian probability
In probability, there are two interpretations of probability
1. Frequentist approach
probability is the limiting success value of a repeated experiment
$$p = \lim_{n \rightarrow \infty} \frac{k(n)}{n}$$
It can be estimated as \(\hat{p} = \frac{X_1+\dots+X_n}{n}\) and it does not make any sense to consider parametric distribution.
2. Bayesian approach
probability quantifies our uncertainty about the experiment. By observing the experiment we can update our knowledge about it
$$\underbrace{f(p|\hat{p})}_{posterior} = \underbrace{\frac{f(\hat{p}|p)}{f(\hat{p})}}_{likelihood \ ratio} \underbrace{f(p)}_{prior} $$
LLN
Thermodynamic limit
By using the relation \(n_i/n = p_i\), we actually used the frequentist definition of probability. As a consequence, it means that \(n \rightarrow \infty\) (in practical situations \(n \gg 1\)). This limit is in physics called thermodynamic limit.
There are a few natural questions:
Does it mean that the entropy can be used only in the thermodynamic limit?
Does the entropy measure the uncertainty of a single particle in a large system or some kind of average probability over many particles?
(LLN & CLT)
N.B.: does anybody recognize what is \(H_G - H_B\)?
Resolution: what are the states?
Do we consider states of a single dice?
Do we consider states of a pair of dices?
etc.
Do we consider states of an n-tuple of dices?
...
Excercise activity for you: can you derive Gibbs entropy from considering the state space of n-tuples of dices?
Two related issues
1. Gibbs paradox
\(\Delta S = k N\ln 2\)
Resolution
1. Simply multiply entropy by
\(1/N!\) - due to "quantum" reasons (indistinguishability)
2. Swendsen approach
Two related issues
2. Additivity and Extensivity
(we will come back to it later)
Additivity: We have two independent systems \(A\) and \(B\) $$S(A,B) = S(A) + S(B)$$
Extensivity: We have a system of N particles, then
$$S(kN) = k \cdot S(N)$$
Summary
Foundations of Entropy I
By Jan Korbel
Foundations of Entropy I
- 256