Foundations of Entropy II

 

 Let's calculate 'em

 

 

Lecture series at the 

School on Information, Noise, and Physics of Life

Nis 19.-30. September 2022

 

by Jan Korbel

all slides can be found at: slides.com/jankorbel

Activity II

 

You have 3 minutes to write down on a piece of paper:

             

             What is the most unexpected/surprising application

of entropy

             you have seen? (unexpected field, unexpected result)

 

 

 

 

Using Boltzmann's formula for non-multinomial systems

As we saw in the previous lecture, the multinomial multiplicity

$$W(n_1,\dots,n_k) = \binom{n}{n_1,\dots,n_k}$$

leads to Boltzmann-Gibbs-Shannon entropy.

 

 

Are there systems with non-multinomial multiplicity?

What is their entropy?

Ex. I: MB, FD & BE statistics

1. Maxwell-Boltzmann statistics - \(N\) distinguishable particles, \(N_i\) particles in state \(\epsilon_i\)

Multiplicity can be calculated as

$$W(N_1,\dots,N_k) = \binom{N}{N_1} \binom{N-N_1}{N_2} \dots \binom{N-\sum_{i=1}^{k-1}N_1}{N_k} = N! \prod_{i=1}^k \frac{1}{N_i!}$$

If \(\epsilon_i\) has degeneracy \(g_i\), then

$$W(N_1,\dots,N_k) = N! \prod_{i=1}^k \frac{g_i^{N_i}}{N_i!}$$

Then, we get

$$S_{MB} = - N \sum_{i=1}^k p_i \log \frac{p_i}{g_i}$$

Ex. I: MB, FD & BE statistics

2. Bose-Einstein statistics - \(N\) indistinguishable particles, \(N_i\) particles in state \(\epsilon_i\) with  degeneracy \(g_i\),

 

Multiplicity can be calculated as

$$W(N_1,\dots,N_k) = \prod_{i=1}^k \binom{N_i + g_i-1}{N_i}$$

Let us introduce \(\alpha_i = g_i/N\). Then, we get

$$S_{BE} = N \sum_{i=1}^k  \left[(\alpha_i + p_i) \log (\alpha_i +p_i) - \alpha_i \log \alpha_i - p_i \log p_i\right]$$

(|*)

3. Fermi-Dirac statistics - \(N\) indistinguishable particles, \(N_i\) particles in state \(\epsilon_i\) with  degeneracy \(g_i\), maximally 1 particle per sub-level (thus \(N_i \leq g_i\))

 

Multiplicity can be calculated as

$$W(N_1,\dots,N_k) = \prod_{i=1}^k \binom{g_i}{N_i}$$

Let us introduce \(\alpha_i = g_i/N\). Then, we get

$$S_{FD} = N \sum_{i=1}^k  \left[-(\alpha_i - p_i) \log (\alpha_i -p_i) + \alpha_i \log \alpha_i - p_i \log p_i\right]$$

Ex. I: MB, FD & BE statistics

Ex. II: structure-forming systems

Let us start with a simple example of a coin tossing.

States are:  

But! let's make a small change, we consider magnetic coins

The bound (or sticky) state is simply  

State space grows super-exponentially  (\(W(n) \sim n^n \sim e^{n \log n}\) )

picture taken from: Jensen et al 2018 J. Phys. A: Math. Theor. 51 375002

Multiplicity

2 x           1x

1 x          1x

3

3

Microstate

Mesostate

Mesostate

2 x           1x

1 x          1x

1    1   2   2    3   3

2   3   1   3    1    2

3   2   3   1    2    1

 

1    1   2   2    3   3

2   3   1   3    1    2

3   2   3   1    2    1

 

= (1,2,3) , (2,1,3)

= (1,3,2) , (3,1,2)

= (2,3,1) , (3,2,1)

= (1,2,3) , (1,3,2)

= (2,1,3) , (2,3,1)

= (3,1,2) , (3,2,1)

How to calculate multiplicity

General formula: \(W(n_i^{(j)}) = \frac{n!}{\prod_{ij} n_i^{(j)}!  {\color{red} (j!)^{n_i^{(j)}}}}\)  

we have \(n_i^{(j)}\) molecules of size \(j\) in a state \(s_i^{(j)}\)

Multiplicity

Boltzmann's 1884 paper

Entropy of structure-forming systems

$$ S =  \log W \approx n \log n - \sum_{ij} \left(n_i^{(j)} \log n_i^{(j)} - n_i^{(j)} + {\color{red} n_i^{(j)} \log j!}\right)$$

Introduce "probabilities" \(\wp_i^{(j)} = n_i^{(j)}/n\)

$$\mathcal{S} = S/n = - \sum_{ij} \wp_i^{(j)} (\log \wp_i^{(j)} {\color{red}- 1}) {\color{red}- \sum_{ij} \wp_i^{(j)}\log  \frac{j!}{n^{j-1}}}$$

Finite interaction range: concentration \(c = n/b\) 

$$\mathcal{S} = S/n = - \sum_{ij} \wp_i^{(j)} (\log \wp_i^{(j)} {\color{red}- 1}) {\color{red}- \sum_{ij} \wp_i^{(j)}\log  \frac{j!}{{\color{purple}c^{j-1}}}}$$

 

 Nat. Comm. 12 (2021) 1127

Ex. III: sample-space reducing processes (SSR)

 Corominas Murtra, Hanel, Thurner PNAS 112(17) (2015) 5348

Multiplicity

The number of states is \(n\). Let us denote the states as \(x_n \rightarrow \dots \rightarrow x_1\), where \(x_1\) is the ground state, where the process restarts. Let us sample \(R\) relaxation sequences \(x = (x_{k_1},\dots,x_1)\). 

The sequences can be visualised as

 

 

 

 

 

 

 

How many of these sequences contain  a state \(x_{j}\) exactly \(k_j\) times?  

Each run must contain \(x_1\)

Multiplicity

Number or runs \(R \equiv k_1\), number of them containing \(x_j\) is \(k_j\)

Multiplicity of these sequences: \(\binom{k_1}{k_j}\)

By multiplying the multiplicity for each state we get

$$W(k_1,\dots,k_n) = \prod_{j=2}^n \binom{k_1}{k_j}$$

$$ \log W \approx \sum_{j=2}^n \left[k_1 \log k_1 - \cancel{k_1} - k_j \log k_j + \cancel{k_j} - (k_1-k_j) \log (k_1 - k_j) + \cancel{(k_1-k_j)}\right]$$

$$\approx \sum_{j=2}^n \left[ k_1 \log k_1 \textcolor{red}{-k_j \log k_1} - k_j log \frac{k_j}{\textcolor{red}{k_1}} - (k_1-k_j) \log(k_1-k_j) \right]$$

By introducing \(p_i = k_i/N\) where \(N\) is the total number of steps, we get

$$S_{SSR}(p) = - N \sum_{j=2}^n \left[p_i \log \left(\frac{p_i}{p_1}\right) + (p_1-p_i) \log \left(1-\frac{p_i}{p_1}\right)\right]$$

Ex. IV: Pólya urns

Hanel, Corominas Murtra, Thurner New J. Phys. (2017) 19 033008

Probability of a sequence 

We have  \(c\) colors, initially \(n_i(0) \equiv n_i\) balls of color \(c_i\). After a ball is drawn, we return \(\delta\) balls of the same color to the urn.

After \(N\) draws, the number of balls in the urn is

$$n_i(N) = n_i + \delta k_i$$

where \(k_i\) is the number of draws of color \(c_i\). The total number of balls is \(n(R) = \sum_{c} n_c(N) = N + \delta N\) 

The probability of drawing a ball of color \(c_i\) in \(N\)-th run, is \(p_i(N) = n_i(N)/n(N)\). The probability of sequence

\(\ \mathcal{I} =\{i_1,\dots,i_N\} \) is 

$$p(\mathcal{I}) = \prod_{j=1}^c \frac{n_{j}^{(\delta,k_j)}}{n^{(\delta,N)}}$$

where \(m^{(\delta,r)} = m(m+\delta)\dots(m+r\delta)\) 

Probability of a histogram

A histogram \(\mathcal{K} = \{k_1,\dots,k_c\}\) is defined as \(k_c = \sum_{j = 1}^N \delta(i_j,c) \)

Thus the probability of observing a histogram is

$$ p(\mathcal{K}) =  \binom{N}{k_1,\dots,k_c} p(\mathcal{I}) $$ 

\(n_{j}^{(\delta,k_{j})} \approx k_{j}! \delta^{k_{j}}  (k_i+1)^{n_i/\delta} \) 

\( p(\mathcal{K}) = \frac{N!}{\prod_{j=1}^c \cancel{k_j!}} \frac{\prod_{j=1}^c \cancel{k_j!} \delta^{k_j} (k_j+1)^{n_j/\delta}}{n^{(\delta,N)}} \)

... technical calculation ...

 

$$S_{Pólya}(p) = \log p(\mathcal{K}) \approx - \sum_{i=1}^c \log(p_i + 1/N)$$

Ex. IV: q-deformations

This example is rather theoretical, but provides us a useful hint of what happens if there are correlations in the sample space

Motivation: finite versions of \(\exp\) and \(\log\)

\(\exp(x) = \lim_{n \rightarrow \infty} \left(1+\frac{x}{n}\right)^n\)

So define

\(\exp_q(x) := \left(1 + (1-q) x\right)^{1/(1-q)}\)

\(\log_q(x) := \frac{x^{1-q}-1}{1-q}\)

Let us find an operation s.t.

\(\exp_q(x) \otimes_q \exp_q(y) \equiv \exp_q(x+y)\)

\(\Rightarrow a \otimes_q b = \left[a^{1-q} + b^{1-q}-1\right]^{1/(1-q)}\)

Suyari, Physica A 368 (2006) 63-82

Calculus of q-deformations

In analogy to \(n! = 1 \cdot 2 \cdot \dots \cdot n\) introduce \(n!_q := 1 \otimes_q 2 \otimes_q  \dots \otimes_q n\)

It is than easy to show that \(\log_q n!_q = \frac{\sum_{k=1}^n k^{1-q} - n}{1-q}\) which can be used for generalized Stirling's approximation \(\log_q n!_q \approx \frac{n}{2-q} \log_q n\)

Let us now consider a q-deformed multinomial factor

$$\binom{n}{n_1,\dots,n_k}_q := n!_q \oslash_q (n_1!_q \otimes_q \dots \otimes_q n_k!_q) $$

$$ =  \left[\sum_{l=1}^n l^{1-q} - \sum_{i_1}^{n_1} i_i^{1-q} - \dots - \sum_{i_k}^{n_k} i_k^{1-q}\right]^{1/(1-q)}$$

 

Tsallis entropy

Let us consider that the multiplicity is given by a q-multinomial factor \( W(n_1,\dots,n_k) = \binom{n}{n_1,\dots,n_k}_q\)

 

In this case, it is more convenient to define entropy as \(S = \log_q W\), which gives us:

 

$$\log_q W = \frac{n^{2-q}}{2-q} \frac{\sum_{i=1}^k p_i^{2-q}-1}{q-1} = \frac{n^{2-q}}{2-q} S_{2-q}(p)$$

This entropy is known as Tsallis entropy

Note that the prefactor is not \(n\) but \(n^{2-q}\)

(non-extensivity) - we will discuss this later

Entropy and energy

 

Until now, we have been just counting states; let us now discuss the relation with energy

 

We consider that the states describe the energy of the system (either Hamiltonian or more generalized energy functional)

 

Therefore, entropy is defined as

$$ S(E) := \log W(E)$$

Ensembles

There are a few typical situations:

1. Isolated system = microcanonical ensemble

Let \(H(s)\) be the energy of a state \(s\). Multiplicity is then \(W(E) = \sum_{s} \delta(H(s) - E)\)

Phenomena like negative "temperature" \(T = \frac{\mathrm{d} S(E)}{\mathrm{d} E} < 0\)

2. closed system = canonical ensemble

Total system is composed of the system of interest (S) and the heat reservoir/bath (B). They are weakly coupled i.e., \(H_{tot}(s,b) = H_S(s) + H_B(b)\) (no interaction energy)

$$W(E_{tot}) = \sum_{s,b} \delta(H_{tot}(s,b) -E_{tot})$$

3. open system = grandcanonical ensemble

This can be further rewritten as 

$$= \int \mathrm{d} E_S\,  \sum_s \delta (H_{S}(s)-E_S) \sum_b(H_B(b)-(E_{tot}-E_S)) $$

$$= \int \mathrm{d} E_S \, W_S(E_S) W_B(E_{tot}-E_S) $$

Entropy in canonical ensemble

This is hard to calculate. Typically, the dominant contribution is from the maximal configuration of the integrand, which we obtain from 

$$\frac{\partial W_S(E_S) W_B(E_{tot}-E_S)}{\partial E_S} \stackrel{!}{=} 0 \Rightarrow \frac{W'_S(E_S)}{W_S(E)} = \frac{W'_B(E_{tot}-E_S)}{W_B(E_{tot}-E_S)}$$

As a consequence \(\frac{\partial S_E(E_S)}{\partial E_S} \stackrel{!}{=} \frac{\partial S_B(E_{tot}-E_S)}{\partial E_S} := \frac{1}{k_B T}\) 

and \(\underbrace{S_B(E_{tot} - E_S)}_{free \ entropy} = \underbrace{S_B(E_{tot})}_{bath \ entropy}  - \frac{\partial S}{\partial E_S} E_S + \dots\)

This is the emergence of Maximum entropy principle

Canonical ensemble & Coarse-graining

S

B

weak coupling

S

coarse-graining

Deterministic picture

Statistical

 picture

Summary

Foundations of Entropy II

By Jan Korbel

Foundations of Entropy II

  • 129