Boltzmann Entropy Formula – Derivation

Click For Summary

Discussion Overview

The discussion revolves around the Boltzmann entropy formula, specifically its definition, derivation, and the meaning of the variable W in the context of statistical mechanics. Participants explore theoretical implications, definitions, and connections to information theory, as well as the distinctions between distinguishable and indistinguishable particles.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • Some participants define Boltzmann entropy as $$ S = k_B lnW $$, where W represents the number of microstates compatible with macroscopic parameters (E, V, N).
  • Others argue that the definition provided is not exact and seek clarification on the derivation and meaning of W.
  • One participant suggests that W is related to probability, specifically that it is proportional to the probability of a macro-state.
  • Another participant mentions that for indistinguishable particles, W is the number of ways a macro state can be realized, leading to extensive entropy.
  • Some participants discuss the implications of using the Boltzmann formula as a definition versus deriving it from other principles, such as the Shannon entropy formula.
  • There is a suggestion that the Boltzmann formula may not be derivable from other principles, while others propose that it can be connected to information theory.
  • Concerns are raised about the non-extensiveness of entropy for distinguishable particles when applying Boltzmann's formula.
  • Participants explore the relationship between W and N in the context of the Shannon formula, with some concluding that they are equivalent.

Areas of Agreement / Disagreement

Participants express differing views on the derivation of the Boltzmann entropy formula and the interpretation of W. There is no consensus on whether the formula can be derived from other principles or if it stands as a definition. The discussion remains unresolved regarding the implications of distinguishable versus indistinguishable particles on entropy.

Contextual Notes

Participants note that the definition and derivation of entropy may depend on specific assumptions and contexts, such as the treatment of particles as distinguishable or indistinguishable. The discussion highlights the complexity of relating statistical mechanics to thermodynamic principles.

Dario56
Messages
289
Reaction score
48
Boltzmann entropy definition is given by: $$ S = k_B lnW $$ where ##W## is the weight of the configuration which has the maximum number of microstates.

This equation is used everywhere in statistical thermodynamics and I saw it in the derivation of Gibbs entropy. However, I can't find the derivation of this equation or where does it come from. Do you know a source?
 
  • Like
Likes   Reactions: Delta2
Science news on Phys.org
Your definition is not exact. In statistical mechanics, one defines the Boltzmann entropy as ##S=k_BlnW##, where ##W## is the number of microstates compatible with some macroscopic parameters ##(E,V,N)##.
 
  • Like
Likes   Reactions: dextercioby
Lord Jestocost said:
Your definition is not exact. In statistical mechanics, one defines the Boltzmann entropy as ##S=k_BlnW##, where ##W## is the number of microstates compatible with some macroscopic parameters ##(E,V,N)##.
You know maybe where can derivation of this equation be found? I can't find it anywhere.
 
Lord Jestocost said:
On http://www.energyandentropy.com/page/index.html, Harvey S. Leff shortly sketches the thinking behind the definition of Boltzmann’s entropy.
So it is actually a definition, it cannot be proved by other principles/theorems?
 
  • Like
Likes   Reactions: Philip Koeck
Delta2 said:
So it is actually a definition, it cannot be proved by other principles/theorems?
I would also like to find out more about that.
It seems clear that entropy is proportional to some probability or weight W, but what exactly does W stand for and why?
 
  • Like
Likes   Reactions: Delta2
I believe Boltzmann was thinking of probability when he chose the letter W (Wahrscheinlichkeit), but clearly it's not a number between 0 and 1.
It's more like something proportional to the probability of a macro-state.
As far as I know it's simple for indistinguishable particles: W is just the number of ways a macro state can be realized. Entropy defined like that becomes an extensive quantity.

Robert Swendsen, mentioned in the article shared by Lord Jestocost, also wants an extensive definition of statistical entropy for distinguishable particles. He defines entropy as S = kB ln ( W f(N) ) in order to make it extensive.
For example f(N) could be 1/N! Other researchers do the same.

I believe the only justification is that statistical entropy has the same properties as thermodynamic entropy. I don't know of any derivation from something else.
 
  • Like
Likes   Reactions: Dario56
I think the most intuitive definition of entropy is via information theory based on Shannon's definition of the measure of information. At least for me it was a revelation, when I attended a special lecture about statistical physics from an information-theoretical point of view. A good testbook is

A. Katz, Principles of Statistical Mechanics, W. H. Freeman and Company, San Francisco and London (1967).

The point is that thermodynamics can be derived from the underlying microscopic physics using statistical methods. It may sound a bit strange, but one should be aware that from a conceptional point of view quantum statistical physics is simpler than classical statistical physics. The latter can be easily derived from the former as an appropriate approximation for situations, where this approximation is justified (i.e., occupation numbers for each microstate that are much smaller than 1 on average; e.g., for large temperatures and small densities in the equilibrium case).
 
  • Like
  • Informative
Likes   Reactions: dextercioby, Philip Koeck, hutchphd and 1 other person
vanhees71 said:
... quantum statistical physics is simpler than classical statistical physics. The latter can be easily derived from the former as an appropriate approximation for situations, where this approximation is justified (i.e., occupation numbers for each microstate that are much smaller than 1 on average; e.g., for large temperatures and small densities in the equilibrium case).
Do you mean "... occupation numbers for each quantum state (or single particle state) that are much smaller than 1 on average ..."?

I'm not sure I can see how low occupation turns quantum particles into classical ones.
For example photons in a cavity should remain indistinguishable and non-classical even at low occupancy, or not?

If we apply the low occupancy limit to the expression for W we don't get the same result for distinguishable and indistinguishable particles either. The results differ by a factor N!.
(See https://www.researchgate.net/publication/330675047_An_introduction_to_the_statistical_physics_of_distinguishable_and_indistinguishable_particles_in_an_ideal_gas)
 
  • #10
Take an ideal For bosons (upper sign) and fermions (lower sign) you have
$$f=\frac{1}{\exp[\vec{p}^2/(2mT)-\mu/T] \mp 1}.$$
If the exponential is much larger than ##1##, i.e., if ##\mu<0## with ##|\mu|/T \gg 1## (dilute limit) you can neglect the ##\pm 1## and get the classical Maxwell-Boltzmann distribution.

The ##1/N!## is a feature, not a bug. We had this discussion already some time ago in this forum!
 
  • #11
vanhees71 said:
Take an ideal For bosons (upper sign) and fermions (lower sign) you have
$$f=\frac{1}{\exp[\vec{p}^2/(2mT)-\mu/T] \mp 1}.$$
If the exponential is much larger than ##1##, i.e., if ##\mu<0## with ##|\mu|/T \gg 1## (dilute limit) you can neglect the ##\pm 1## and get the classical Maxwell-Boltzmann distribution.

The ##1/N!## is a feature, not a bug. We had this discussion already some time ago in this forum!
No problem with that.
The original question was how one can motivate S = k ln(W) and I would say a major part of this question is what W stands for.
If W is the number of microstates for a given macrostate then there is a factor N! difference between distinguishable and indistinguishable particles in the low occupancy limit.
This means that S becomes non-extensive for distinguishable particles.

I can see several possible solutions for this, for example:
There are no distinguishable particles.
W is not the number of microstates.
The Boltzmann-formula has to be modified.
Entropy doesn't need to be extensive.
 
Last edited:
  • #12
Delta2 said:
So it is actually a definition, it cannot be proved by other principles/theorems?
Well, yes and no. To derive a formula for entropy, one must first define entropy somehow. Often it is convenient to start from the Boltzmann formula as a definition, but it's not necessary. For example, one can start from the Shannon formula
$$S=-\sum_{i=1}^N p_i \ln p_i$$
as a definition and then obtain the Boltzmann formula as follows. A priori all the probabilities ##p_i## are the same, so from the constraint ##\sum_{i=1}^N p_i=1## it follows that ##p_i=1/N##. Inserting this into the Shannon formula, one gets
$$S=\ln N$$
which is almost the Boltzmann formula. To get the Boltzmann formula one has to change the units, i.e. redefine entropy as ##S=k\ln N##, and define ##W=N##.
 
  • Like
Likes   Reactions: vanhees71
  • #13
Demystifier said:
Well, yes and no. To derive a formula for entropy, one must first define entropy somehow. Often it is convenient to start from the Boltzmann formula as a definition, but it's not necessary. For example, one can start from the Shannon formula
$$S=-\sum_{i=1}^N p_i \ln p_i$$
as a definition and then obtain the Boltzmann formula as follows. A priori all the probabilities ##p_i## are the same, so from the constraint ##\sum_{i=1}^N p_i=1## it follows that ##p_i=1/N##. Inserting this into the Shannon formula, one gets
$$S=\ln N$$
which is almost the Boltzmann formula. To get the Boltzmann formula one has to change the units, i.e. redefine entropy as ##S=k\ln N##, and define ##W=N##.
What does N and pi stand for in Shannon's formula?
 
  • #14
Philip Koeck said:
What does N and pi stand for in Shannon's formula?
##N## is the number of different microscopic states, ##p_i## are the probabilities of those states.
 
  • Like
Likes   Reactions: Philip Koeck
  • #15
Demystifier said:
##N## is the number of different microscopic states, ##p_i## are the probabilities of those states.
So N was really W to start with?
 
  • #16
Philip Koeck said:
So N was really W to start with?
Yes.
 
  • Like
Likes   Reactions: Philip Koeck
  • #17
If you apply Boltzmann's formula to an ideal gas of distinguishable particles you get an entropy that's not extensive, I believe.

A quick explanation: For an ideal gas of indistinguishable particles you get an extensive entropy.
At low occupancy Wd = n! Wi, where Wd is the number of microstates for distinguishable and Wi the same for indistinguishable particles. n is the number of particles.

This means that S can't be extensive for an ideal gas of distinguishable particles if S = k ln Wd.

Do you have a solution to this problem? (Or is it a problem?)
 
  • #18
Philip Koeck said:
If you apply Boltzmann's formula to an ideal gas of distinguishable particles you get an entropy that's not extensive, I believe.
I don't.
Philip Koeck said:
A quick explanation: For an ideal gas of indistinguishable particles you get an extensive entropy.
At low occupancy Wd = n! Wi, where Wd is the number of microstates for distinguishable and Wi the same for indistinguishable particles. n is the number of particles.
So far so good.
Philip Koeck said:
This means that S can't be extensive for an ideal gas of distinguishable particles if S = k ln Wd.
By your formula the entropy increases with ##n##, which implies that it is "extensive" in the sense that it is not intensive. Perhaps by "not extensive" you mean that it does not grow linearly with ##n##. However, by Stirling formula, ##\ln n!## grows approximately linearly with ##n##. For large ##n## we have the Stirling approximation
$$\ln n! = n \ln n -n$$
so the derivative over ##n## of this is ##\ln n##, which is a slowly growing function so can be approximated by a constant. For example, when ##n## doubles then ##\ln n## increases by ##\ln 2 =0.69##, which is a totally negligible number compared to ##n## when ##n## is large.
 
  • Like
Likes   Reactions: Philip Koeck
  • #19
Demystifier said:
I don't.

So far so good.

By your formula the entropy increases with ##n##, which implies that it is "extensive" in the sense that it is not intensive. Perhaps by "not extensive" you mean that it does not grow linearly with ##n##. However, by Stirling formula, ##\ln n!## grows approximately linearly with ##n##. For large ##n## we have the Stirling approximation
$$\ln n! = n \ln n -n$$
so the derivative over ##n## of this is ##\ln n##, which is a slowly growing function so can be approximated by a constant. For example, when ##n## doubles then ##\ln n## increases by ##\ln 2 =0.69##, which is a totally negligible number compared to ##n## when ##n## is large.
Thanks! That's a really helpful answer.
So the only thing that changes is that S gets an offset that is roughly proportional to n.
In total S remains roughly proportional to n (that's what I meant by extensive) even if you decide to apply Boltzmann's formula to distinguishable particles (aerosols maybe?).
 
  • Like
Likes   Reactions: vanhees71 and Demystifier

Similar threads

Replies
4
Views
1K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 22 ·
Replies
22
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 6 ·
Replies
6
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K