I Reading course in statistical physics

Click For Summary
The online reading course on "Statistical Physics" by Reif has opened discussions specifically about chapter 3, focusing on the quantum and classical descriptions of an ideal gas. It highlights that while quantum mechanics requires 3N quantum numbers for N particles, classical mechanics needs 6N numbers (3N positions and 3N momenta). The conversation also delves into the nature of indistinguishable particles, emphasizing that for bosons, the state is defined by occupation numbers, while for fermions, antisymmetrization applies. A key point of discussion is the equivalence of equilibrium states and uniform probability distributions, with participants debating the probabilities of microstates for indistinguishable particles. The thread continues to explore the implications of these concepts in statistical mechanics and their mathematical formulations.
  • #31
vanhees71 said:
From this
$$Y' \partial_x \sigma = X' \partial_y \sigma,$$
and this is a function of ##(x,y)##, which I can write in the Form ##X' Y'/\tau##. This shows that for functions with 2 independent variables for any inexact differential there's always an integrating factor, ##\tau##, such that
$$\mathrm{d} G/\tau=\mathrm{d} \sigma$$
is an exact differential.
Got it. The two expressions are the same so I can replace both with a 3rd expression.
The integration factor τ, which depends on x and y in general, is used to make the replacement correct. Something like that.

It will be interesting to read in the next chapter how this is applied to dS = d-Q / T.
(Couldn't find a crossed-out d.)
Somehow one also has to realize that this equation is only valid for a d-Q in a reversible process.
 
  • Like
Likes vanhees71
Science news on Phys.org
  • #32
I have a rather general question about the definition of entropy used in chapter 5:
S = k ln Ω, where Ω is the number of available microstates.
Boltzmann wrote W rather than Ω, and I believe this stood for probability (Wahrscheinlichkeit).
Obviously this is not a number between 0 and 1, so it's more like something proportional to probability.

Probability would be number of available microstates divided by total number of microstates (including those that are not available).
Now for distinguishable particles both these numbers are bigger than for indistinguishable particles, by a factor N!, where N is the number of particles, in the case of low occupancy.

Would it make sense therefore to use the following definition of entropy for distinguishable particles to make sure that this "probability" W is calculated correctly?
S = k ln (Ω / N!) for distinguishable particles at low occupancy.
 
  • #33
Philip Koeck said:
I have a rather general question about the definition of entropy used in chapter 5:
S = k ln Ω, where Ω is the number of available microstates.
Boltzmann wrote W rather than Ω, and I believe this stood for probability (Wahrscheinlichkeit).
Obviously this is not a number between 0 and 1, so it's more like something proportional to probability.

Probability would be number of available microstates divided by total number of microstates (including those that are not available).
Now for distinguishable particles both these numbers are bigger than for indistinguishable particles, by a factor N!, where N is the number of particles, in the case of low occupancy.

Would it make sense therefore to use the following definition of entropy for distinguishable particles to make sure that this "probability" W is calculated correctly?
S = k ln (Ω / N!) for distinguishable particles at low occupancy.
Maybe I should move this to a new post.
 

Similar threads

  • · Replies 29 ·
Replies
29
Views
3K
  • · Replies 15 ·
Replies
15
Views
1K
  • · Replies 2 ·
Replies
2
Views
562
Replies
0
Views
1K
Replies
4
Views
445
  • · Replies 6 ·
Replies
6
Views
2K
Replies
4
Views
324
  • · Replies 16 ·
Replies
16
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
Replies
5
Views
4K