Mathematica Mathematical detail regarding Boltzman's H thm

  • Thread starter Thread starter quasar987
  • Start date Start date
  • Tags Tags
    Mathematical
AI Thread Summary
The discussion centers on the definition of H in Boltzmann's H theorem, specifically questioning whether H can be accurately described as the "mean value of ln P_r." The formula H = ∑ P_r ln P_r is identified as Shannon's information entropy, which connects to statistical mechanics through Gibbs' entropy S = -k ∑ P_r ln P_r. The confusion arises from the notation, as the correct interpretation involves summing over states rather than variables. While the professor defends the definition, it is suggested that conceptualizing H as a mean value may not be particularly insightful. Ultimately, the focus should be on entropy as a measure of the number of microstates corresponding to a macrostate, emphasizing its connection to maximum entropy in equilibrium.
quasar987
Science Advisor
Homework Helper
Gold Member
Messages
4,796
Reaction score
32
In the "proof" of the theorem, my course notes defines P_r(t) as the probability to find the system is state r at time t, and it defined H as the mean value of \ln P_r over all acesible states:

H \equiv \sum_r P_r\ln P_r

Is is right to call the above sum the "mean value of ln P_r" ?! Cause given a quantity u, the mean value of f(u) is defined as

\sum_i P(u_i)f(u_i)

So the mean value of \ln P_r should be

\sum_r P(P_r)\ln P_r

But P(P_r) does not make sense.I confessed my confusion to the professor in more vague terms (at the time, I only tought the equation looked suspicious), but he said there was nothing wrong with it. I say, H could be called at best "some kind" of mean value of ln(Pr).
 
Last edited:
Physics news on Phys.org
The formula you are asking about
H=SUM(P_r * ln(P_r))
is Shannon's definition of information entropy; it is related back to statistical mechanics by multiplying by the Boltzmann factor -k to give Gibbs' entropy
S=-k*SUM(P_r * ln(P_r)).
H is correctly called the mean of ln(P_r). Don't get confused by your notation--you must sum over states, not variables. P(u_i) is the probability of finding u in the ith state, so your first and second equations are actually the same.

You may prefer a less confusing way of writing the mean or expectation
E[f(x)] = SUM(P_r * f(x_r)),
as used in Jaynes, Phys. Rev. 106:620-630 (1957) (who discusses the connection between information H and stat mech S)
or Chandler, Intro to Modern Stat Mech, ch. 3 (1987).

Having defended the correctness of the definition, I have to add that it isn't a very useful way of thinking of H. Take a very simple case, that of an ideal gas, as an example. One sums r over W equally probable microstates (p=1/W) so the entropy of a macrostate of the gas system reduces to
H = -lnW;
multiplying by the constant -k gives exactly Boltzmann's entropy H=k*lnW. But how is thinking of H as the mean value of log of probability helpful or insightful? Instead, entropy reflects the number of possible ways W that a macrostate can be realized, and the second law ensures that the macrostate adopted in equilibrium is that which can be realized in the most number of ways. To tie it to information theory, this is the maximum entropy state.

Hope this helps.
 

Similar threads

Back
Top