High School Is Entropy Truly Undefined in Physical Systems?

  • Thread starter Thread starter lukephysics
  • Start date Start date
  • Tags Tags
    Entropy
Click For Summary
SUMMARY

Entropy in physical systems is fundamentally a measure of uncertainty regarding predictions made by an observer, rather than a strictly defined quantity. The discussion highlights that terms like "low entropy" in the context of the early universe lack clarity without specifying the observer and the predictions involved. The treatment of entropy varies between classical thermodynamics and statistical physics, with the latter defining it in relation to the phase-space distribution function. Key references include N. G. van Kampen's work on relativistic thermodynamics and the von Neumann-Shannon-Jaynes definition of entropy.

PREREQUISITES
  • Understanding of information theory concepts
  • Familiarity with classical thermodynamics and statistical physics
  • Knowledge of relativistic hydrodynamics
  • Basic grasp of quantum statistical mechanics
NEXT STEPS
  • Study N. G. van Kampen's paper on relativistic thermodynamics of moving systems
  • Explore the von Neumann-Shannon-Jaynes definition of entropy in quantum statistical physics
  • Research the implications of phase-space distribution functions in statistical mechanics
  • Investigate the relationship between entropy and information theory
USEFUL FOR

Physicists, researchers in thermodynamics, and students of information theory seeking to deepen their understanding of entropy in both classical and relativistic contexts.

lukephysics
Messages
60
Reaction score
15
TL;DR
why do they say things have entropy such as ‘the early universe has low entropy’ when they don’t specify who is the observer and what thing are they predicting?
I always got a bit confused when listening to podcasts about arrow of time and entropy in the universe. So I was reading more about information theory. I learned today that for physical systems entropy is not defined. All it means is how much uncertainty an observer has when making a prediction about something of particular interest to them.

So why do they say things have entropy such as ‘the early universe has low entropy’ when they don't say who is the observer and what thing are they predicting?

Another example is entropy in chemical reactions. Is that a different definition of entropy? Or is it fundamentally the same?
 
Last edited:
Physics news on Phys.org
Can you please say (exactly quote) the specific statements that you're examining?
 
Interesting question - maybe @vanhees71 has some insights here, being the resident expert on relativistic hydrodynamics. Planck proved in this paper that entropy is a Lorentz invariant scalar, but also regards the measured temperature as transforming as ##T \rightarrow T(1-v^2)^{-1/2}## between observers. On the other hand, it seems more natural to say that temperature is only defined in the rest frame of the body?
 
The relativistic treatment of thermodynamics before van Kampen is a mess, though I'm not sure whether van Kampen is really the first who introduced our modern view. A kind of review is

N. G. van Kampen, Relativistic thermodynamics of moving
systems, Phys. Rev. 173, 295 (1968),
https://doi.org/10.1103/PhysRev.173.295.

Today we use the definition, as given in Sect. 9 of this paper, that the thermodynamic quantities are defined in the (local) restframe of the medium. Entropy is a scalar quantity. The paper also gives examples for two historical treatments by Ott and Planck.

Another approach is of course statistical physics. There the key is that the phase-space distribution function is a scalar quantity. For a manifestly covariant treatment of elementary relativistic transport theory, see

https://itp.uni-frankfurt.de/~hees/publ/kolkata.pdf

Concerning the more general questions of the OP, it's clear that entropy in the information theoretical sense (and that seems to be the best and most comprehensive approach we have) is always a measure for the missing information (more intuitively, it's a measure for the "surprise" you have from a specific outcome of a random experiment), given some information about the system, relative to the case of complete information. E.g., in quantum statistical physics the entropy is always relative to the preparation of a pure state, i.e., any pure state has entropy 0, which leads to the von Neumann-Shannon-Jaynes definition of entropy,
$$S=-k_{\text{B}} \mathrm{Tr} (\hat{\rho} \ln \hat{\rho} ),$$
where ##\hat{\rho}## is the statistical operator, describing the state of the system.
 
Last edited:
  • Like
Likes ergospherical
In sci-fi when an author is talking about space travellers or describing the movement of galaxies they will say something like “movement in space only means anything in relation to another object”. Examples of this would be, a space ship moving away from earth at 100 km/s, or 2 galaxies moving towards each other at one light year per century. I think it would make it easier to describe movement in space if we had three axis that we all agree on and we used 0 km/s relative to the speed of...

Similar threads

  • · Replies 51 ·
2
Replies
51
Views
5K
  • · Replies 5 ·
Replies
5
Views
583
  • · Replies 6 ·
Replies
6
Views
1K
  • · Replies 7 ·
Replies
7
Views
1K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 4 ·
Replies
4
Views
1K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 17 ·
Replies
17
Views
3K