Is Entropy Truly Undefined in Physical Systems?

  • Context: High School 
  • Thread starter Thread starter lukephysics
  • Start date Start date
  • Tags Tags
    Entropy
Click For Summary

Discussion Overview

The discussion revolves around the concept of entropy in physical systems, particularly in relation to the arrow of time and information theory. Participants explore the definitions and interpretations of entropy, questioning its applicability and meaning in various contexts such as the early universe and chemical reactions.

Discussion Character

  • Exploratory
  • Conceptual clarification
  • Debate/contested
  • Technical explanation

Main Points Raised

  • One participant expresses confusion about the definition of entropy, suggesting it is not well-defined in physical systems and relates to an observer's uncertainty in predictions.
  • Another participant requests specific statements to clarify the discussion's focus.
  • A participant references Planck's work, noting that entropy is a Lorentz invariant scalar but questions the definition of temperature across different observers' frames.
  • There is mention of historical treatments of relativistic thermodynamics and the definition of thermodynamic quantities in the local rest frame of the medium, emphasizing that entropy is a scalar quantity.
  • Another participant introduces the statistical physics perspective, suggesting that entropy measures missing information relative to complete information, with specific reference to quantum statistical physics and the von Neumann-Shannon-Jaynes definition of entropy.

Areas of Agreement / Disagreement

Participants do not reach a consensus on the definition and implications of entropy, with multiple competing views and interpretations presented throughout the discussion.

Contextual Notes

There are unresolved questions regarding the definitions of entropy in different contexts, such as information theory versus thermodynamics, and the implications of observer-dependent measurements in relativistic frameworks.

lukephysics
Messages
60
Reaction score
15
TL;DR
why do they say things have entropy such as ‘the early universe has low entropy’ when they don’t specify who is the observer and what thing are they predicting?
I always got a bit confused when listening to podcasts about arrow of time and entropy in the universe. So I was reading more about information theory. I learned today that for physical systems entropy is not defined. All it means is how much uncertainty an observer has when making a prediction about something of particular interest to them.

So why do they say things have entropy such as ‘the early universe has low entropy’ when they don't say who is the observer and what thing are they predicting?

Another example is entropy in chemical reactions. Is that a different definition of entropy? Or is it fundamentally the same?
 
Last edited:
Science news on Phys.org
Can you please say (exactly quote) the specific statements that you're examining?
 
Interesting question - maybe @vanhees71 has some insights here, being the resident expert on relativistic hydrodynamics. Planck proved in this paper that entropy is a Lorentz invariant scalar, but also regards the measured temperature as transforming as ##T \rightarrow T(1-v^2)^{-1/2}## between observers. On the other hand, it seems more natural to say that temperature is only defined in the rest frame of the body?
 
  • Like
Likes   Reactions: vanhees71
The relativistic treatment of thermodynamics before van Kampen is a mess, though I'm not sure whether van Kampen is really the first who introduced our modern view. A kind of review is

N. G. van Kampen, Relativistic thermodynamics of moving
systems, Phys. Rev. 173, 295 (1968),
https://doi.org/10.1103/PhysRev.173.295.

Today we use the definition, as given in Sect. 9 of this paper, that the thermodynamic quantities are defined in the (local) restframe of the medium. Entropy is a scalar quantity. The paper also gives examples for two historical treatments by Ott and Planck.

Another approach is of course statistical physics. There the key is that the phase-space distribution function is a scalar quantity. For a manifestly covariant treatment of elementary relativistic transport theory, see

https://itp.uni-frankfurt.de/~hees/publ/kolkata.pdf

Concerning the more general questions of the OP, it's clear that entropy in the information theoretical sense (and that seems to be the best and most comprehensive approach we have) is always a measure for the missing information (more intuitively, it's a measure for the "surprise" you have from a specific outcome of a random experiment), given some information about the system, relative to the case of complete information. E.g., in quantum statistical physics the entropy is always relative to the preparation of a pure state, i.e., any pure state has entropy 0, which leads to the von Neumann-Shannon-Jaynes definition of entropy,
$$S=-k_{\text{B}} \mathrm{Tr} (\hat{\rho} \ln \hat{\rho} ),$$
where ##\hat{\rho}## is the statistical operator, describing the state of the system.
 
Last edited:
  • Like
Likes   Reactions: ergospherical

Similar threads

  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 1 ·
Replies
1
Views
944
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 5 ·
Replies
5
Views
736
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K