2nd law of thermodynamics -- why?

Click For Summary

Discussion Overview

The discussion revolves around the Second Law of Thermodynamics, specifically exploring its origins, implications, and the concept of entropy. Participants delve into theoretical aspects, empirical findings, and the relationship between entropy and knowledge about a system.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • Some participants describe the Second Law as an empirical finding accepted as an axiom of thermodynamics, linked to the irreversibility of thermodynamic processes.
  • One participant proposes that irreversibility arises from the asymmetric character of thermodynamic operations.
  • A participant uses the example of dice to illustrate entropy as a measure of knowledge about a system, suggesting that entropy increases as knowledge decreases.
  • Another participant argues that from a microscopic perspective, the Second Law relates to the probability of a system being in a certain state, emphasizing that while heat could theoretically flow from cold to hot, the probability is extremely low.
  • There is a discussion about whether entropy is a property of the observer or the system, with some asserting it is a property of the system based on the number of microstates associated with a macrostate.
  • Participants explore the implications of knowing the initial state of a system on its entropy, questioning the relationship between knowledge and entropy.
  • Irreversibility is attributed to various forms of dissipation, including mechanical energy, temperature gradients, concentration gradients, and chemical potential gradients.
  • One participant raises a question about how to determine the probability of heat flowing from a cold body to a hot one, leading to a discussion about the multiplicity of macrostates.
  • Another participant introduces the concept of pure and mixed states in relation to quantum mechanics and entropy, suggesting that the understanding of entropy may depend on the interpretation of quantum mechanics.

Areas of Agreement / Disagreement

Participants express differing views on the nature of entropy, whether it is an observer-dependent property or a property of the system itself. The discussion remains unresolved regarding the relationship between knowledge and entropy, as well as the implications of quantum mechanics on the understanding of entropy.

Contextual Notes

Participants acknowledge limitations in their understanding of the relationship between entropy and knowledge, as well as the complexities introduced by quantum mechanics. There is also a recognition of the need for conventions in defining state variables in classical thermodynamics.

Einstein's Cat
Messages
182
Reaction score
2
I am aware of the Second Law of Theromodynamics and I understand it to a certain extent, although still I am burdened with the frustration of being ignorant to why such a law exists. Please assist me.
 
Science news on Phys.org
It is an empirical finding and it is accepted as an axiom of thermodynamics. Its microscopic origin is explained by statistical thermodynamics. It basically has to do with the irreversibility of a thermodynamical process, something observed in our macroscopic reality.
 
QuantumQuest said:
It is an empirical finding and it is accepted as an axiom of thermodynamics. Its microscopic origin is explained by statistical thermodynamics. It basically has to do with the irreversibility of a thermodynamical process, something observed in our macroscopic reality.
Why does this irreversibility exist?
 
Einstein's Cat said:
Why does this irreversibility exist?

Because of the asymmetric character of thermodynamic operations.
 
Consider a box of six sided dice. Initially, they all have "1" facing up. Now shake the box very gently, so dice only have a small probability of flipping over time. As time passes, the dice look more and more randomly distributed. Entropy can be thought of as your lack of knowledge about a system. If you know the initial state exactly, then the initial entropy is 0. As time passes, you know less and less about the state of the dice, and entropy increases. You never regain knowledge you have lost unless you make a measurement (which means you don't have a closed system). This is at the heart of the second law.
 
From the microscopic point of view, the 2nd law is a combination of the basic statistical fact that the higher the probability of a system being in a state, the more probable it is the system will be found in that state when measured, with the fact that thermodynamic systems are so huge that the probability distribution is extremely narrowly peaked around a given state, to the point that any fluctuation away from that state is too small to measure.

Heat could flow from a cold body to a hot one, but the probability of that happening is so small that in practice, it will never be observed.
 
Khashishi said:
If you know the initial state exactly, then the initial entropy is 0. As time passes, you know less and less about the state of the dice, and entropy increases.
So in this perspective, entropy is a property of an observer? For example, at any moment I could look inside and know the exact state, and hence claim the entropy is zero, but for you the entropy would still be nonzero. So the entropy is "of me" or "of you" and not "of the dice." But in thermodynamics entropy is presented as a state variable, isn't it? The entropy is "of the state of the system" and not "of me" or "of you."

I've yet to really understand entropy; I'm curious how these two perspectives of entropy (informational and thermodynamic) fit together.
 
Nathanael said:
So in this perspective, entropy is a property of an observer?

No, it is a property of the system. In statistical thermodynamics the entropy of a macro-state depends on the logarithm of the number of the possible micro-states. In Khashishi's example the initial macro-state is "1" facing up for all dices. As this macro-state has only a single micro-state it's entropy is zero.
 
  • Like
Likes   Reactions: Nathanael
DrStupid said:
No, it is a property of the system. In statistical thermodynamics the entropy of a macro-state depends on the logarithm of the number of the possible micro-states. In Khashishi's example the initial macro-state is "1" facing up for all dices. As this macro-state has only a single micro-state it's entropy is zero.
So would you disagree with Khashishi in saying "If you know the initial state exactly, then the initial entropy is 0." ?
In other words, you are saying the entropy is zero because there is only 1 possible configuration which gives all 1s, whereas Khashishi said the entropy is zero because we know exactly what all the dice are.
 
  • #10
Nathanael said:
whereas Khashishi said the entropy is zero because we know exactly what all the dice are.

He didn't say "because" but "if".
 
  • #11
Khashishi said:
Entropy can be thought of as your lack of knowledge about a system. If you know the initial state exactly, then the initial entropy is 0.
This quote gives me the impression that it was a more general statement, but perhaps it was meant only for the specific example of all 1s.

At any rate, I appreciate your clarification of entropy as the logarithm of the number of micro states which produce a certain macro state.
 
  • #12
  • Like
Likes   Reactions: Nathanael
  • #13
Einstein's Cat said:
Why does this irreversibility exist?
Irreversibility is the result of (a) viscous dissipation of mechanical energy to internal energy, associated with rapid deformation of a material (b) dissipation of temperature gradients, associated with rapid (spontaneous) conduction of heat in a material, (c) dissipation of concentration gradients, associated with (spontaneous) diffusional transport and mixing of chemical species in a material, and (d) dissipation of chemical potential driving force resulting from (spontaneous) chemical reactions in a system.
 
  • Like
Likes   Reactions: QuantumQuest
  • #14
DrClaude said:
Heat could flow from a cold body to a hot one, but the probability of that happening is so small that in practice, it will never be observed.
How would one determine such a probability?
 
  • #15
Einstein's Cat said:
How would one determine such a probability?
By considering the multiplicity of each macrostate (i.e., the number of microstates associate with each macrostate).
 
  • #16
Actually, when I say "you know it to be in this state", what I mean is that the system is in a pure state. It's not necessary for a conscious observer to exist. When I say, "you don't know the state", what I mean is that the system is in a mixed state*. Presumably, the quantum state of a system is an objective statement, but frankly, we are getting into territory we don't know how to experimentally test, and the answer depends on what interpretation of quantum mechanics you take. (For example, in the many worlds hypothesis, entropy will depend on what apparent "world" you are in.)

When we talk about entropy in classical mechanics and chemistry, we use certain conventions to agree on what the state variables of the system are (i.e. pressure, volume, number of particles, temperature), and we assume the exact position/momentum of every particle is in a heavily mixed state (since it is thoroughly entangled with the surrounding environment). With these conventions, we can talk objectively about the entropy of a classical system.

*a mixed state can be subjective or objective -- the math is the same. It can mean that we don't know the state, so it is in a mix of possibilities, or it can mean that the system is actually entangled with a larger system, so we can't express the complete state of the system within the confines of the smaller system. Macroscopic systems are always entangled with a larger environment, and it is not possible to fully describe the state of any macroscopic system without detailing the state of the entire universe.
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 33 ·
2
Replies
33
Views
3K
  • · Replies 100 ·
4
Replies
100
Views
9K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 20 ·
Replies
20
Views
3K
  • · Replies 13 ·
Replies
13
Views
3K
  • · Replies 65 ·
3
Replies
65
Views
10K
  • · Replies 7 ·
Replies
7
Views
16K
  • · Replies 6 ·
Replies
6
Views
2K