Development of 2nd Law Of Thermodynamics

Click For Summary
The discussion revolves around the development and understanding of the Second Law of Thermodynamics, particularly its qualitative and quantitative aspects. The original qualitative postulates, such as the Kelvin-Planck and Clausius statements, are contrasted with the later quantitative formulations, including the efficiency of the Carnot cycle. The thread raises a critical question about the experimental validation of the Second Law, specifically whether there was a rigorous demonstration of its efficiency equation, eff = 1 - TL/TH, akin to Joule's demonstration of the First Law. Participants note that while Carnot made early measurements, the lack of precise experimental methods in the 19th century limited rigorous validation of the Second Law. The conversation highlights the evolution of thermodynamic principles and the ongoing exploration of their foundational experiments.
  • #31
wcg1989 said:
<snip>

Maybe his math was poor but I'm understanding better why Carnot was held in such high regard as one of the fathers of thermodynamics.

Excellent!
 
Science news on Phys.org
  • #32
Andy Resnick said:
This is anti-science.

Putting forward scientific arguments is not "anti-science". Anyway, just tell me how you got rid of the statistical fluctuations in your energy flows and other quantities?

So, basically my point is that the statistical foundations are simply hidden in the assumptions you make. Not elaborating on those assumptions does not mean that the statistical foundations aren't there, even if scientists haven't worked out yet how to do this rigorously.
 
  • #33
Count Iblis, classical thermodynamics - about which we are talking here - always refers to the limit of the system size to become infinite (at fixed density), where fluctuations become irrelevant.
 
  • #34
DrDu said:
Count Iblis, classical thermodynamics - about which we are talking here - always refers to the limit of the system size to become infinite (at fixed density), where fluctuations become irrelevant.

Yes, I largely agree, but the transport phenomena which Andy no doubt does want to describe in his thermodynamic treatment do depend on the fluctuations. What happens here is that one considers a deviation from global thermal equilibrium and then the fluctuations are essential to understand how the system will tend to restore global thermal equilibrium.
 
  • #35
Count Iblis said:
Putting forward scientific arguments is not "anti-science". Anyway, just tell me how you got rid of the statistical fluctuations in your energy flows and other quantities?

So, basically my point is that the statistical foundations are simply hidden in the assumptions you make. Not elaborating on those assumptions does not mean that the statistical foundations aren't there, even if scientists haven't worked out yet how to do this rigorously.

Your 'arguments' are not anywhere close to scientific, they are a logorrhea of random rhetorical statements.

But fine, now you actually bring up something concrete. Whence fluctuations?

Transport is handled in continuum mechanics by means of balance equations. Brenner's text "Macrotransport Processes" is an excellent starting point, and Slattery's "Interfacial transport phenomena" may even be better. But how do we model fluctuations in the continuum picture?

I'll first note that the fluctuations are occurring *about equilibrium values*. Fluctuations cannot be defined without equilibrium.

Answer: By means of the fluctuation-dissipation theorem. I'm not sure what fluctuations mean in a fully dynamic situation. But if you want to restrict yourself to near-equilibrium conditions, go for it.
 
  • #36
I really enjoyed the part of Carnot's paper that made an analogy with the Paddle wheel. (I can wrap my engineer brain around that one!) For the heck of it, I did my own calculations showing the analogy. No integrals or statistical equations! Just some really basic undergraduate stuff.

The 2nd law says the maximum efficiency possible of a heat engine is

eff = 1 - TL/TH

where TH is the high temperature of the heat going into the engine and TL is the low temperature. And eff is defined as

W = Eff * Qin

where W is the useful work and Qin is the heat energy going into the process also called QH.

This is analogous to how a paddle wheel works. The maximum possible work rate (power) from a flow over a paddle wheel is defined by the potential energy

W = mgh

where W is a work rate and m is the mass flow rate. That’s the maximum possible. Friction in the wheel causes the real work to be less but this is the maximum possible.

But look at h. If we think of it as over sea level and not just the height of the wheel we can form this another way and think about the maximum potential work of the paddle wheel vs the potential work of the wheel if the wheel extended all the way down to sea level. For this example assume sea level is as absolutely low as one could possibly go, no holes in the ground!. Then we can say

W = mghH - mghL

where hH is the top of the wheel and hL is the bottom of the wheel and both are defined as the height over sea level.

If hL is 0 or sea level then the equation above would be the maximum amount of work possible based on the initial height (hH) of the water going into the wheel.

I can rewrite the equation above through algebra

W = mghH * (1-mghL/mghH)

or

W = mghH * (1-hL/hH)

And if I defined best efficiency possible as (1-hL/hH) then

W = eff * mghH

And that is the same as the thermal efficiency through analogy! The potential energy of the flow going into the wheel mghH is analogous to the thermal energy going in the system, QH. The best efficiency based on height difference (1-hL/hH) is analogous to the best efficiency possible through temperature difference, 1-TL/TH

And this makes sense too. The objects in the room around us have a tremendous amount of thermal energy in them. But they are all about the same temperature. If there is no temperature difference, no where for the thermal energy in the room to "flow down" then no work can be made. And that's the analogy.
 
  • #37
That's really excellent- mind if I use it in my Physics I class next semester?
 
  • #38
Thanks for the compliment!

I think it would be cool if that was taught as a lesson in a physics class. This stuff is more interesting then the daily grind of being an engineer. But I suppose that is what I'm suppose to be doing right now! Gotta go.
 
  • #39
Andy Resnick,

Thank you for this discussion. I am certain that blis is wrong and you are correct. It is not often that such diverse views are openly debated about something so fundamental as Thermodynamic Entropy. The difference that you point out between Clausius' definition and alternative definitions that followed is, I think, very important to fundamental scientific learning. The difference between the obvious inclusion of time in Clausius' definition and the leap to other definitions that are static is critical to fundamental understanding. I think it demonstrates that we do not yet know what Thermodynamic Entropy is. I think that understanding the role of 'time' in Clausius' definition is important for explaining why Thermodynamic Entropy is a process that takes time. The question, as I see it, is why does Thermodynamic Entropy require time? I would appreciate reading more of your input even if what I have said does not conform to it.

James Putnam
 
  • #40
Well I don't think there are any real disagreements about the physics. From a fundamental POV, the fluctuations cause the transport phenomena, but you can describe these phenomena in an effective way.
 
Last edited:
  • #41
Iblis,

Sorry for missing the 'I' in your name.

"..From a fundamental POV, the fluctuations cause the transport phenomena, but you can describe these phenomena in an effective way."

How does static produce fluctuations?" Clausius did not require fluctuations. In the ideal situation that he relied upon there were no fluctuations. Is this incorrect?

James
 
  • #42
Andy Resnick said:
<snip>

Transport is handled in continuum mechanics by means of balance equations. Brenner's text "Macrotransport Processes" is an excellent starting point, and Slattery's "Interfacial transport phenomena" may even be better. But how do we model fluctuations in the continuum picture?

I'll first note that the fluctuations are occurring *about equilibrium values*. Fluctuations cannot be defined without equilibrium.

Answer: By means of the fluctuation-dissipation theorem. I'm not sure what fluctuations mean in a fully dynamic situation. But if you want to restrict yourself to near-equilibrium conditions, go for it.


I was going to edit this, but since there have already been comments, I'll just do it this way:

It makes no sense to speak of fluctuations in a fully dynamical theory. What you are asking for is to first, remove the dynamics and then add them back in, in such a manner that requires statistical notions. In which case you can start with the Langevin equation or Smoluchowski equation. In the end, continuum mechanics is recovered.

Fluctuations 'cause' transport *only in a static theory*. In a dynamic theory, we have balance equations- have you heard of flux?. To be sure, in the quantum mechanical view, fluctuations occur in a more fundamental sense- but that's when I would rely on the fluctuation-dissipation theorem. And again, continuum mechanics is recovered.

In the end, Iblis, I really don't understand what your objection is to anything that I have written.
 
  • #43
James A. Putnam said:
Andy Resnick,

<snip>
I think it demonstrates that we do not yet know what Thermodynamic Entropy is. I think that understanding the role of 'time' in Clausius' definition is important for explaining why Thermodynamic Entropy is a process that takes time. The question, as I see it, is why does Thermodynamic Entropy require time? I would appreciate reading more of your input even if what I have said does not conform to it.

James Putnam

I think I understand your question, and I would answer it this way:

Asking what entropy 'is' (or what temperature 'is'), is very similar to asking (in the context of mechanics) what mass 'is'. In mechanics, we can't answer very much beyond 'mass is how much stuff there is'. And that is generally accepted without further comment. Recall that the first and second laws of thermodynamics are assertions- they are not derived from anything.

Now, it is clear there is some connection between entropy and *temperature*. In thermodynamics, there is not (AFAIK) a 'problem of the direction of time'. Irreversible processes (and time-dependent processes) pose no real conceptual issue: see my results on linear friction posted earlier. Put another way, When Joule's weight drops and the water heats up, we are not surprised. But we would be *very* surprised if instead, we heated the water and the mass raised up, cooling the water.

There are many unsolved problems in Thermodynamics (limit of small size is an obvious one, wetting is another), but relating entropy change and time is not one of them.
 
Last edited:
  • #44
Andy Resnick,

Absorbing energy under equilibrium conditions requires time. As thermodynamic entropy changes so does something that relates to time. Why do you say that a process that requires time does not relate to time? Perhaps I do not understand your point. How does thermodynamic entropy change without relating it to a change in time?

James
 
  • #45
Andy Resnick,

I should have added that the following questions are important to answer before moving on to higher level theory that includes them.

"...Asking what entropy 'is' (or what temperature 'is'), is very similar to asking (in the context of mechanics) what mass 'is'. ..."

What is Thermodynamic Entropy? What is temperature? What is mass? I think they are each required to be answered in order to advance theoretical physics. However, the most immediate one is: What is mass? The answer to this question could change almost everything.

James
 
  • #46
Iblis,

The truth is that I would like to read your answers first. The subject of this thread justifies responding only to: What is Thermodynamic Entropy? What is it that Clausius discovered? Did he discover microstates?

James
 
  • #47
James, I think one of the reasons entropy is less understood and has acquired a special mystique is the question of the system boundary.

If we think of energy we are comfortable with notions of where it come from and where it goes to. We have energy conservation laws. We can collect it in a container, store it, transport it, release it at will.

But if we think of entropy where is it located?
When entropy increases where does it come from?
Is it possible to fill my beer glass or other container with entropy?
 
  • #48
James A. Putnam said:
Andy Resnick,

Absorbing energy under equilibrium conditions requires time. <snip>

James

This is one of the notions I hope to disabuse you of. Maybe you are thinking of steady-state conditions, but at equilibrium, there can be *no* absorption. That's part of the definition of equilibrium!
 
  • #49
James A. Putnam said:
Andy Resnick,

I should have added that the following questions are important to answer before moving on to higher level theory that includes them.

"...Asking what entropy 'is' (or what temperature 'is'), is very similar to asking (in the context of mechanics) what mass 'is'. ..."

What is Thermodynamic Entropy? What is temperature? What is mass? I think they are each required to be answered in order to advance theoretical physics. However, the most immediate one is: What is mass? The answer to this question could change almost everything.

James

I disagree we need any definition other than:

"mass" is how much stuff there is
"temperature" is how hot something is

I confess, I don't have such a short, clean, definition of entropy (yet). Some people claim we need to understand where mass 'comes from' (i.e. Higgs mechanism), and while I don't disagree with their overall goal, I also note that continuum mechanics will be changed not a whit by future theories about the origin of mass, and thermodynamics will be changed not a whit by "understanding the origin of entropy and temperature".

It's similar to the occasional post around here, claiming that to understand how light reflects or refracts, we first have to understand quantum electrodynamics. It's a foolish notion and unhelpful for students.
 
  • #50
Andy Resnick,

"...This is one of the notions I hope to disabuse you of. Maybe you are thinking of steady-state conditions, but at equilibrium, there can be *no* absorption. That's part of the definition of equilibrium! ..."

Clausius defined thermodynamic entropy by allowing for absorbtion under conditions of equilibrium. Is this an incorrect statement?

From your following message:

"...Some people claim we need to understand where mass 'comes from' (i.e. Higgs mechanism), and while I don't disagree with their overall goal, I also note that continuum mechanics will be changed not a whit by future theories about the origin of mass, and thermodynamics will be changed not a whit by "understanding the origin of entropy and temperature. ..."

Ok I get your point, though I think it has to do with what is debatable here and not with what is settled science. I disagree completely with it, but I respect your position. I will not press the issue here.

James
 
  • #51
Andy, I don't think we disagree on the physics itself. It is more of a disagreement of what follows from what (in principle), and that can be a matter of taste.

In my book: Entropy is the number of bits contained in the information you would need to specify in order to give the exact description of the physical state of an object given its macroscopic description.

Also: The mass of an object is the total rest energy of the object.

And, of course, I agree with this:

http://insti.physics.sunysb.edu/~siegel/history.html
 
  • #52
Iblis,

"...Also: The mass of an object is the total rest energy of the object. ..."

What is the total rest energy of the object? In other words, what is the material composition of energy whether rest or otherwise?

James
 
  • #53
James A. Putnam said:
Iblis,

"...Also: The mass of an object is the total rest energy of the object. ..."

What is the total rest energy of the object? In other words, what is the material composition of energy whether rest or otherwise?

James

That depends on the interactions in the system. Particles get their rest energy from the interaction with the Higgs field as Andy pointed out. But bound states of particles will contain extra energy (e.g. the mass of the proton is mostly due to the interaction between the quarks and not the masses of the quarks).

Exotic example. Take 6 square mirrors and make a cube out of it that encloses a perfect vacuum. Then the total energy of the cube is due to the rest energy (mass) of the 6 mirrors plus the Casimir energy of the enclosed vacuum. The mass of the empty cube is thus slightly more than the mass of the 6 mirrors.
 
  • #54
Count Iblis said:
That depends on the interactions in the system. Particles get their rest energy from the interaction with the Higgs field as Andy pointed out. But bound states of particles will contain extra energy (e.g. the mass of the proton is mostly due to the interaction between the quarks and not the masses of the quarks).

Exotic example. Take 6 square mirrors and make a cube out of it that encloses a perfect vacuum. Then the total energy of the cube is due to the rest energy (mass) of the 6 mirrors plus the Casimir energy of the enclosed vacuum. The mass of the empty cube is thus slightly more than the mass of the 6 mirrors.

I am pretty sure you have this backwards. Bound states have lower masses than their composite parts, because the lower potential energy represents a negative contribution to the total energy. That is why the most stable nucleus that can be formed by normal stellar fusion (nickel-56, which decays rapidly by successive anti-beta decays to form iron-56) also has just about the most negative mass defect of any isotope (nickel-62 and iron-58 have lower masses, but are not accessible by fusion pathways).

Similarly with the cube example, the mirrors form a cavity that restrict the wavelengths of the virtual particles that can exist inside the cube ... I guess that is what you mean by the Casimir energy, correct? Since the Casimir force in such a cavity is attractive, it should also make a negative contribution to the total energy, and so the mass of the empty cube would be slightly lighter than the mass of the component mirrors.
 
  • #55
Yes, bound states of protons and neutrons forming a nucleus will have a lower mass than the sum of the masses of the neutrons and protons. However, in case of the QCD forces between quarks you get the opposite effect, see here:

http://en.wikipedia.org/wiki/Proton#Quarks_and_the_mass_of_the_proton

So, you need to compare the energy of the system relative to some reference state. Now, you don't get free quarks when you pump more energy in a proton, you actually get further away from such a state. But perhaps some expert in QCD can explain that better than I can...

About the Casimir energy, I remember reading somewhere that for a cube this is positive, but I'm not 100% sure...
 
  • #56
Count Iblis said:
Andy, I don't think we disagree on the physics itself. \

I wouldn't be too sure about that. Specifically:

Count Iblis said:
Also, strictly speaking there is no such thing as thermodynamics. Strictly speaking it is always thermostatics.

The entropy of a system can only be rigorously defined within the statistical framework. There is no rigorous definition of entropy within "thermodynamics".

Count Iblis said:
Thermodynamics withoiut a statistical foundation does not make sense.

Count Iblis said:
So, basically my point is that the statistical foundations are simply hidden in the assumptions you make.

Count Iblis said:
From a fundamental POV, the fluctuations cause the transport phenomena, but you can describe these phenomena in an effective way.

You have a fatally flawed understanding of the subject, and are apparently not interested in correcting it.
 
Last edited:
  • #57
Count Iblis said:
Yes, bound states of protons and neutrons forming a nucleus will have a lower mass than the sum of the masses of the neutrons and protons. However, in case of the QCD forces between quarks you get the opposite effect, see here:

http://en.wikipedia.org/wiki/Proton#Quarks_and_the_mass_of_the_proton

So, you need to compare the energy of the system relative to some reference state. Now, you don't get free quarks when you pump more energy in a proton, you actually get further away from such a state. But perhaps some expert in QCD can explain that better than I can...

About the Casimir energy, I remember reading somewhere that for a cube this is positive, but I'm not 100% sure...

Yes, I agree that you are right about the quarks .. the strong force is special, and I don't really understand it either ... just that the harder you pull them apart, the more they binding force increase, until enough energy is put in for spontaneous quark-pair production.

I was just pointing out that in cases not involving quarks, things work the other way, and more strongly bound systems have lower masses.
 
  • #58
You have a fatally flawed understanding of the subject, and are apparently not interested in correcting it.


I think you fail to appreciate that in theoretical physics we have a different perspective than in engineering. I have learned the topic this way from books and lecture notes. I have taught this subject this way at Uni.

The view in theoretical physics will typically be that a the general state of system consisting of 10^23 degrees of freedom cannot be desribed using a handful of variables. Your state space is then simply not large enough to cover all the states the system can be in. This is the sort of perspective any decent lecture note on theoretical physics will start with in the introduction. Thermodynamics will then be developed as a coarse grained description of the system in some sense. We certainly do not start with vague undefined concepts like heat, temperature work etc. etc.

It may well be that this approach involves quite a bit of hand waving that you can then criticize. Certainly there are many open problems in non eqilibrium statistical mechanics. However, as I have pointed out many times now, Reif always invokes an ensemble even when he discusses non equilibrium phenomena.


And entropy always counts the number of microstates compatible with the macrostate the system is in, i.e. the amount of information contained in the exact description of the system. Ignore that and you are vulnerable to a Maxwell Demon type argument. So, I can lower the entropy of any system at the expense of filling the memory of a computer with random information that comes from the system.

And such thought experiments are perfectly feasible for non equilibrium states too. That's not an issue at all.
 
  • #59
Count Iblis, I think that the phenomenological concepts like heat, temperature, work that you mention are experimentally very well defined quantities while e.g. microstates are an idealization that does not exist for a real macroscopic body. The problem is that you cannot, even theoretically, decouple a macroscopic body from its environment which has infinite degrees of freedom. Thermodynamic concepts cannot be derived from a microscopic theory. You have to use these very concepts as an input to do statistical mechanics. See e.g. the very interesting article by Hans Primas, "Emergence in exact sciences" which can be downloaded e.g. here: http://philsci-archive.pitt.edu/archive/00000953/00/EmergenceInExactSciences.pdf
 
Last edited by a moderator:
  • #60
Dr. Du, thanks for the link to that article, I'll read it in my next break. But I will note that a theorist is not so interested with the experimental or pactical reality (even at the level of theory itself, i.e. that in practice you do smuggle in vague intuitions from thermodynamics).

We know the structure of the laws of physics at the microlevel and that alone allows you to set up certain thought experiments which show that entropy must be a measure of the information conained in the system or else you can violate the second law in principle.

Also note that the whole issue about information loss in black hole evaporation is only an issue because of the purely theoretical problems that would pose. So, the objections you and Andy have are not seen to be relevant in the way we view this subject.

But that doesn't mean that you don't have a point regarding the way we do physics in practice. So, that's why I wrote that on the physics itself there are no disagreements. It is just one of these intepretational issues like what the meaning of thought experiments that purport to show something.

E.g. I can always imagine completely isolating any thermodynamic system by imposing the right boundary conditions far away. Then I can use some Maxwell's Demon that uses a quantum computer to cool the system down to its quantum mechanical ground state. The memory of the quantum computer will then be filled by the information that initially resided in the system.

This transformation from initial to final state has to be a unitiary transformation. The minimum size of the memory needed is te entropy of the system (in the final state we're in the static case where this is not controversial). So this then means that the entropy of any arbitrary system has to be identified with the dimensionality of the Hilbert space that you need to describe it.


So, this can be the only theoretically correct definition of entropy (it is in fact the standard definition of entropy used by theoretical physicists). But its relevance may be limited to solve purely theoretical problems.

But then such purely theoretical way of looking at things is important. Special Relativity would not have been discovered in 1905 if everyone had objected to doing thought experiments that cannot even remotely be realized.
 

Similar threads

  • · Replies 4 ·
Replies
4
Views
5K
Replies
1
Views
2K
  • · Replies 20 ·
Replies
20
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
16K
Replies
7
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 15 ·
Replies
15
Views
3K