# Entropy: Definition, Misconceptions & Increase in Closed System

In summary: Is water an example of an ordered material?Water is an example of an ordered material, because it does have a lowest point in its phase diagram. However, water is not a closed system, because it is open to the atmosphere. The air molecules can randomly enter and exit the water molecule, contributing to its entropy.
A lot of the less maths-y definitions of entropy talk about disorder and how disordered a system is. I'm given to understand that entropy is a measure of energy over temperate. Could someone clear up these misconceptions? I don't understand why 'disorder' is used. Isn't that subjective?

Second question. I don't understand why entropy has to increase or stay constant in a closed system over time. Surely say, in a room where air molecules are bouncing around, they would at some point move to a more 'ordered' state?

So, there's two definitions of entropy.
The first (and oldest) one is based on heat and temperature.
The second, more modern definition is based on "disorder".

"disorder" is a confusing term.
To put it a touch more precisely, entropy is a measure of statistical dispersion, or how spread out the system is, among all its possible configurations.

The formal definition of entropy $S$ if a system is:
$S=k_{B}log(W)$
where $k_{B}$ is Boltzmann's constant, and $W$ is the number of "ways" that all the states of the atoms that make the system up can be arranged so as to give the same overall state of the object (total energy, particle number, volume, etc). Boltzmann's constant is there to relate the modern definition of entropy to the old definition of entropy, so everything works out right. There is a little bit of arbitrariness in what counts as a "way", but whether or not we include, say, nuclear spin into the entropy, only shifts it by a constant amount, and doesn't affect most calculations.

Technically, there is no dynamical reason why the entropy has to increase or stay constant in a closed system over time. It is a fact of statistics, rather than something written into the laws of motion that it does.

The only reason we see that the entropy of large scale systems seems to inexorably increase or stay constant in a closed system is that it is overwhelmingly more likely than all other possibilities.

For a small chamber with an ultra high vacuum of maybe a dozen atoms, you could wait long enough and see at some point of time, all of those atoms being on one side of the container, and not the other. It may take a while, but it almost certainly will happen.

If you instead take a liter size vessel with air at room temperature and atmospheric pressure, you're waiting for the time when roughly a hundred billion trillion atoms will happen to be on one side of the bottle and not the other. Although there is a nonzero probability, you would have to wait an unbelievably long time to see such an occurrence (unbelievably long, even compared to the age of the Universe).

On the less absurd scale of things, you would probably be able to see minute fluctuations back and forth in the entropy of a closed system. The entropy of a closed system is not truly nondecreasing, but that is overwhelmingly the most likely thing that we see.

I suggest you start at entropy disambiguation on Wikipedia, and follow some of the links. There are many definitions of entropy, all correct but with differing contexts and emphasis.

In some contexts the disorder makes sense. In others the unknown information makes sense. In thermo, the definitions jfizzix gave make sense.

The important thing IMHO is that the second law of thermodynamics applies to all the definitions.

Thank you both. That helped a lot.

Hi, this is also one that I struggle with. And every time I read something that defines it, I get lost in the jargon and the formulas. I'm good with algebra, but I struggle to tie an equation to a theory or law, which is why analogies work much better for me. There are a few aspects of entropy I'd like to understand and please please do correct me.

The first aspect is the closed system explanation. Is there anywhere outside experimental situations where you would find a closed system? Whenever I try to think of one, I think it's closed but then realize there are outside factors. For example, a kettle that's cooling is doing so because of the room it is in, and the room temperature is affected by airconditioning, external temperatures, insulation etc. And the external temperatures are affected by local weather, heat energy from the sun etc. If there are no closed systems in real situations, are the experimental ones really only for establishing baseline formulas etc?

Secondly, my understanding of entropy is twofold, yet related. I understand it as a kind of equilibrium. Like the boiled kettle going from a higher temperature on a steady slope down to room temperature. In the same way, I understand it as analogous to water finding its lowest place to settle. In that same way, all "ordered" materials will eventually find their lowest point by breaking down and becoming disordered, like a piece of steel or a human body will both eventually become part of the soil again.

One definition I've read is the hardest one of all in some respects. I've read it as the decrease in energy available to do work. Conservation of energy says that we never lose energy, it just changes form. So it's my understanding that "available" energy is like electricity in a battery, which is converted into work. But after the battery is depleted of electricity, it still has energy, but not in a form we can use. In some respects, a battery is both an analogy of entropy and also an observable mechanism of that decrease in available energy, except that it breaks down as both analogy and true observation on so many levels. Energy has been removed from (is no longer part of) the battery and been converted to heat or movement or photons, and so that energy has gone elsewhere. What remains of the depleted battery are elements that can no longer offer energy. And the energy that has gone elsewhere has dispersed, but into what I don't know. Does all "available" energy eventually radiate out to space as heat or photons? I guess this is where a closed system becomes a useful tool.

Apologies for the length and meandering.

narrator said:
Hi, this is also one that I struggle with. And every time I read something that defines it, I get lost in the jargon and the formulas. I'm good with algebra, but I struggle to tie an equation to a theory or law, which is why analogies work much better for me. There are a few aspects of entropy I'd like to understand and please please do correct me.

The first aspect is the closed system explanation. Is there anywhere outside experimental situations where you would find a closed system? Whenever I try to think of one, I think it's closed but then realize there are outside factors. For example, a kettle that's cooling is doing so because of the room it is in, and the room temperature is affected by airconditioning, external temperatures, insulation etc. And the external temperatures are affected by local weather, heat energy from the sun etc. If there are no closed systems in real situations, are the experimental ones really only for establishing baseline formulas etc?
You are mistaken about the definition of a Closed System. A closed system is one that does not exchange mass with the surroundings, but is fully capable to exchanging energy with the surroundings in the form of both work and heat.
Secondly, my understanding of entropy is twofold, yet related. I understand it as a kind of equilibrium. Like the boiled kettle going from a higher temperature on a steady slope down to room temperature. In the same way, I understand it as analogous to water finding its lowest place to settle. In that same way, all "ordered" materials will eventually find their lowest point by breaking down and becoming disordered, like a piece of steel or a human body will both eventually become part of the soil again.

One definition I've read is the hardest one of all in some respects. I've read it as the decrease in energy available to do work. Conservation of energy says that we never lose energy, it just changes form. So it's my understanding that "available" energy is like electricity in a battery, which is converted into work. But after the battery is depleted of electricity, it still has energy, but not in a form we can use. In some respects, a battery is both an analogy of entropy and also an observable mechanism of that decrease in available energy, except that it breaks down as both analogy and true observation on so many levels. Energy has been removed from (is no longer part of) the battery and been converted to heat or movement or photons, and so that energy has gone elsewhere. What remains of the depleted battery are elements that can no longer offer energy. And the energy that has gone elsewhere has dispersed, but into what I don't know. Does all "available" energy eventually radiate out to space as heat or photons? I guess this is where a closed system becomes a useful tool.

Apologies for the length and meandering.
The concept of entropy originally developed from the theoretical understanding that evolved during the development of the second law of thermodynamics. The second law was the result of experimental observations. These experiments indicated that all materials and closed systems must exhibit a unique physical property that was dubbed entropy. This physical property applies to thermodynamic equilibrium states of the material or system. The statement of the second law as captured mathematically by noting that the change in entropy in going from one thermodynamic equilibrium state to another is greater or equal to the heat transferred to the system in any process transitioning the system from one equilibrium state to the other divided by the temperature at the boundary of the system through which the heat is flowing. For a so-called reversible process path (a subset of all the possible process paths between the two equilibrium states), the equal sign applies. This provides a way of measuring or calculating the entropy change between the two equilibrium states.

Chet

Last edited:
Reading about Maxwell's demon helped me understand better why entropy doesn't decrease in closed systems. Its a thought experiment, very easy to visualize and follow.

Enclose said:
Reading about Maxwell's demon helped me understand better why entropy doesn't decrease in closed systems. Its a thought experiment, very easy to visualize and follow.
I think you mean isolated systems, not closed systems. Entropy can certainly decrease in closed systems.

Chet

Oops, indeed I do Chestermiller, thanks for pointing that out.

## 1. What is entropy?

Entropy is a measure of the disorder or randomness in a system. It is a thermodynamic property that describes the amount of energy that is unavailable for work in a system.

## 2. What are some common misconceptions about entropy?

One common misconception is that entropy only applies to physical systems. In reality, entropy can also be used to describe the amount of disorder or randomness in other systems such as information or social systems. Additionally, some people mistakenly believe that entropy always increases, when in fact it can decrease in certain circumstances.

## 3. How is entropy defined in thermodynamics?

In thermodynamics, entropy is defined as the ratio of heat energy to temperature. This means that as the temperature of a system increases, its entropy also increases. It is also related to the number of possible microstates that a system can have at a given energy level.

## 4. Can entropy be decreased in a closed system?

In a closed system, where no energy or matter can enter or leave, the total entropy of the system will always increase. However, it is possible for the entropy of a specific component or subsystem within the closed system to decrease, as long as the overall entropy of the entire system increases.

## 5. How can entropy be increased in a closed system?

Entropy can be increased in a closed system through processes such as heat transfer, chemical reactions, and mixing of different components. These processes all lead to an increase in the disorder or randomness of the system, resulting in an increase in entropy.

Replies
1
Views
933
Replies
3
Views
1K
Replies
12
Views
2K
Replies
23
Views
1K
Replies
10
Views
2K
Replies
4
Views
1K
Replies
13
Views
2K
Replies
13
Views
2K
Replies
3
Views
1K
Replies
8
Views
1K