Show extremum of the Entropy is a maximum

Click For Summary
SUMMARY

The discussion focuses on demonstrating that the extremum of the entropy function, defined as S = -k_B ∑ p(i) ln(p(i)), is a maximum. Participants confirm that the extremum S' = k_B ln(N) is positive and explore the inequality S' - S = k_B ln(N) + ∑ p(i) ln(p(i)) > 0. They suggest using Lagrange multipliers and induction to analyze boundary conditions where one or more p(i) are zero, ultimately concluding that transferring probability from higher to lower p(i) values increases entropy, reinforcing the maximum nature of S'.

PREREQUISITES
  • Understanding of entropy in statistical mechanics
  • Familiarity with Lagrange multipliers for constrained optimization
  • Knowledge of Taylor series expansions
  • Basic principles of probability distributions
NEXT STEPS
  • Study the application of Lagrange multipliers in optimization problems
  • Learn about Taylor series and their use in approximating functions
  • Investigate the properties of probability distributions and their boundaries
  • Explore entropy maximization techniques in statistical mechanics
USEFUL FOR

Students and researchers in physics, particularly those studying statistical mechanics, thermodynamics, and information theory, will benefit from this discussion.

dipole
Messages
553
Reaction score
149

Homework Statement



The Entropy of a probability distribution is given by,

S = -k_B \sum _{i=1}^N p(i)\ln{p(i)}

I've shown that the extremum of such a function is given by,

S' = k_B \ln{N} (which is a positive quantity)

Now I want to show that this is a maximum by showing that

S' - S = k_B \ln{N} + \sum _{i=1}^N p(i)\ln{p(i)} > 0



Homework Equations



The p(i)'s are constrained by

\Sum_{i=1}^N p(i) =1

The Attempt at a Solution



I'm kind of stuck here. The second term is inherently negative, so it's not a priori obvious that S' - S > 0. I would probably want to take the ratio and show \frac{S'}{S} \geq 1 but I'm not sure how to do this.

Any ideas?
 
Physics news on Phys.org
I suggest writing a Taylor series for S in terms of the p_i and looking at the second-order term.
 
dipole said:

Homework Statement



The Entropy of a probability distribution is given by,

S = -k_B \sum _{i=1}^N p(i)\ln{p(i)}

I've shown that the extremum of such a function is given by,

S' = k_B \ln{N} (which is a positive quantity)

Now I want to show that this is a maximum by showing that

S' - S = k_B \ln{N} + \sum _{i=1}^N p(i)\ln{p(i)} > 0



Homework Equations



The p(i)'s are constrained by

\Sum_{i=1}^N p(i) =1

The Attempt at a Solution



I'm kind of stuck here. The second term is inherently negative, so it's not a priori obvious that S' - S > 0. I would probably want to take the ratio and show \frac{S'}{S} \geq 1 but I'm not sure how to do this.

Any ideas?

You only found one extremum which has a positive entropy (using Lagrange multipliers, I presume). The only other place you could have an extremum is on the boundary. The boundary of your set of p(i) would be the case where one of the p(i) is 1 and the rest are 0. What's the entropy there? You have to think about limits, since log(0) is undefined.
 
Dick said:
The boundary of your set of p(i) would be the case where one of the p(i) is 1 and the rest are 0.

Hello Dick!

The boundary of the space in question is actually much bigger than this - it's the set of points for which at least one p(i) is zero. So doing it this way, some work remains.
 
Oxvillian said:
Hello Dick!

The boundary of the space in question is actually much bigger than this - it's the set of points for which at least one p(i) is zero. So doing it this way, some work remains.

Good point. But if you fix one of your N p(i) to be zero. Then you are in the N-1 case with the remaining p(i). Suggests you use induction. In the case N=2, the boundary is the two points (p(1),p(2)) equal to (1,0) or (0,1).
 
Dick said:
Good point. But if you fix one of your N p(i) to be zero. Then you are in the N-1 case with the remaining p(i). Suggests you use induction. In the case N=2, the boundary is the two points (p(1),p(2)) equal to (1,0) or (0,1).

Correct :smile:

But (at the risk of giving away the plot), I think the easiest way to demonstrate the global nature of the max is to notice that you can always increase the entropy by transferring probability from somewhere with a large p(i) to somewhere with a small p(i).
 
Oxvillian said:
Correct :smile:

But (at the risk of giving away the plot), I think the easiest way to demonstrate the global nature of the max is to notice that you can always increase the entropy by transferring probability from somewhere with a large p(i) to somewhere with a small p(i).

That is a simpler way to look at it.
 

Similar threads

  • · Replies 7 ·
Replies
7
Views
3K
Replies
4
Views
5K
Replies
1
Views
2K
  • · Replies 10 ·
Replies
10
Views
5K
Replies
6
Views
2K
Replies
2
Views
2K
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
Replies
12
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K