Concavity of Entropy: Is it True?

  • Thread starter Thread starter emma83
  • Start date Start date
  • Tags Tags
    Entropy
Click For Summary
SUMMARY

The discussion centers on the concavity of Shannon entropy, specifically its conditional form. Shannon entropy, defined as H(X)=-∑p(x)log p(x), is established as a concave function. Participants explore the relationship between conditional Shannon entropy H(X|Y) and its components, debating the inequality ∑p(y)H(X|Y=y) ≥ H(X|Y). The consensus is that the right-hand side must also be summed over y for the inequality to hold true.

PREREQUISITES
  • Understanding of Shannon entropy and its mathematical formulation
  • Familiarity with conditional probability and random variables
  • Knowledge of concave functions in mathematical analysis
  • Basic skills in summation notation and inequalities
NEXT STEPS
  • Study the properties of concave functions in information theory
  • Explore the implications of conditional entropy in statistical mechanics
  • Learn about the applications of Shannon entropy in machine learning
  • Investigate the relationship between entropy and information gain
USEFUL FOR

Mathematicians, information theorists, and students studying probability and statistics will benefit from this discussion, particularly those interested in the properties of entropy and its applications in various fields.

emma83
Messages
31
Reaction score
0

Homework Statement


Shannon entropy is a concave function defined as follows:
H(X)=-\sum_{x}p(x)\log p(x)

Conditional Shannon entropy is defined as follows:
H(X|Y)=\sum_{y} p(y) H(X|Y=y)=-\sum_{y} p(y)\sum_{x}p(x|Y=y)\log p(x|Y=y)

Can we deduce that:
\sum_{y} p(y)H(X|Y=y)\geq H(X|Y=y)

Homework Equations


The Attempt at a Solution


I would say yes because of the concavity but I am confused with the 2 random variables.
 
Physics news on Phys.org
I'm assuming the RHS of your final expression is also meant to be summed over y, otherwise it doesn't make much sense...

The way to go about this is to think about the nature of p(y). What constraints do you know about the values that p(y) can take?

PS Do you really mean \geq in the last line, or do you mean \leq?
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 5 ·
Replies
5
Views
1K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 18 ·
Replies
18
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
Replies
1
Views
2K
  • · Replies 22 ·
Replies
22
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K