Conditional Entropy and Kullback–Leibler divergence

In summary, the relation between conditional entropy and KL divergence can be expressed as H[X | Y] = H[X,Y] - H[Y] = H[p(x,y) || p(x)p(y)].
  • #1
avkr
12
0

Homework Statement


To find relation between conditional (Shanon) entropy and KL divergence.

Homework Equations


Conditional Entropy: H[X | Y] = H[X,Y] - H[Y]
KL Divergenece: H[X || Y] = -H[X] - Σx ln(y)

The Attempt at a Solution


H[p(x,y) || p(x)p(y)] = -H[p(x,y)] + H[p(x)] + H[p(y)]
 
Physics news on Phys.org
  • #2
H[p(x,y) || p(x)p(y)] = - H[X,Y] + H[X] + H[Y] H[p(x,y) || p(x)p(y)] = - H[X,Y] + H[X|Y] + H[Y] H[p(x,y) || p(x)p(y)] = - (H[X,Y] - H[Y]) + H[Y] H[p(x,y) || p(x)p(y)] = - H[X|Y]Hence, H[X | Y] = H[X,Y] - H[Y] = H[p(x,y) || p(x)p(y)]
 

1. What is conditional entropy?

Conditional entropy is a measure of the uncertainty or randomness of a random variable given the value of another random variable. It quantifies the amount of information needed to describe a random variable Y, given that another random variable X has already been observed.

2. How is conditional entropy calculated?

The formula for conditional entropy is H(Y|X) = -Σ p(y,x)log p(y|x), where p(y,x) is the joint probability distribution of X and Y, and p(y|x) is the conditional probability of Y given X. This formula can be applied to both discrete and continuous random variables.

3. What is Kullback-Leibler divergence?

Kullback-Leibler (KL) divergence is a measure of how different two probability distributions are. It is calculated as DKL(P||Q) = Σ p(x)log(p(x)/q(x)), where P and Q are two probability distributions. KL divergence is often used to compare a model's predicted distribution to the true distribution.

4. How is Kullback-Leibler divergence related to conditional entropy?

KL divergence can be thought of as a measure of the additional uncertainty that results from using one probability distribution (Q) to approximate another (P). In other words, KL divergence can be seen as a measure of the difference between the conditional entropy of P given Q and the conditional entropy of P. This relationship is expressed as DKL(P||Q) = H(P|Q) - H(P).

5. What are some real-world applications of conditional entropy and Kullback-Leibler divergence?

Conditional entropy and KL divergence have various applications in fields such as information theory, statistics, and machine learning. They are commonly used in data compression, pattern recognition, and information retrieval. KL divergence is also used in probabilistic programming and Bayesian inference to measure the difference between model predictions and actual data.

Similar threads

Replies
1
Views
639
  • Advanced Physics Homework Help
Replies
14
Views
2K
  • Advanced Physics Homework Help
2
Replies
39
Views
9K
  • Advanced Physics Homework Help
Replies
1
Views
730
  • Advanced Physics Homework Help
Replies
3
Views
194
  • Advanced Physics Homework Help
Replies
12
Views
2K
  • Advanced Physics Homework Help
Replies
1
Views
1K
  • Advanced Physics Homework Help
Replies
9
Views
1K
  • Advanced Physics Homework Help
Replies
2
Views
1K
  • Advanced Physics Homework Help
Replies
8
Views
1K
Back
Top