Chernoff Bounds using importance sampling identity

  • Thread starter Thread starter WMDhamnekar
  • Start date Start date
Click For Summary
The discussion focuses on using the importance sampling identity to derive Chernoff bounds for a binomial random variable. It outlines the process of bounding the probability that the random variable deviates from its mean by more than a specified amount. By defining the binomial variable as a sum of independent Bernoulli trials and applying the importance sampling technique, the probability is expressed in terms of expectations under a different measure. The resulting Chernoff bound is formulated, allowing for the selection of a suitable parameter to achieve a tight bound on tail probabilities. The solution is presented as correct, inviting further questions for clarification.
WMDhamnekar
MHB
Messages
376
Reaction score
28
How to use importance sampling identity to obtain the Chernoff bounds as given below?

Let X have moment generating function ##\phi(t)= E[e^{tX}]##. Then, for any c > 0 ,

##P[X\geq c ]\leq e^{-tc} \phi(t), \text{if t > 0}##

##P[X \leq c]\leq e^{-tc}\phi(t), \text{if t<0} ##

Solution:
Here's an example of using the importance sampling identity to obtain Chernoff bounds for a binomial random variable.

Suppose we have a binomial random variable ##X \sim \text{Bin}(n,p)##, where n is the number of trials and p is the probability of success on each trial. We want to bound the probability that X deviates from its mean by more than a certain amount, i.e., we want to bound ##P(X \geq (1+\delta)np)## for some ##\delta > 0##.

Let ##X = \sum_{i=1}^n X_i##, where the ##X_i## are independent Bernoulli random variables with parameter p. Let ##Q_i## be a Bernoulli distribution with parameter q, and let ##Q = Q_1 \times \cdots \times Q_n##.

Then, by the importance sampling identity, we have ##P(X \geq (1+\delta)np) = E_Q\left[\prod_{i=1}^n \frac{dP_i}{dQ_i}I_{X \geq (1+\delta)np}\right]##

Now, let ##\phi_i(t) = E[e^{tX_i}] = 1 + p(e^t - 1)## and ##\psi_i(t) = E_Q[e^{tX_i}] = 1 + q(e^t - 1)##.

By choosing ##Q_i## such that ##\frac{dP_i}{dQ_i} = e^{tX_i - \log \phi_i(t)}##, we obtain ##P(X \geq (1+\delta)np) = E_Q\left[e^{tX - n\log \phi(t)}I_{X \geq (1+\delta)np}\right]##

Applying Markov's inequality, we get ##P(X \geq (1+\delta)np) \leq e^{-t(1+\delta)np}E_Q\left[e^{tX - n\log \phi(t)}\right] = e^{-t(1+\delta)np}\prod_{i=1}^n\psi_i(t)e^{-\log \phi_i(t)} = e^{-t(1+\delta)np}\left(\frac{1+q(e^t-1)}{1+p(e^t-1)}\right)^n##

This is the Chernoff bound for a binomial random variable. By choosing an appropriate value of t, we can obtain a tight bound on the tail probability.

In my opinion, this solution seems to look correct. Do you have any doubts? Let me know, so that I shall clear them.
 
Last edited:
There is a nice little variation of the problem. The host says, after you have chosen the door, that you can change your guess, but to sweeten the deal, he says you can choose the two other doors, if you wish. This proposition is a no brainer, however before you are quick enough to accept it, the host opens one of the two doors and it is empty. In this version you really want to change your pick, but at the same time ask yourself is the host impartial and does that change anything. The host...

Similar threads

Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K