Convergence of the Expected Value of a Function

Click For Summary

Discussion Overview

The discussion revolves around the convergence of the expected value of a function of a binomially distributed variable as the number of trials approaches infinity. Participants explore the implications of different types of functions and the conditions under which convergence occurs, including the use of integrals and limits.

Discussion Character

  • Exploratory
  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • One participant proposes that for a binomially distributed variable, the expected value of a non-linear, decreasing, and continuous function converges to f(p) as n approaches infinity, while noting that linear functions yield a different result.
  • Another participant questions the clarity of the initial formulation, asking for clarification on the parameters of the binomial distribution and the treatment of "nx".
  • A later reply clarifies the goal of evaluating a specific sum involving binomial coefficients and a decreasing function, aiming to show that the limit leads to p = f^{-1}(c) as n approaches infinity.
  • Further elaboration discusses the need to define an interval around the mean of the binomial distribution that grows with n, while also becoming a smaller proportion of n, to facilitate convergence of the sum to an integral.
  • Participants discuss bounding the sum by analyzing the contributions from different segments of the distribution and establishing upper and lower bounds based on the behavior of the function within those segments.

Areas of Agreement / Disagreement

Participants express differing views on the formulation and interpretation of the binomial distribution and the expected value of the function. There is no consensus on the approach to proving convergence or the specific conditions required for the expected value to equal f(p).

Contextual Notes

Participants highlight potential ambiguities in the definitions and assumptions regarding the binomial distribution parameters, the treatment of the function f, and the limits involved in the convergence process. The discussion includes various mathematical steps that remain unresolved.

ntg865
Messages
8
Reaction score
0
Suppose that nx is binomially distributed: B((n-1)p, (n-1)p(1-p))
I wish to find the expected value of a function f(x), thus
[itex]\sum_{nx=0}^{n-1} B() f(x)[/itex]
Assume that f() is non-linear, decreasing and continuous, f(x) = c is [0,1] to [0, ∞)
I want to show that the above sum converges to f(p) if n→∞
By computation using all types of different functions (even very extreme ones), it is clear that it converges to that. The only thing is that while the proof looks like it'd be simple, I just couldn't figure it out. If f() is linear, then we just get f((n-1)p/n) which converges to f(p), if it's not, Jensen's inequality tells us that we will get something else.

If we are not happy with the sum we can use n→∞ to show the sum goes to:
[tex]\int_{0}^{∞} \text{(normal distribution of nx with same mean and variance as above)} f(nx/n) dnx[/tex]

Just wondering if anyone has a tip or so of how I can approach this. Thanks
 
Physics news on Phys.org
ntg865 said:
Suppose that nx is binomially distributed: B((n-1)p, (n-1)p(1-p))
I'm used to seeing the binomial written as B(n,p) or, in your case, it woud be B(n-1,p). I think you have given the parameters as the mean and variance instead of the number of trials and the probability of succes.

I wish to find the expected value of a function f(x), thus
[itex]\sum_{nx=0}^{n-1} B() f(x)[/itex]

It is completely unclear what you mean by that.

Are you treating "nx" as one variable or is it the product (n)(x)?

What is the argument of the function B() ? B(x)? B(nx)? B(n)?

Are you trying to use the density of the binomial distribution:
[itex]b(k,n,p) = C^n_k p^k (1-p)^{n-k}[/itex] ?
or, in your case [itex]B(k) = b(k,n-1,p)[/itex].
 
Thanks for the reply.

I am trying to valuate the following sum:

[tex]\sum_{i=0}^{n-1} C_i^{n-1} \; p^i (1-p)^{n-1-i} f(\frac{i}{n}) = c[/tex]

where c is a real, positive constant, f is a decreasing function, and I am trying to evaluate p. I know that the end result when n → ∞ is [itex]p = f^{-1} (c)[/itex] and I want to prove it.

I previously defined [itex]x= \frac{i}{n}[/itex] (thus i = xn) to make f() independent of n, which allows me to take the limit of n and converge the sum into an integral using demoivre-laplace theorem.

And sorry, by binomial I did mean b(n-1, p). Hope this makes it clearer
 
So rephrasing, you want to find:
[tex]V = lim_{n \rightarrow \infty} \sum_{i=0}^{n} {{n}\choose{i}} p^i (1-p)^{n-i} f(\frac{i}{n})[/tex]

Our goal is now to pick an interval around the mean np, which includes an increasing number of standard deviations around the mean as n goes to infinity (which means the interval covers an amount of probability that goes to 1), but where the interval also becomes a smaller proportion of n as n goes to infinity (which means the interval includes only values from a small local part of f). Then we can show that the sum can be approximated only by values from within this interval, and provide converging upper and lower bounds for the sum within the interval.

Here's a picture of the interval (labeled B) and the parts beside it (A and C):
attachment.php?attachmentid=51324&stc=1&d=1348989737.png


The variance of the binomial distribution is np(1-p), for a standard deviation of [itex]\sqrt{np(1-p)} = \sqrt{n} \sqrt{p(1-p)}[/itex]. So the size of the interval must grow faster than [itex]\sqrt{n} \sqrt{p(1-p)}[/itex] in order to include an increasing number of standard deviations around the mean. It must also grow slower than n, in order to decrease relative to n. So by picking any function asymptotically between [itex]sqrt{n}[/itex] and n, let's choose the width of the interval to be [itex]2 n^{2/3}[/itex]. We want the interval to be centered about the mean, and since it's used in the sum the endpoints must be integers, so that gives us [itex](\lceil np - n^{2/3} \rceil, \lceil np + n^{2/3} \rceil)[/itex]

Now rewrite the limit in terms of the parts of the sum outside the interval, and inside the interval:
[tex]lim_{n \rightarrow \infty} \left( <br /> \sum_{i=0}^{\lceil{np-n^{2/3}}\rceil} {{n}\choose{i}} p^i (1-p)^{n-i} f(\frac{i}{n})<br /> + \sum_{i=\lceil{np-n^{2/3}}\rceil + 1}^{\lceil{np + n^{2/3}}\rceil} {{n}\choose{i}} p^i (1-p)^{n-i} f(\frac{i}{n})<br /> + \sum_{i=\lceil{np + n^{2/3}}\rceil + 1}^{n} {{n}\choose{i}} p^i (1-p)^{n-i} f(\frac{i}{n})<br /> \right)[/tex]
And let's label those three parts of the limit as follows, to refer to them:
[tex]lim_{n \rightarrow \infty} \left( A(n) + B(n) + C(n) \right)[/tex]

The total amounts of probability contained within A, B, and C are:
[tex]A_p(n) = \sum_{i=0}^{\lceil{np-n^{2/3}}\rceil} {{n}\choose{i}} p^i (1-p)^{n-i}[/tex]
[tex]B_p(n) = \sum_{i=\lceil{np-n^{2/3}}\rceil + 1}^{\lceil{np + n^{2/3}}\rceil} {{n}\choose{i}} p^i (1-p)^{n-i}[/tex]
[tex]C_p(n) = \sum_{i=\lceil{np + n^{2/3}}\rceil + 1}^{n} {{n}\choose{i}} p^i (1-p)^{n-i}[/tex]

Also, denote the maximum and minimum values for f within A, B, and C by [itex]A_{fmin}(n), B_{fmin}(n), C_{fmin}(n), A_{fmax}(n), B_{fmax}(n) C_{fmax}(n)[/itex].

Now it's in a solvable form. To solve it, first find a lower bound on the limit, by finding lower bounds on A, B, and C. A is lower bounded by [itex]A_{fmin}(n) A_p(n)[/itex]. And similarly for B and C.
[tex]L = lim_{n \rightarrow \infty} ( A_{fmin}(n) A_p(n) + B_{fmin}(n) B_p(n) + C_{fmin}(n) C_p(n))[/tex]
Also find an upper bound:
[tex]U = lim_{n \rightarrow \infty} ( A_{fmax}(n) A_p(n) + B_{fmax}(n) B_p(n) + C_{fmax}(n) C_p(n))[/tex]

Now you must prove that L = f(p). First, since [itex]A_p(n)[/itex] and [itex]B_p(n)[/itex] go to 0 as n increases, the first and third parts of the sum go to 0. Second, [itex]B_p(n)[/itex] goes to 1 as n increases (due to careful choice of the interval, earlier). And third, f is continuous, so [itex]B_{fmin}[/itex] approaches f(p) as the interval shrinks around p.

In a similar fashion you prove that U = f(p). Then since the limits in U and L sandwich V from above and below, you can conclude V = f(p).
 

Attachments

  • abc.png
    abc.png
    2.8 KB · Views: 480

Similar threads

  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 14 ·
Replies
14
Views
2K
  • · Replies 19 ·
Replies
19
Views
4K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 9 ·
Replies
9
Views
3K