Understanding Mixed Conditional PDFs in Continuous and Discrete Variables

  • Context: Graduate 
  • Thread starter Thread starter pluviosilla
  • Start date Start date
  • Tags Tags
    Conditional Mixed
Click For Summary
SUMMARY

The discussion centers on the identity for conditional probability density functions (PDFs) involving continuous random variable X and discrete variable N. The identity is expressed as f_{X|Y}(x|n) = \frac{P(N=n|X=x)}{P(N=n)}f(x), which illustrates how a conditional PDF can be represented as a product of a non-conditional PDF and the right-hand side of Bayes' Theorem. The conversation also touches on the application of this identity to derive the beta distribution from a binomial distribution when the success probability is uniformly distributed. Key points include the transformation of conditions and the significance of the binomial coefficient in the context of the example problem.

PREREQUISITES
  • Understanding of conditional probability and probability density functions (PDFs)
  • Familiarity with Bayes' Theorem and its applications
  • Knowledge of beta distributions and their properties
  • Basic concepts of binomial distributions and combinatorial coefficients
NEXT STEPS
  • Study the derivation of conditional PDFs in continuous and discrete variables
  • Explore the applications of Bayes' Theorem in statistical inference
  • Learn about the properties and applications of beta distributions in Bayesian statistics
  • Investigate the role of binomial coefficients in probability distributions and their significance in statistical modeling
USEFUL FOR

Statisticians, data scientists, and researchers in fields involving probability theory and Bayesian statistics will benefit from this discussion, particularly those interested in the relationship between conditional and marginal distributions.

pluviosilla
Messages
17
Reaction score
0
I ran across this identity for a conditional PDF where the dependent random variable X is continuous and the independent variable N is discrete:

\frac{P(x<X<x+dx|N=n)}{dx}=\frac{P(N=n|x<X<x+dx)}{P(N=n)}\frac{P(x<X<x+dx)}{dx}

In the limit as dx approaches 0 this yields:

f_{X|Y}(x|n) =\frac{P(N=n|X=x)}{P(N=n)}f(x)

I think I understand the 2nd step, but not the initial identity. The reversing of the conditions (from X dependent on N to N dependent on X) reminds me of Bayes Law, but if he is using Bayes Law here, it is not clear to me exactly how. Could someone help me understand this identity?
 
Physics news on Phys.org
The dx is probably misleading here, the first identity is a Newton quotient,
<br /> \frac{Pr(x&lt;X&lt;x+y)}{y}=\frac{Pr(X&lt;x+y)-Pr(X&lt;x)}{y}
This when y tends to 0 gives the derivative. You can condition this on N but I still don't see why. The second identity comes from Bayes' theorem as you said, i.e.
Pr(A|B)=\frac{Pr(B|A) Pr(A)}{Pr(B)}
If you let
A=x&lt;X&lt;x+y \quad B=N=n
you get exacly the right hand side of the equation.
 
  • Like
Likes   Reactions: pluviosilla
Thanks for this reply!

How about the right hand term? I refer to this:


\frac{Pr(x &lt; X &lt; x + dx)}{dx}

This term gives the non-conditional marginal PDF f(x), and the point of the identity seems to be to show that you can represent a conditional PDF as a product of a non-conditional PDF and the RHS of Bayes Theorem. But where does that righthand term come from?

Here's a problem this identity is reputed to solve. Consider n + m trials having a common probability of success. Suppose, however, that this success probability is not fixed in advance but is chosen from a uniform (0,1) population. We want to determine the conditional PDF for success given that the n + m trials result in n successes.

Using the identity established above:

f_{X|Y}(x|n) =\frac{P(N=n|X=x)}{P(N=n)}f(x)

we have

\frac{\left(\begin{array}{c}n+m\\n\end{array}\right)x^n (1-x)^m }{Pr(N=n)}=cx^n (1-x)^m

which is the PDF of a beta random variable with parameters n + 1 and m + 1.

This is supposed to show that when PDF of a trial success is uniformly distributed over (0,1) prior to the collection of data, then given n sucesses in n+m trials the posterior or conditional PDF is a beta distribution with parameters n + 1 and m + 1.
 
Please disregard my question about the righthand term. I'm used to seeing Bayes THM with an AND term in the numerator, so my left brain took over and I did not *see* the conditional term, which of course, explains why you need the righthand term.

APOLOGIES!

However, I am also not sure I understand how he applies the identity to obtain the solution of the example problem. Once x established, the number of successes is modeled by the the binomial distribution. O.k. But he feels free to absorb the binomial coefficient and the denominator into a constant c as if it were irrelevant to the main point of showing that this is a beta distribution:

\frac{\left(\begin{array}{c}n+m\\n\end{array}\right)}{Pr(N=n)}

Why is the author so dismissive of the binomial coefficient and the denominator?
 

Similar threads

  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 5 ·
Replies
5
Views
5K
  • · Replies 4 ·
Replies
4
Views
3K
Replies
5
Views
2K
  • · Replies 8 ·
Replies
8
Views
4K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K