Combination of two dependant discrete random variables

In summary, the conversation discusses the problem of combining two discrete random variables with a known correlation between them. The goal is to find the joint probability distribution or complete probability distribution of the combination. To solve this problem, simultaneous equations can be set up using the known marginal probabilities and correlation function. However, the ill posed nature of the problem means there may be infinitely many solutions or no solution at all. Additionally, the probability distribution for the combined variable can be determined using the characteristic function and the normal approximation.
  • #1
simcc
1
0
Hi,
I’m looking for a way to combine two discrete random variables (which I have as probability distributions). The combination should be the product (or other operation) of the two variables.
This would be easy if they were independent, but they’re not. There is a known correlation between the variables.

Question: how to combine two discrete random variables with correlation?
Given: The marginal probabilities of the two variables & a correlation function
Result: either the individual probabilities in a probability table or the complete probability distribution of the combination.

Simple example:
Variables A and B are the distributions:
PA(a=1, 4) = [0.75, 0.25]
PB(b=4, 8, 10) = [0.25, 0.25, 0.5]

Their joint probability function is shown in their joint probability table and joint value table:
P B=4 8 10
A=1 ? ? ? 0.75
4 ? ? ? 0.25
0.25 0.25 0.5 1

value B=4 8 10
A=1 4 8 10
4 16 32 40

(tables are clearer in attached file)

The correlation between the two variables is: b = 10 – 2/3*a

P(A*B)(4, 8, 10, 16, 32, 40) = ?
 

Attachments

  • Combining two dependant discrete random variables.doc
    27.5 KB · Views: 194
Physics news on Phys.org
  • #2
You've described an interesting type of problem. This general type of problem is "ill posed", meaning that there are examples of it that have infinitely many solutions. However, ill posed problems arise in many real world situation, such as in mathematics of computing CAT scans, MRI scans etc., so you shouldn't let the ill posed nature of the problem deter you from thinking about it if you find it interesting.

To solve for the joint probability distribution (or determine that there are no solutions or infinitely many solutions), set up the simultaneous equations that the entries in the joint probability table must satisfy. Each entry in the joint probability table is an unknown. The fact that each row sum is known gives you an equation for each row. Likewise the totals for each column give you an equation for each column.

You need to clarify what you mean by "the correlation function". If you mean the line that is computed by doing linear regression ( to get a least squares fit), there is some ambiguity about that line. The line computed by treating B as the independent variable is not the same as the line you get by treating A as the independent variable. There is also a method called "total least squares" that fits a regression line that may be different from both the aforementioned lines. (If you intended to say "correlation coefficient", that is a single quantity, not a line. Likewise, the "covariance" of A and B is not a line.)

How you define the "correlation function" will give you more equations for the unknowns in the joint probability table.

You may find that in some cases, the simultaneous equations have no solution and in some cases they may have infinitely many solutions.

As to the probability distribution for the quantity AB, it would be defined by a table that gave all the possible values of AB and their probabilities. It would not list a value twice. So if your "joint value" table for AB had several entries all equal to the same number, then the final table for the random variable AB would list that number as a value only once. The probability of that vaue would be the sum of all probabilities in the joint distribution table that corresponded to that "joint vaue".
 
Last edited:
  • #3
If you can use the normal approximation to your two distributions and know the correlation [itex]\rho[/itex], you should be able to use the characteristic function:

[tex]\phi(t_{1},t_{2})=exp[i(t_{1}\mu_{1}+t_{2}\mu_{2})-1/2(\sigma_{1}^{2}t_{1}^{2}+2\rho \sigma_{1} \sigma_{2} t_{1}t_{2}+\sigma_{2}^{2} t_{2}^{2})][/tex]
 
Last edited:

1. What is a combination of two dependent discrete random variables?

A combination of two dependent discrete random variables is a statistical concept that involves analyzing the relationship between two variables that have a limited number of possible outcomes. These variables are considered dependent because the outcome of one variable affects the outcome of the other.

2. How do you determine the probability distribution of a combination of two dependent discrete random variables?

To determine the probability distribution of a combination of two dependent discrete random variables, you need to first understand the relationship between the two variables. Then, you can use mathematical formulas or statistical software to calculate the joint probability distribution, which shows the likelihood of each possible combination of outcomes for the two variables.

3. Can you give an example of a combination of two dependent discrete random variables?

One example of a combination of two dependent discrete random variables is the number of heads and tails when flipping two coins. The outcome of one coin affects the outcome of the other, as the number of heads and tails must add up to two. This relationship can be represented by a joint probability distribution.

4. How is the covariance of two dependent discrete random variables calculated?

The covariance of two dependent discrete random variables is calculated by taking the sum of the products of the differences between each possible outcome and the mean of each variable, multiplied by their respective probabilities. This measures the degree to which the two variables vary together.

5. What is the significance of studying the combination of two dependent discrete random variables?

Studying the combination of two dependent discrete random variables allows us to better understand the relationship between these variables and make predictions about their joint behavior. This can be useful in various fields, such as finance, economics, and biology, where variables are often dependent and have discrete outcomes. It also helps in decision-making and risk analysis.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
30
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
6
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
954
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
477
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
899
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
2K
Back
Top