The covariance of a sum of two random variables X and Y

  • #1
Ad VanderVen
169
13
TL;DR Summary
Suppose you have three random variables X, Y and K. Suppose X and Y are independent, but each correlated with K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?
Suppose X and Y are random variables. Is it true that

Cov (Z,K) = Cov(X,K)+Cov(Y,K) where Z=X+Y?
 
Last edited:
Physics news on Phys.org
  • #2
The summary:

Suppose you have three random variables X, Y and K. Suppose X and Y are independent, but each correlated with K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?

is incorect and should be:

Suppose you have three random variables X, Y and K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?
 
  • #3
Covariance is linear in each variable. The random variables do not have to be independent. See this.
 
  • #5
I remember something along the lines that correlation( covariance?) was an inner -product in some space of Random Variables. I guess we have Cov<X,X>=Var(X)=norm(X)?
 
  • #6
How can X and Y be independent (cov=0) if they are both correlated with K?
 
  • Like
Likes Filip Larsen
  • #7
BWV said:
How can X and Y be independent (cov=0) if they are both correlated with K?
Suppose X and Y are any two independent variables and K = X+Y. Then cov(X,K) = cov(X, X+Y) = cov(X,X) + cov(X,Y) = cov(X,X) + 0 > 0.
 
  • Like
Likes Filip Larsen and BWV
  • #8
Let X and Y be independent uniformly distributed RVs and let K be some function of them.

Z=X+Y

But k could be anything, if it is X+Y it obviously perfectly correlated, but is could also be X-Y, or sin(x+y) etc, no?
 
  • #9
I believe the best we can use is bilinearity of coefficients, i.e.,
Cov(aX, bY)=abCov(X,Y).
But you're right, beyond that, I think there are no rules for f with f=f(X,Y).

Edit: I suspect the answer here may fall under propagation of errors/uncertainty







Wikipedia
Search

Propagation of uncertainty​


Article Talk

For the propagation of uncertainty through time, see Chaos theory § Sensitivity to initial conditions.
In statistics, propagation of uncertainty (or propagation of error) is the effect of variables' uncertainties (or errors, more specifically random errors) on the uncertainty of a function based on them. When the variables are the values of experimental measurements they have uncertainties due to measurement limitations (e.g., instrument precision) which propagate due to the combination of variables in the function.
The uncertainty u can be expressed in a number of ways. It may be defined by the absolute error Δx. Uncertainties can also be defined by the relative errorx)/x, which is usually written as a percentage. Most commonly, the uncertainty on a quantity is quantified in terms of the standard deviation, σ, which is the positive square root of the variance. The value of a quantity and its error are then expressed as an interval x ± u. However, the most general way of characterizing uncertainty is by specifying its probability distribution. If the probability distribution of the variable is known or can be assumed, in theory it is possible to get any of its statistics. In particular, it is possible to derive confidence limits to describe the region within which the true value of the variable may be found. For example, the 68% confidence limits for a one-dimensional variable belonging to a normal distribution are approximately ± one standard deviation σ from the central value x, which means that the region x ± σ will cover the true value in roughly 68% of cases.
If the uncertainties are correlated then covariance must be taken into account. Correlation can arise from two different sources. First, the measurement errors may be correlated. Second, when the underlying values are correlated across a population, the uncertainties in the group averages will be correlated.[1]
In a general context where a nonlinear function modifies the uncertain parameters (correlated or not), the standard tools to propagate uncertainty, and infer resulting quantity probability distribution/statistics, are sampling techniques from the Monte Carlo method family.[2] For very expansive data or complex functions, the calculation of the error propagation may be very expansive so that a surrogate model[3] or a parallel computing strategy[4][5][6] may be necessary.
In some particular cases, the uncertainty propagation calculation can be done through simplistic algebraic procedures. Some of these scenarios are described below.

Linear combinations​

Non-linear combinations​


Example formulae​

Example calculations​


See also​















References​



















Further reading​







External links​







Last edited 1 month ago by Hellacioussatyr
Wikipedia

 
Last edited:
  • #10
BWV said:
Let X and Y be independent uniformly distributed RVs and let K be some function of them.

Z=X+Y

But k could be anything,
Well, it can be a lot of things. If you start with any two uncorrelated variables, X and Y, many variables, Z, will have some correlation with both of them.
BWV said:
if it is X+Y it obviously perfectly correlated,
Not perfectly. X's correlation with X+Y is not perfect. The independent Y variable term prevents that.
BWV said:
but is could also be X-Y, or sin(x+y) etc, no?
Yes, there are a lot of examples where there is clearly a connection. There are also a lot of examples where two uncorrelated variables, X and Y might be correlated with a third variable, K, with no apparent reason.
 
  • #11
FactChecker said:
Well, it can be a lot of things. If you start with any two uncorrelated variables, X and Y, many variables, Z, will have some correlation with both of them.

Not perfectly. X's correlation with X+Y is not perfect. The independent Y variable term prevents that.

Yes, there are a lot of examples where there is clearly a connection. There are also a lot of examples where two uncorrelated variables, X and Y might be correlated with a third variable, K, with no apparent reason.
Was thinking of the correlation of z and k per the OP, obviously if Z=K the correlation is 1
 

1. What is the covariance of a sum of two random variables X and Y?

The covariance of a sum of two random variables X and Y is given by the formula Cov(X+Y, X+Y). This can be expanded using the bilinearity of covariance as Cov(X, X) + Cov(X, Y) + Cov(Y, X) + Cov(Y, Y). This simplifies to Var(X) + 2Cov(X, Y) + Var(Y), where Var(X) and Var(Y) are the variances of X and Y, respectively, and Cov(X, Y) is the covariance between X and Y.

2. How does covariance help in understanding the relationship between X and Y?

Covariance provides a measure of how much two random variables change together. If the covariance is positive, it indicates that as one variable increases, the other tends to increase as well. A negative covariance indicates that as one variable increases, the other tends to decrease. A covariance of zero suggests that the variables do not exhibit any linear relationship. However, it is important to note that non-zero covariance does not imply causation.

3. What is the difference between covariance and correlation?

Covariance and correlation both measure the linear relationship between two variables. However, covariance is affected by the scale of the variables, meaning that it does not provide a standardized measure of the relationship. Correlation, on the other hand, normalizes covariance by the standard deviations of the variables, providing a dimensionless value that ranges from -1 to 1, which allows for easier comparison across different datasets.

4. Can covariance be used for more than two variables?

Yes, covariance can be extended to more than two variables. For a set of variables, the covariances between each pair of variables can be arranged in a matrix known as the covariance matrix. Each element of this matrix represents the covariance between a pair of variables, and the diagonal elements of the matrix are the variances of each variable.

5. What does a covariance of zero mean in practical terms?

A covariance of zero between two variables indicates that there is no linear relationship between them. However, it's important to note that this does not imply there is no relationship at all; rather, it suggests that any relationship is not linear. Non-linear dependencies or relationships involving more complex interactions might still exist, and other statistical methods may be required to detect these relationships.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
452
  • Set Theory, Logic, Probability, Statistics
Replies
30
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
486
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
12
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
849
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
13
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
551
Back
Top