The covariance of a sum of two random variables X and Y

Click For Summary

Discussion Overview

The discussion revolves around the covariance of a sum of two random variables, X and Y, particularly in relation to a third variable K. Participants explore the properties of covariance, independence, and correlation among these variables, addressing both theoretical and practical implications.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants propose that the covariance of Z (where Z = X + Y) with K can be expressed as the sum of the covariances of X with K and Y with K, regardless of independence.
  • Others argue that the initial summary of the covariance relationship was incorrect and should clarify that Z = X + Y is the basis for the discussion.
  • It is noted that covariance is linear in each variable, and the independence of X and Y does not negate their correlation with K.
  • Some participants question how X and Y can be independent if both are correlated with K, suggesting a contradiction in the definitions.
  • A later reply discusses the implications of K being a function of X and Y, suggesting that various forms of K could yield different correlation behaviors.
  • One participant mentions the bilinearity of covariance coefficients, indicating a potential limitation in applying rules to nonlinear functions of X and Y.
  • There is a discussion about the correlation of Z with K, noting that while Z can be correlated with K, this correlation is not necessarily perfect due to the independent nature of X and Y.

Areas of Agreement / Disagreement

Participants express differing views on the relationship between independence and correlation, particularly regarding how X and Y can be independent while also being correlated with K. The discussion remains unresolved with multiple competing perspectives on the covariance properties.

Contextual Notes

Some participants highlight the need for clarity regarding the definitions of independence and correlation, as well as the implications of different forms of K on the covariance relationships discussed.

Ad VanderVen
Messages
169
Reaction score
13
TL;DR
Suppose you have three random variables X, Y and K. Suppose X and Y are independent, but each correlated with K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?
Suppose X and Y are random variables. Is it true that

Cov (Z,K) = Cov(X,K)+Cov(Y,K) where Z=X+Y?
 
Last edited:
Physics news on Phys.org
The summary:

Suppose you have three random variables X, Y and K. Suppose X and Y are independent, but each correlated with K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?

is incorect and should be:

Suppose you have three random variables X, Y and K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?
 
Covariance is linear in each variable. The random variables do not have to be independent. See this.
 
I remember something along the lines that correlation( covariance?) was an inner -product in some space of Random Variables. I guess we have Cov<X,X>=Var(X)=norm(X)?
 
How can X and Y be independent (cov=0) if they are both correlated with K?
 
  • Like
Likes   Reactions: Filip Larsen
BWV said:
How can X and Y be independent (cov=0) if they are both correlated with K?
Suppose X and Y are any two independent variables and K = X+Y. Then cov(X,K) = cov(X, X+Y) = cov(X,X) + cov(X,Y) = cov(X,X) + 0 > 0.
 
  • Like
Likes   Reactions: Filip Larsen and BWV
Let X and Y be independent uniformly distributed RVs and let K be some function of them.

Z=X+Y

But k could be anything, if it is X+Y it obviously perfectly correlated, but is could also be X-Y, or sin(x+y) etc, no?
 
I believe the best we can use is bilinearity of coefficients, i.e.,
Cov(aX, bY)=abCov(X,Y).
But you're right, beyond that, I think there are no rules for f with f=f(X,Y).

Edit: I suspect the answer here may fall under propagation of errors/uncertainty







Wikipedia
Search

Propagation of uncertainty​


Article Talk

For the propagation of uncertainty through time, see Chaos theory § Sensitivity to initial conditions.
In statistics, propagation of uncertainty (or propagation of error) is the effect of variables' uncertainties (or errors, more specifically random errors) on the uncertainty of a function based on them. When the variables are the values of experimental measurements they have uncertainties due to measurement limitations (e.g., instrument precision) which propagate due to the combination of variables in the function.
The uncertainty u can be expressed in a number of ways. It may be defined by the absolute error Δx. Uncertainties can also be defined by the relative errorx)/x, which is usually written as a percentage. Most commonly, the uncertainty on a quantity is quantified in terms of the standard deviation, σ, which is the positive square root of the variance. The value of a quantity and its error are then expressed as an interval x ± u. However, the most general way of characterizing uncertainty is by specifying its probability distribution. If the probability distribution of the variable is known or can be assumed, in theory it is possible to get any of its statistics. In particular, it is possible to derive confidence limits to describe the region within which the true value of the variable may be found. For example, the 68% confidence limits for a one-dimensional variable belonging to a normal distribution are approximately ± one standard deviation σ from the central value x, which means that the region x ± σ will cover the true value in roughly 68% of cases.
If the uncertainties are correlated then covariance must be taken into account. Correlation can arise from two different sources. First, the measurement errors may be correlated. Second, when the underlying values are correlated across a population, the uncertainties in the group averages will be correlated.[1]
In a general context where a nonlinear function modifies the uncertain parameters (correlated or not), the standard tools to propagate uncertainty, and infer resulting quantity probability distribution/statistics, are sampling techniques from the Monte Carlo method family.[2] For very expansive data or complex functions, the calculation of the error propagation may be very expansive so that a surrogate model[3] or a parallel computing strategy[4][5][6] may be necessary.
In some particular cases, the uncertainty propagation calculation can be done through simplistic algebraic procedures. Some of these scenarios are described below.

Linear combinations​

Non-linear combinations​


Example formulae​

Example calculations​


See also​















References​



















Further reading​







External links​







Last edited 1 month ago by Hellacioussatyr
Wikipedia

 
Last edited:
  • #10
BWV said:
Let X and Y be independent uniformly distributed RVs and let K be some function of them.

Z=X+Y

But k could be anything,
Well, it can be a lot of things. If you start with any two uncorrelated variables, X and Y, many variables, Z, will have some correlation with both of them.
BWV said:
if it is X+Y it obviously perfectly correlated,
Not perfectly. X's correlation with X+Y is not perfect. The independent Y variable term prevents that.
BWV said:
but is could also be X-Y, or sin(x+y) etc, no?
Yes, there are a lot of examples where there is clearly a connection. There are also a lot of examples where two uncorrelated variables, X and Y might be correlated with a third variable, K, with no apparent reason.
 
  • #11
FactChecker said:
Well, it can be a lot of things. If you start with any two uncorrelated variables, X and Y, many variables, Z, will have some correlation with both of them.

Not perfectly. X's correlation with X+Y is not perfect. The independent Y variable term prevents that.

Yes, there are a lot of examples where there is clearly a connection. There are also a lot of examples where two uncorrelated variables, X and Y might be correlated with a third variable, K, with no apparent reason.
Was thinking of the correlation of z and k per the OP, obviously if Z=K the correlation is 1
 

Similar threads

  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 30 ·
2
Replies
30
Views
4K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
Replies
4
Views
1K
  • · Replies 9 ·
Replies
9
Views
2K
Replies
13
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K