Proving Correlation Coefficient Relationships for 'x' and 'y' with Zero Mean

Click For Summary

Discussion Overview

The discussion revolves around the relationships between two random variables, 'x' and 'y', with zero mean, specifically focusing on the correlation coefficient 'c' and the implications of splitting 'x' into correlated and uncorrelated components. Participants explore mathematical expressions and assumptions related to expectations and variances.

Discussion Character

  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • Some participants propose that if 'x' is split into correlated term 'xc' and uncorrelated term 'xu', then the expectations satisfy certain relationships involving the correlation coefficient 'c'.
  • Others question the implicit assumption that the variances of 'x' and 'y' are equal, noting that in their specific cases, \(\overline{x^2} \neq \overline{y^2}\).
  • There is a suggestion that if 'xc' equals 'cy', then the correlation coefficient between 'xc' and 'y' must be either 1 or -1.
  • Participants express confusion regarding the algebraic proof of the uncorrelated nature of 'xu' and 'y', with some asserting that the correlation does not imply a direct linear relationship.
  • Some participants seek clarification on the definition of the correlation coefficient being discussed, questioning whether it refers to Pearson's product-moment correlation or another form.

Areas of Agreement / Disagreement

Participants express differing views on the assumptions regarding variances and the nature of correlation. There is no consensus on the implications of the correlation coefficient or the algebraic proofs related to the uncorrelated terms.

Contextual Notes

Limitations include the ambiguity in the term "correlation coefficient" and the varying assumptions about the relationship between 'x' and 'y'. Some mathematical steps remain unresolved, particularly regarding the expectations and variances of the components.

iVenky
Messages
212
Reaction score
12
Let 'x' and 'y' be two random variables with zero mean.

We find that 'x' is related to 'y' with a correlation coefficient 'c'.

Now let us say we are splitting 'x' into correlated term 'xc'and uncorrelated term 'xu'
Then we have

[tex] x= x_c + x_u<br /> \\<br /> \overline{x^2}= \overline{(x_c + x_u)^2} = \overline{x_c^2} + \overline{x_u^2}<br /> [/tex]
Then does it mean the following is true? -
[tex] <br /> \overline{x_c^2}= \overline{x^2}(|c|^2)<br /> \\<br /> \overline{x_u^2}=\overline{x^2}(1-|c|^2)<br /> [/tex]

If so how would you prove it?
I am sure that the above result is true as I saw it in a book.

Thanks a lot
 
Physics news on Phys.org
Please explain your notation of upper bar.
 
x = cy +(x - cy), where xc = cy.

Implicit is the assumption that |x|2 = |y|2 = 1
 
Last edited:
ssd said:
Please explain your notation of upper bar.

Upper bar indicates expectation E().
 
mathman said:
x = cy +(x - cy), where xc = cy.

Implicit is the assumption that |x|2 = |y|2 = 1
Are you sure about that implicit assumption that |x|2 = |y|2 because

In my problem

[tex] <br /> \overline{x^2} \neq \overline{y^2}<br /> [/tex]

Thanks
 
mathman said:
x = cy +(x - cy), where xc = cy.

Implicit is the assumption that |x|2 = |y|2 = 1

If xc=cy, then correlation coefficient between xc and y is 1 or -1.
 
ssd said:
If xc=cy, then correlation coefficient between xc and y is 1 or -1.

Yes. Good point. Didn't notice.
 
iVenky said:
Let 'x' and 'y' be two random variables with zero mean.

We find that 'x' is related to 'y' with a correlation coefficient 'c'.

Now let us say we are splitting 'x' into correlated term 'xc'and uncorrelated term 'xu'
Then we have

[tex] x= x_c + x_u<br /> \\<br /> \overline{x^2}= \overline{(x_c + x_u)^2} = \overline{x_c^2} + \overline{x_u^2}<br /> [/tex]

I am confused here. How do you take E[xcxu]=0?
 
Last edited:
ssd said:
I am confused here. How do you take E[xcxu]=0?

Because correlated term and uncorrelated term are uncorrelated.
 
  • #10
My turn to guess what the question is,

Let [itex]X[/itex] be a real valued random variable with mean zero and variance [itex]\sigma^2_X[/itex].
Let [itex]Y[/itex] be a real valued random variable with mean zero and variance [itex]\sigma^2_Y.[/itex].

Let the [itex]Cov(X,Y)[/itex] denote the covariance of [itex]X[/itex] and [itex]Y[/itex].

Let [itex]c = \frac {Cov(X,Y)}{\sigma_X \sigma_Y}[/itex].

Let [itex]a = \frac{ Cov(X,Y)}{\sigma^2_Y}[/itex].

Define [itex]X_c = a Y[/itex].

Define [itex]X_u = X - X_c[/itex],

Define [itex]W = X^2[/itex]
Define [itex]W_c = {X_c}^2[/itex]
Define [itex]W_u = {X_u}^2[/itex]

Is it true that the mean of [itex]W_c[/itex] is equal to the sum of the means of [itex]W_c[/itex] and [itex]W_u[/itex] ?


How about that?

Or is the question about sample statistics from random variables rather than about population values?
 
  • #11
But the same question rises...

when

Xc= a Y

Then Xc and Y are completely correlated with a correlation coefficient 1 no matter what the value of a is. (If a is negative then Xc and Y are negatively correlated and if a is positive then they are positively correlated.)
 
  • #12
iVenky said:
Because correlated term and uncorrelated term are uncorrelated.

Can you show this algebraically, please? Long way back I went through descriptive stats. Neither have the references at hand nor my mind readily accepts a linguistic statement like this without proof.

Simple correlation is nothing but degree of linear dependence. The correlated term may have a partially related (if |r|<1 ) linear part and a non-separable non linearly related part. The uncorrelated term may have a similar type of non linearly related part too. There by, the correlated and uncorrelated part might be linearly related.
 
Last edited:
  • #13
ssd said:
Can you show this algebraically, please? Long way back I went through descriptive stats. Neither have the references at hand nor my mind readily accepts a linguistic statement like this without proof.

I don't know how to prove it algebraically basically because I am really confused with how you would represent Xc in terms of Y but I can give you some other kind of proof.

Let's say Xc is correlated with Y which means that if Y increases by some amount Xc will also increase and how it increases depends on the correlation coefficient. Now we know that Xu is uncorrelated with Y which means that changes in Xu is no way related with changes in Y which in turn means that it is no way related to variations in Xc. So we can say that they both are uncorrelated.
 
  • #14
iVenky said:
I don't know how to prove it algebraically basically because I am really confused with how you would represent Xc in terms of Y but I can give you some other kind of proof.

Let's say Xc is correlated with Y which means that if Y increases by some amount Xc will also increase and how it increases depends on the correlation coefficient. Now we know that Xu is uncorrelated with Y which means that changes in Xu is no way related with changes in Y which in turn means that it is no way related to variations in Xc. So we can say that they both are uncorrelated.

Unacceptable.

1/ Xc is correlated with Y which DOES NOT mean that if Y increases by some amount Xc will also always increase. Will always increase if r=1.

2/ "... how it increases depends on the correlation coefficient..."... Not at all, if I interpret the word "how" as the rate of increase.

3/ "Now we know that Xu is uncorrelated with Y which means that changes in Xu is no way related with changes in Y which in turn means that it is no way related to variations in Xc"...
No, not again... Xu is uncorrelated with Y but still changes in Xu may be WELL related with changes in Y, but of course non linearly. Or, even strictly linearly in some part of the domain and in some other part of the domain in opposite direction.

Conclusion: need an algebraic proof before proceeding further.
 
  • #15
iVenky said:
We find that 'x' is related to 'y' with a correlation coefficient 'c'.

The term "correlation coefficient" is ambiguous. Can you specify which meaning of "correlation coefficient" you are using? Do you mean Perason's product-moment correlation ? (http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient)

Or are you talking about the slope of a line in a linear regression?
 
  • #16
Stephen Tashi said:
The term "correlation coefficient" is ambiguous. Can you specify which meaning of "correlation coefficient" you are using? Do you mean Perason's product-moment correlation ? (http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient)

Or are you talking about the slope of a line in a linear regression?

Unless otherwise mentioned, we take "correlation coefficient" as simple correlation coefficient (the prod-moment one by Pearson).
 

Similar threads

Replies
1
Views
4K
  • · Replies 43 ·
2
Replies
43
Views
6K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 16 ·
Replies
16
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 17 ·
Replies
17
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K