Stats: would the sum of the variances be 1 in this case?

Click For Summary

Discussion Overview

The discussion revolves around the relationship between the variances explained by multiple factors in a statistical model, particularly in the context of how these variances relate to the overall variance of a dependent variable V. Participants explore whether the individual variances of factors X1, X2, ..., Xn necessarily sum to 1, depending on their independence and the nature of the function relating them to V.

Discussion Character

  • Debate/contested
  • Mathematical reasoning
  • Conceptual clarification

Main Points Raised

  • Some participants propose that if the factors Xn are independent, the individual variances can sum to 1, while others argue that if they are not independent, the fractions may not be meaningful.
  • A participant questions whether the conclusion about the sum of variances being 1 is specific to linear combinations of Xn or if it applies to any deterministic function of independent Xn.
  • Another participant suggests that the relationship between the variances depends on how the calculated quantity changes with the uncertainties of the Xn, indicating that the function's details are relevant.
  • One participant asserts that the fractions cannot be added if the independent variables are strongly correlated, providing an example where two variables explain the same variance in V.
  • A clarification is made regarding the term "independent variables," emphasizing that it typically refers to statistical independence, which may not align with the initial usage in the discussion.
  • A request for book recommendations on variance and correlation measures is made, indicating a desire for deeper conceptual understanding.

Areas of Agreement / Disagreement

Participants express differing views on whether the variances of the factors can be summed to 1, with some asserting it is possible under certain conditions while others contest this notion, particularly in the presence of correlation among the factors. The discussion remains unresolved regarding the general applicability of the claims made.

Contextual Notes

There are limitations regarding the assumptions about the independence of the factors and the nature of the function relating them to V, which are not fully explored or defined in the discussion.

JesseM
Science Advisor
Messages
8,519
Reaction score
17
Often in empirical studies you see statements that factor X explains some fraction of the variance in some other variable V, and thinking about what this means intuitively made me curious about the following question. Suppose you have a model where the values of some set of factors X1, X2, ..., Xn in a given member of the population, taken together, completely determine that member's value for V with probability 1. And say we look at how much of the variance in V is explained by each factor individually in the population as a whole, and find X1 accounts for a fraction F1 of the variance in V, X2 accounts for F2 of the variance in V, and so forth. Is it necessarily the case here that all the individual variances add to 1, i.e. F1 + F2 + ... + Fn = 1? Or would this depend on the exact nature of the function that takes the values of X1, X2, ..., XN as input and gives you the value for V as output? (and if so, are there some types of simple functions--like if the value of V is just a linear sum of the values of X1, X2, ..., Xn--where it would be the case that the individual variances would add to 1?)
 
Physics news on Phys.org
If the Xn are independent, that approach works and the sum (or quadratic sum, depending on how you define the fraction) is 1. If they are not independent, I don't think fractions would be meaningful.
 
  • Like
Likes   Reactions: FactChecker
mfb said:
If the Xn are independent, that approach works and the sum (or quadratic sum, depending on how you define the fraction) is 1. If they are not independent, I don't think fractions would be meaningful.
Thanks--is this answer specific to the case I mentioned where the value of V is a linear sum of the values of the Xn, or is it general? If it's general, could V be any sort of deterministic function of several independent Xn (including things like arbitrary computational algorithms), or are there still some broad restrictions (say, V being determined by a polynomial equation on the Xn)?
 
It does not depend on the Xn and their combination, just on the change of the calculated quantity if the Xn change within their uncertainties (which then depends on your function, of course, but that detail is not relevant).
 
The short answer is no. You can not add the fractions. Suppose two of your independent variables are strongly correlated. Then they are almost the same variable. If one of those explains most of the variance of V, then the other other will also.
An extreme case is X2=-X1=V. Then both X2 and X1 explain 100% of the variance of V.
Sorry, I did not notice mfb's answer above, which implies the same thing.
 
FactChecker said:
Suppose two of your independent variables are strongly correlated.
Doesn't the term "independent variables" normally mean they are statistically independent of one another, i.e. uncorrelated? mfb already mentioned that the variances only add to 1 "if the Xn are independent", that's what I took "independent" to mean there.
 
JesseM said:
Doesn't the term "independent variables" normally mean they are statistically independent of one another, i.e. uncorrelated? mfb already mentioned that the variances only add to 1 "if the Xn are independent", that's what I took "independent" to mean there.
Oh. Of course, you are right. I used the term "independent" in the non-statistical way (input variables of a function) and I shouldn't have. I should have said "Suppose two of your Xi variables are strongly correlated." Your problem statement does not say the Xs are statistically independent.
 
Can anyone recommend a good book that would cover this, along with other aspects of variance and other measures of correlation like Pearson's and Spearman's that might be helpful in building conceptual intuitions about how to interpret them? (for example, this comment uses a simple example to point out a conceptual difference between Pearson and Spearman correlation measures...for another example of the kind of conceptual intuitions I'm looking for, http://www.ncbi.nlm.nih.gov/books/NBK21866/ on 'heritability', which is just variance due to genes, mentions that 'if the heritability of performance on an IQ test were, say, 0.4, then, if all children had the same developmental and social environment as the “average child,” about 60 percent of the variation in IQ test performance would disappear and 40 percent would remain')
 
Last edited:

Similar threads

  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 8 ·
Replies
8
Views
12K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 5 ·
Replies
5
Views
1K
  • · Replies 13 ·
Replies
13
Views
3K
  • · Replies 2 ·
Replies
2
Views
1K