Intuitively, shouldn't variance always be 0?

  • Thread starter Thread starter samh
  • Start date Start date
  • Tags Tags
    Variance
AI Thread Summary
Variance is not always zero because it measures the average squared deviation of a random variable from its mean, which accounts for both positive and negative differences. The logic that variance should be zero stems from a misunderstanding of the definition; variance applies to random variables, not constants. In a random process, values will fluctuate around the mean, leading to positive squared differences that contribute to a non-zero variance. The squaring of differences ensures that all contributions to variance are positive, reflecting the distribution's nature. Understanding that variance captures the spread of values around the mean clarifies why it cannot be zero in non-constant scenarios.
samh
Messages
46
Reaction score
0
Ok I know that Var[X] = E[(X-E[X])^2]. But I just can't help but think that the variance should always be zero. I think it makes so much sense, but obviously the formula says otherwise... But look, my reasoning seems so perfect:

1) The variance is the expected difference from the mean, squared

2) The expected value of X is the mean

3) So shouldn't we always expect X to be the mean, and so (X - mean)^2 = 0^2 = 0?

But obviously it doesn't work out that way...it's so weird. Does anyone know an intuitive reason why Var[X] shouldn't be zero? What's wrong with my logic?
 
Physics news on Phys.org
samh said:
1) The variance is the expected difference from the mean, squared.

What's wrong with my logic?
The first item is what is wrong with your logic. While the expected difference from the mean is tautologically zero, but that is not what the variance is. The variance is the expected value of the square of the difference between the random variable and the mean.

What does this mean? The difference between a random variable and its mean is identically zero only for a constant random variable. In a way it is a bit silly to even talk about a constant random variable because there is no randomness. The cumulative probability function (CDF) for a constant is the step function. Not a very interesting random variable.

Now consider drawing a set of sample values from a random process whose underlying CDF is a smooth function. Some of these sample values will be above the expected mean, some under the expected mean. While the difference between the ith sampled value and the mean might be positive or negative, the square of this difference is always positive. The mean of a bunch of positive values is positive. It is this mean that forms the variance.
 
Variance is an average of the square of differences of observed values from the mean. It depends on the nature of statistical distribution. In other words, it is a measure of the deviation from the average.
The squaring part makes all the difference because when summing and taking average, we square the difference so that only positive quantities are added. Had we taken -ve sign for values below the mean, we would obviously have obtained the value equal to 0 as you said.
 
Ok I see. I forgot that E[...] is kind of like the average of many repetitions. I see what you guys are saying, it makes sense that the average of a bunch of positive numbers is positive. Thanks :)
 
I was reading documentation about the soundness and completeness of logic formal systems. Consider the following $$\vdash_S \phi$$ where ##S## is the proof-system making part the formal system and ##\phi## is a wff (well formed formula) of the formal language. Note the blank on left of the turnstile symbol ##\vdash_S##, as far as I can tell it actually represents the empty set. So what does it mean ? I guess it actually means ##\phi## is a theorem of the formal system, i.e. there is a...
Back
Top