What is the Difference between Square & Absolute Deviation?”

  • Context: Undergrad 
  • Thread starter Thread starter icystrike
  • Start date Start date
  • Tags Tags
    deviation Statistic
Click For Summary

Discussion Overview

The discussion centers around the differences between square deviation and absolute deviation, particularly in the context of statistical measures. Participants explore the implications of using squared differences versus absolute differences, touching on historical preferences and mathematical properties.

Discussion Character

  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants note that the square of the difference between a data point and the mean is used to avoid negative values, while others question why absolute values are not used instead.
  • It is highlighted that the measures based on squared deviations and absolute deviations are not equivalent, as indicated by the expressions \(\sqrt{\sum(x-\mu)^2} \ne \sum |x - \mu |\).
  • Some participants argue that the preference for squared deviations is historically rooted in the assumption of normal distribution of data, which aligns with Gaussian noise.
  • An analogy is drawn between using squared distances in statistics and the Pythagorean theorem, suggesting a conceptual similarity in why squares are used.
  • There are claims that squared distances are preferred due to their continuity, while absolute distances have discontinuities that complicate optimization processes.
  • However, some participants counter that the absolute distance function is not problematic in statistics, as other measures like the median and median deviation utilize absolute values.
  • One participant mentions a third measure involving the maximum absolute deviation, indicating that there are multiple approaches to this topic.
  • There are discussions about the lack of a derivative for the absolute value function, with varying opinions on its significance in statistical contexts.

Areas of Agreement / Disagreement

Participants express differing views on the appropriateness and implications of using squared versus absolute deviations. There is no consensus on which measure is superior or more applicable in all contexts, and the discussion remains unresolved.

Contextual Notes

Some participants point out limitations regarding the assumptions made about data distributions and the mathematical properties of the functions involved, but these remain unresolved within the discussion.

icystrike
Messages
444
Reaction score
1
I went for a lecture and the lecturer said that the square of the difference between the x sub i and the mean is the take precaution of the negative value. This has been bugging me , i was wondering why don't they just take absolute because there is a difference between :
[tex]\sqrt{\frac{\sum(x-\mu)^2}{f}}[/tex] and[tex]\frac{\sum \left|(x-\mu)\right|}{f}[/tex]
 
Physics news on Phys.org
Yes, there is a difference, as

[tex] \sqrt{\sum(x-\mu)^2} \ne \sum |x - \mu |[/tex]

There is actually quite a history about whether a measure based on

[tex] \sqrt{\frac{\sum (x-\mu)^2 }{f}}[/tex]

or

[tex] \sqrt{\frac{\sum |x-\mu|}{f}}[/tex]

should be used. Basically, the measure based on the sum of squared deviations won out because, statistically, when it is assumed that the data are drawn from a normal distribution (equivalently, when it is assumed the random noise is Gaussian).
 
statdad said:
Basically, the measure based on the sum of squared deviations won out because, statistically, when it is assumed that the data are drawn from a normal distribution (equivalently, when it is assumed the random noise is Gaussian).

Thanks for your help :smile:
Random noise, i got to check this out !
 
heh, i remember my stats lecturer said that too.
an analogy can be drawn with why we take the squares of the sides (pythagoras) to work out the hypotenuse and not the absolute value.
 
The squared distance is also used because it is continuous, where the absolute distance function has a discontinuity. This is a big problem in optimization.
 
daviddoria said:
The squared distance is also used because it is continuous, where the absolute distance function has a discontinuity. This is a big problem in optimization.

Not really the case in statistics - the median, median deviation, and other procedures use the absolute value.
 
daviddoria said:
The squared distance is also used because it is continuous, where the absolute distance function has a discontinuity. This is a big problem in optimization.
The absolute distance function does not have a derivative at a point. There is no discontinuity.
 
statdad said:
Yes, there is a difference, as

[tex] \sqrt{\sum(x-\mu)^2} \ne \sum |x - \mu |[/tex]

There is actually quite a history about whether a measure based on

[tex] \sqrt{\frac{\sum (x-\mu)^2 }{f}}[/tex]

or

[tex] \sqrt{\frac{\sum |x-\mu|}{f}}[/tex]
When you sum the absolute values, you should not have a square root.

should be used. Basically, the measure based on the sum of squared deviations won out because, statistically, when it is assumed that the data are drawn from a normal distribution (equivalently, when it is assumed the random noise is Gaussian).

There is a third used occasionally:
[tex]\frac{max |x-\mu|}{f}[/tex]

The end of your last sentence seems to be missing!
 
Halls, i wish i had your proof-reading skills. Thanks for catching my missed comment.

You are also correct that the absolute value expression has no derivative, but again, for statistics, I'd add that really isn't a problem.

Why did I miss the unneeded square root? Let me know when you figure it out, because I can't.
 

Similar threads

  • · Replies 42 ·
2
Replies
42
Views
6K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 15 ·
Replies
15
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K