Prem1998 said:
So, since we frequently encounter distributions of the form p(x)=Ne^-(ax-b)^2 so it is convenient to use s=1/a or s=(summation[X(mean)-x(i)]^2) and c=b/a or c= X(mean).
Hold up ... not finished yet.
Once deciding on ##c=b/a## and ##s=1/a^2## ... we could use an understanding of how to extract those parameters from an arbitrary equation.
To get c out of p(x), it turns out we need to do ##c = \int_\infty xp(x)\; dx## ... which you see from the previous posts is the definition of the mean.
So for the gaussian, the intuitive estimate for the central value is exactly the mean.
To get s out, you have to do: ##s = 2\int_{\infty} (x-\mu)^2p(x)\; dx## ... from prev posts the RHS is twice the definition of the variance.
So ##s=2\sigma^2## ... now do you see how mean and standard deviation come about?
Since the normal distribution is so very very common, this way of estimating central values and spread is very useful a lot of the time.
But it is not always useful ... which you will learn when you get to the weirder distributions. Right now it looks like you are just starting out on your journey and you are asking about stuff from much farther on. The best I can do is the picture postcard version.
And, these two definitions are good for comparing and combining the normal distributions. Maybe, they are good for combining because of the property s(x)^2=s(y)^2+s(z)^2 pointed out by Stephen Tashi but I don't agree with the comparing part. In comparing, these quantities are just to compare which distribution is more spread, right?
Wrong. We don't use the parameters just to describe the distribution, we use them to do things.
Example:
When you do an experiment to measure the value of something, you do not get a single exact value (in fact, that would be suspicious). Instead you get a distribution of values which has a mean and a standard deviation (usually quoted as a value and an uncertainty). You will want to compare that value with what other researchers get for the same measurement, but using different methods. So they get different distributions of values.
You want to be able to say if your measurement agrees with the other researcher's measurements.
You need to use the spread to decide that. It is probably easiest to see by comparing box plots.
What about non-continuous discrete distributions? Is the usual formula also more convenient there or Will all these formulas be equivalent there for calculating spread?
The short answer is "yes" (often) - the reason the normal distribution crops up so much is because when you combine lots of different things that are each distributed randomly, you end up with the result being distributed, well, normally. A normal distribution.
An example of a non continuous distribution would be rolling some dice and adding them up.
One 6-sided die has a flat distribution so the mean and standard deviation don't work well for describing it
... but what if you roll 3x 6-sided dice and add them up?
Draw a histogram of the frequency a number appears vs the number and see what shape you get.
What if you use 100 dice? What if the dice are all different shapes and sizes?
Now google for "mean value theorem".
And what if the frequently encountered distribution in practical applications was Ne^(-k*(x-u)^4) ? Do different formulas become convenient for different curves?
That is speculation - the reality is that it is not frequently encountered. But let's say you happened to find one and you wanted to describe it:
You'd probably intuitively want to use s=1/k for a spread estimate ... but IRL we would express 1/k in terms of the defined variance since that is a standard.
You notice I keep calling "s" a spread
estimate ... ie it is not to be considered an exact value or a specific property of something in real life.
It also means that any method that gets sort-of the same thing also counts as well as a spread estimate.
So, usually, the variance works just fine anyway.
Go back to the examples for central value and spread ... think about why you'd want to pick different ones.
Just for central value - there are lots of ways. How do you decide which way to choose just for that one?
For discrete distribution, which you seem to be more familiar with, you know that the average is not always the best approach for finding a central value, you are often better to use the median instead (and the interquartile range is better as an estimate of spread). Unless you haven't got as far as box and whisker plots yet?
So the short answer is "yes" you use the estimate that is best for the situation and what you want it to tell you.
It happens to be (see "mean value theorem") that the gaussian form is extremely common.
In physics though there are other very common distributions ... ie the laplace distribution.
That one has form ##p(x) = \frac{1}{2b}e^{-|x-\mu|/b}## ... here the parameter "b" (called the diversity) is a handy measure of spread ...
We encounter laplace distribution is any particle detector ... the detected energy of the particle is just about always distributed like that and the diversity is a characteristic of the detector.
Another example, occurring whenever something is being counted over time, is the Poisson distribution ... for large numbers it is well approximated by a normal distribution but for small number is is not - so other things need to be done to deal with them.
Those "other things" are college level, will be covered in your course in due time.
Off your questions, you seem to be starting secondary school statistics and have just covered the concept of discrete samples.
Right now they are just teaching you definitions - this part of the course is aimed at people who won't continue with the subject.