chiro said:
Hey dydxforsn.
I don't know about the +- but I'm guessing it's just a natural convention that's been used and still is used.
heh, I didn't realize there was a ± button in the quick symbols of this forum. I just meant to say that you report experimental values that are determined in a lab setting in your formal lab report as like (if it's a position measurement) 4.0
cm ± 0.2cm (where 0.2cm is the standard deviation of the mean. Or rather, 2 times the standard deviation of the mean usually to get like 95% confidence..)
chiro said:
Another to mention is that the distribution may not be for say a sample or collected data, but for a model and this needs to be taken into account. For example you might want to use Gaussian distributions as noise models for a communication system and there are very good reasons for doing this. I imagine a lot of other similar situations apply.
Yeah that is definitely true, like if you were measuring speeds of molecules in a gas or something. There are definitely those situations.
chiro said:
In terms of actually getting the standard deviation of the mean, there are results like the Central Limit Theorem which says the distribution of the mean of any variable given enough data points is normally distributed. This theorem is actually the foundation for a lot of classical frequentist statistics, that many scientists use.
Using this, you can derive distributions for the mean with respect to the sample mean (mean calculated from your data) and from this you get a distribution of your population mean under certain assumptions.
Okay, this is definitely helpful, I clearly need to delve into the Central Limit Theorem some more, I'll look into this. This is definitely along the lines of what I'm looking for. It's this "distribution for the mean" that I'm looking for the proof of (\frac{\sigma}{\sqrt{n}}), seeing as how σ
m along with the average value (which we presumably already know to be the average value of the sample we've already obtained) completely define the theoretical distribution of average (true) values.)
chiro said:
The actual understanding behind this involves the CLT, Random Variables, Transformation of PDF's amongst other things and if you want to understand this, then open up a solid probability/statistics book and take a look.
If you have specific questions, post them in this forum and I'm sure you will get a decent answer that is deeper and more conceptual than what you are getting in your coursework.
I definitely appreciate your response Chiro, it's helped. I may have to go to the university library and read up on some statistics books on the subject, but I was always afraid that if I got a math book on statistics that it might stray too far from what I'm looking for if it gets included at all, but I may give it a try soon if I can't clear everything up with this topic. Currently I'm reading
Experiments in Modern Physics by Melissinos, but the section on the subject of this forum topic is a mere appendix topic and I feel could use further explanation.
Parlyne said:
What we're generally interested in, however, is getting the best possible estimate of the value that our measurements are fluctuating around and characterizing how closely that best-fit value approximates the "true" value. As long as the fluctuations in the individual measurements are random, averaging the measured values will tend to cancel the effects of the random fluctuations, which will mean that the average will tend to be closer to the "true" value than any individual measurement will be. Thus, we should expect that a quantity characterizing the typical deviation of the average of a data set from the "true" value will be smaller than that characterizing the typical deviation of an individual measurement from the "true" value. We can see this by considering error propagation. If we take the uncertainty in each measurement to be \sigma (with the presumption that this represents the standard deviation of the measured values), the uncertainty in the average, \frac{1}{N}\sum_{i=1}^Nx_i, will be \sqrt{\sum_{i=1}^N(\frac{\sigma}{N})^2}, which simplifies to \frac{\sigma}{\sqrt{N}}. You should recognize this as the standard deviation of the mean.
That's it! Okay, I see what you're doing, you're simply assuming each measurement has its own error value associated with it, 'σ', and using propagation of errors to find that there is an error in the average value where the average value is simply a function of each of the independant measurements which allows you to do the familiar \delta_{w}^{2} = \sum_{i}{(\frac{\partial w}{\partial x_i}\delta_{x_{i}})}^2 (where w = x
avg = f(x
1, x
2, .., x
i). I think I've been looking at it from this "Maximum Likelihood Method" and thus didn't really see how this came about.
Parlyne said:
Whatever is actually the case about the statistical properties of our measurements, we need to justify the use of the average as the best estimator of the "true" value, as well as the use of error propagation rules. This is usually done using the Method of Maximum Likelihood, which is just a fancy way of saying that we guess at the form of the distribution from which our measured values are sampled, use that to find the form of the probability of data set, and maximize that probability with respect to the parameters of the distribution. That's a little more involved that I want to try to type up here; but, you can find discussions elsewhere which go through this in some detail.
I sort of talked about this in my original post. I've seen such a process, and I believe it explains why the sample average and sample standard deviation can be used to find the population average (which is simply just the sample average according to this "Maximum Likelihood Method") as well as the population standard deviation (\frac{\sigma}{\sqrt{n}}. However, I don't see how this isn't just another assumption. The Maximum Likelihood Method seems to just say "Well, let's just multiply the probability distribution (using the additional assumption that it's a gaussian) at the specific points that we measured together and try to maximize that product. So basically we're assuming that the 10 points we obtained somehow combine to form the highest likelihood 10 point set that we could have sampled." Or something like that... I dunno, everything about error analysis just seems to boil down to assumptions (btw I very much appreciate your reply Parlyne, it was very useful.)
I'll make my ideas on this subject clear at this point to avoid confusion in this topic and be very specific. I came into this topic having 3 specific questions:
1.) How is it that we can assume an underlying gaussian distribution in the spread of our experimentally measured values?
2.) How can we say that from the sample we can find the population distribution with the average value we obtained being equal to the average value of the population as well as the standard deviation we obtained being equal to the standard deviation of the population?
But wait, we're not finding the average value for the real population distribution, we're finding the average value and
its standard deviation from the sample set (so a distribution involving average values, not a distribution for the population of all values possible to experimentally obtain.)
3.) How is it that the Method of Maximum Likelihood isn't just simply another assumption? How can we possibly be certain that are true value is in the range that we specified? This Method of Maximum Likelihood seems like a nastier assumption than the gaussian distribution itself!
Question 2 has been sort of answered for me by Parlyne. I see how the standard deviation of the mean would be related to the standard deviation of the sample by simply just a factor of 1 over the square root of the sample size. And the proof that the average you obtained via the sample is the average that is the center point for the standard distribution of the mean is still completely unknown to me. How do we know that this average is a sufficient center point for the gaussian specified with standard deviation equal to the standard deviation of the mean?
I'm sorry for such a confusing topic, but this whole subject can get pretty long-winded near as I can tell. My only hope is that I have kept my topic and wording not as jumbled as it seems to me.