Right way to account for non-statistical uncertainties

  • Context: Undergrad 
  • Thread starter Thread starter BillKet
  • Start date Start date
  • Tags Tags
    Uncertainties
Click For Summary

Discussion Overview

The discussion revolves around how to account for non-statistical uncertainties in data from a spectroscopy experiment. Participants explore the implications of incorporating these uncertainties alongside statistical errors when fitting data to a Lorenz distribution.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested

Main Points Raised

  • One participant describes their approach to fitting spectroscopy data, considering statistical errors as Poisson errors and proposing to add non-statistical uncertainties to both x and y variables.
  • Several participants suggest that uncertainties should be squared to obtain variances and then added, assuming independence.
  • There is confusion about what specific values should be squared and how to combine statistical and non-statistical uncertainties.
  • Participants discuss the distinction between known uncertainties and known error values, with one participant providing an example of laser intensity specifications.
  • One participant questions the independence of the y-axis data, suggesting that the counts may depend on both laser intensity and frequency.
  • Another participant references the NIST guide on uncertainty, emphasizing the importance of terminology and the correct assessment of uncertainties.
  • There is a debate about whether the term "non-statistical uncertainties" is appropriate for the discussed uncertainties and how to combine them with statistical uncertainties.

Areas of Agreement / Disagreement

Participants express differing views on how to properly account for non-statistical uncertainties, with no consensus reached on the correct methodology for combining these uncertainties with statistical errors.

Contextual Notes

Participants highlight the need for clarity regarding the definitions of uncertainties and errors, as well as the assumptions underlying their calculations. The discussion reflects varying interpretations of how to handle uncertainties in experimental data.

BillKet
Messages
311
Reaction score
30
Hello! I have data from a spectroscopy experiment of the form ##(\lambda,N)## where ##\lambda## is the laser wavelength and ##N## is the number of photon counts for that given wavelength. The structure of the obtained spectrum is pretty simple, having a few peaks that need to be fitted with a Lorenz distribution. If I account only for the statistical nature of the experiment, the wavelength can be considered fixed (no errors associated to it) while the number of counts would have a Poisson error i.e. ##\sqrt{N}##. Doing the fit like this works well and I get reasonable values (using available fitting packages like the ones in Python for chi-square minimization). However I would like to now add the effects of the non-statistical uncertainties. On the x-axis, they come from the error on the laser frequency, ##\Delta \lambda## and on the y-axis from the uncertainty in the laser intensity ##\Delta I##. For the y-axis we can assume that number of counts is proportional to the intensity so ##\Delta N_I = \frac{\Delta I}{I}N## with ##\Delta I## and ##I## being known. How should I include these error in my analysis? What I am thinking to do is to just add these error to the values I have, so for the x-axis I have for each point ##\lambda \pm \Delta \lambda## and for the y-axis ##N \pm \sqrt{N+(\frac{\Delta I}{I}N)^2}##, where for the counts the first term under the radical is the statistical error (which was there before, too), while the second one is the non-statistical one. So in the end I have a data set with error on both x and y variable. I can also easily fit this using Python, but is my approach correct? Am I accounting the right way for these extra errors? Thank you!
 
Physics news on Phys.org
Usually you should square them to get a variance. Then add the variances. (Assuming independence)
 
Dale said:
Usually you should square them to get a variance. Then add the variances. (Assuming independence)
Thank you for your reply. What should I square?
 
BillKet said:
Thank you for your reply. What should I square?
The standard uncertainty (standard deviation) of the source of uncertainty. For instance, for ##\Delta \lambda## the laser manufacturer probably lists a specification.
 
Dale said:
The standard uncertainty (standard deviation) of the source of uncertainty. For instance, for ##\Delta \lambda## the laser manufacturer probably lists a specification.
Sorry, I am a bit confused. Yes, the non-statistical error ##\Delta \lambda## and ##\Delta I## are known, as well as the ##I## itself (the central value). I am not sure what I should add together. For the x-axis ##\Delta \lambda## is the only source of error, shouldn't I keep it just like that i.e. ##\lambda \pm \Delta \lambda##? And for the y-axis I added them in quadrature and took the square root, as in the original post. What am I missing? Thank you!
 
BillKet said:
I am not sure what I should add together. For the x-axis Δλ is the only source of error, shouldn't I keep it just like that i.e. λ±Δλ?
I didn’t realize it was the only source if error. So indeed the total uncertainty is ##\sqrt{\Delta \lambda^2+0^2}=\Delta \lambda##
 
Dale said:
I didn’t realize it was the only source if error. So indeed the total uncertainty is ##\sqrt{\Delta \lambda^2+0^2}=\Delta \lambda##
And is the one for the y-axis correct, too?
 
Is the y-axis data completely independent?
 
jim mcnamara said:
Is the y-axis data completely independent?
Do you mean if the counts in each bin are independent of the rest? I think so (not sure what independent would mean here). But the number of counts at a frequency depends on the laser intensity and the laser frequency, so I don't think measurement in a bin can affect another bin. Please let me know if this is not what you meant.
 
  • #10
BillKet said:
Yes, the non-statistical error ##\Delta \lambda## and ##\Delta I## are known,
I think you should clarify this. Are these "uncertainties" or are they known? Do you mean that they were errors in the sense of not controlled but you know what they were or that they really are unknown?
 
  • #11
FactChecker said:
I think you should clarify this. Are these "uncertainties" or are they known? Do you mean that they were errors in the sense of not controlled but you know what they were or that they really are unknown?
They are known. They are provided by the manufacturer of the laser system.
 
  • #12
BillKet said:
They are known. They are provided by the manufacturer of the laser system.
Are they known uncertainties or known error values?
 
  • #13
FactChecker said:
Are they known uncertainties or known error values?
I am not sure I understand the difference between the two. For example for the laser intensity (which is actually the power of the laser), the specification is ##20 \pm 0.1## mW. So in the case above, ##I=20## and ##\Delta I = 0.1##
 
  • #14
BillKet said:
I am not sure I understand the difference between the two. For example for the laser intensity (which is actually the power of the laser), the specification is ##20 \pm 0.1## mW. So in the case above, ##I=20## and ##\Delta I = 0.1##
Then you are using ##\Delta I## to represent a range of possible, but unknown, errors. That is what I was wondering. This is different from a known error. That means that the true error is a [EDIT] random, unknown, variable uncertainty (see @Dale 's like in post #16) and should be treated that way.
 
Last edited:
  • #15
FactChecker said:
Then you are using ##\Delta I## to represent a range of possible, but unknown, errors. That is what I was wondering. This is different from a known error. That means that the true error is a random, unknown, variable and should be treated that way.
So how should I combine it with the statistical ##\sqrt{N}## error? Also could you please explain what you mean by the difference between known uncertainty and known error? I am not sure what you mean by "a range of possible, but unknown, errors". Isn't this what any error is i.e. you know a central value, but the true value can be anywhere within a range given by the error associated to that value (I think I missunderstand what exactly you mean by error in this context). Thank you!
 
  • #16
BillKet said:
I am not sure I understand the difference between the two.
The difference is that one source of uncertainty is evaluated through statistical means and the other is not.
FactChecker said:
Then you are using ΔI to represent a range of possible, but unknown, errors. That is what I was wondering. This is different from a known error. That means that the true error is a random, unknown, variable and should be treated that way.
The correct way to assess and report uncertainty is given by the NIST guide on uncertainty:

https://www.nist.gov/pml/nist-technical-note-1297

Note that the term error is no longer recommended. They recommend using the term uncertainty instead
 
  • Informative
Likes   Reactions: FactChecker
  • #17
Dale said:
The difference is that one source of uncertainty is evaluated through statistical means and the other is not. The correct way to assess and report uncertainty is given by the NIST guide on uncertainty:

https://www.nist.gov/pml/nist-technical-note-1297

Note that the term error is no longer recommended. They recommend using the term uncertainty instead
Thanks for your reply! So I am even more confused now. In the title of my post I used the term "non-statistical uncertainties". Is that the right term to use for my case (for ##\Delta \lambda## and ##\Delta I##)? Also, I thought (and I think I have seen this in all the experimental papers I read), that the statistical errors/uncertainty and the non-statistical ones (which in most papers they are still called systematics) should be added in quadradure, as they are independent i.e. the values for ##\Delta I## is the same no matter how many counts we get in a bin (so no matter the value of the statistical error ##\sqrt{N}##). Also, as far as, I understand this is what it is mentioned in the NIST link you sent me, too. So is the thing I did in the original post correct, or am I still missing something (I feel like I am being told that what I did is not right, but I am not sure what is wrong about it)?
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 28 ·
Replies
28
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 30 ·
2
Replies
30
Views
4K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 22 ·
Replies
22
Views
2K
  • · Replies 18 ·
Replies
18
Views
3K
Replies
5
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K