Weighting data based on the errors

In summary: Thank you for your reply. So you are saying that fitting the background right (i.e. the tail) is more important than the peak itself for a Voigt profile? If that's the case, then using the inverse error as the weight will help, I just thought it makes more sense to focus on fitting the peak right (and extracting the parameters of the peak) rather than the background.
  • #1
kelly0303
561
33
Hello! I have some data (counts) with a Poisson error associated to it and I want to make a fit to the data. I am trying to weight the data inversely proportional to the errors, such that the data points with high errors are less important for the fit. However, using the the error on its own, doesn't seem right. If I have a point with a value ##100 \pm 10## and another one with the value ##10000 \pm 100##, the first one has a smaller error, but the second one should be (I think) more important for the fit, as the relative error is much smaller. So, should I weight each data point by the inverse of its percentage error i.e. the first point would have a weight of ##10##, while the other a weight of ##100##? Is this the right way to do it? Thank you!
 
Physics news on Phys.org
  • #2
Maximum Likelihood Estimation is an approach that may accomplish what you want to do. With the scant details you have given, I can't say whether it would be the best possible approach.
 
  • #3
tnich said:
Maximum Likelihood Estimation is an approach that may accomplish what you want to do. With the scant details you have given, I can't say whether it would be the best possible approach.
Thank you for your reply! To give more details: I have some data points to which I want to fit a Voigt profile + background. For each data point I have the energy, the counts and the error (square root of the number of counts). In principle, I am just using an already build Python fitting module (lmfit), and one of the options for that is "weights". I assumed that I have to use the value of the errors as the weights to the fit (1/errors), I am just not sure if I should use the absolute or relative error (or something totally different).
 
  • #4
I don't know what algorithm lmfit uses, or how it uses the weight data, so I don't really know how to answer your question. Since the only error information you have is derived solely from your count data, it would not provide any more information to the fit algorithm than the count data would by itself. I would be tempted not to enter weight data in that case.

I can tell you that the data with low counts is the important data for fitting a pdf curve, especially when it is a Voigt profile. In that case you don't know whether you have a Cauchy distribution, a Gaussian distribution, or some combination of the two. So estimating the tails of the distribution is the important aspect, and that depends on fitting the experimental data you have for the tails. Your fit algorithm may take that into account or it may not.
 
  • #5
tnich said:
I don't know what algorithm lmfit uses, or how it uses the weight data, so I don't really know how to answer your question. Since the only error information you have is derived solely from your count data, it would not provide any more information to the fit algorithm than the count data would by itself. I would be tempted not to enter weight data in that case.

I can tell you that the data with low counts is the important data for fitting a pdf curve, especially when it is a Voigt profile. In that case you don't know whether you have a Cauchy distribution, a Gaussian distribution, or some combination of the two. So estimating the tails of the distribution is the important aspect, and that depends on fitting the experimental data you have for the tails. Your fit algorithm may take that into account or it may not.
Thank you for your reply. So you are saying that fitting the background right (i.e. the tail) is more important than the peak itself for a Voigt profile? If that's the case, then using the inverse error as the weight will help, I just thought it makes more sense to focus on fitting the peak right (and extracting the parameters of the peak) rather than the background.
 

What is weighting data based on the errors?

Weighting data based on the errors is a statistical technique used to adjust the influence of individual data points in a dataset based on their level of uncertainty or error. This helps to improve the accuracy of the overall analysis and reduce the impact of outliers.

Why is weighting data based on the errors important?

Weighting data based on the errors is important because it allows for a more accurate representation of the data by taking into account the degree of uncertainty associated with each data point. This can lead to more reliable and robust conclusions from the analysis.

How is weighting data based on the errors calculated?

The calculation of weighting data based on the errors depends on the specific statistical method being used. In general, it involves assigning a weight to each data point based on its level of uncertainty, and then using these weights in the analysis to give more weight to the more reliable data points.

What are the benefits of weighting data based on the errors?

Weighting data based on the errors can lead to more accurate and precise results in statistical analyses. It can also help to reduce the impact of outliers, which can skew the results of an analysis. Additionally, weighting data can help to account for any systematic errors or biases in the data.

Are there any limitations to weighting data based on the errors?

While weighting data based on the errors can improve the accuracy of an analysis, it is not a perfect solution. It relies on accurate estimates of the uncertainty associated with each data point, which may not always be available. Additionally, weighting data can also introduce some bias into the analysis, so it is important to carefully consider the appropriate weighting method for each dataset.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
932
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
26
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
28
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
16
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
24
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
8
Views
909
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
1K
Back
Top