Analyses in HEP and bugs in codes

  • Context: Graduate 
  • Thread starter Thread starter ChrisVer
  • Start date Start date
  • Tags Tags
    Hep
Click For Summary

Discussion Overview

The discussion revolves around the reliability of analyses in High Energy Physics (HEP) when conducted using software that may contain bugs. Participants explore the implications of software errors on published results, the nature of these bugs, and the practices in place to mitigate their impact.

Discussion Character

  • Debate/contested
  • Technical explanation
  • Conceptual clarification

Main Points Raised

  • One participant questions the reliability of an analysis if a bug in the code used for the analysis is discovered after publication.
  • Another participant suggests that the impact of a bug depends on its nature, noting that a bug affecting the tail of a distribution may be significant only if that tail is the focus of the study.
  • It is mentioned that analyses often utilize multiple methods or independent teams to cross-check results, which may help identify significant bugs.
  • Some participants argue that bugs can lead to incorrect results being published, and while some errors are caught, others may go unnoticed, leading to potential inaccuracies in the literature.
  • A participant shares a personal experience of finding a bug that had a minimal impact on results, indicating that not all bugs are significant enough to affect conclusions.
  • Concerns are raised about the existence of errors in widely used software like Geant, with suggestions that experiments take steps to mitigate the impact of such errors.

Areas of Agreement / Disagreement

Participants express a range of views on the reliability of analyses in the presence of software bugs. While some acknowledge that bugs exist and can affect results, others emphasize the robustness of cross-checking methods and the rarity of significant errors leading to errata. The discussion remains unresolved regarding the overall impact of software bugs on published analyses.

Contextual Notes

Participants note that the nature of bugs and their effects can vary widely, and the discussion highlights the complexity of assessing reliability in the context of software used in HEP analyses.

ChrisVer
Science Advisor
Messages
3,372
Reaction score
465
I have a stupid question that has been bugging me for quiet some time and I wouldn't try to post it if I could give myself a reasonable answer, but here it goes:
In order to make an analysis in HEP, people rely on codes/tools/frameworks and stuff like this... Here goes my thinking:
1. Suppose an analysis A1 was published in 2012 using a code X
2. In 2013 a bug is spotted in that code X which needs some fixing and is fixed
3. How reliable is the result of the analysis A1 since it ran under that bug?
Doesn't the idea that no code is perfect and there are always bugs swarming around (And that's why developments are done on those codes even today), make previous year analyses less reliable for not having spotted it?
 
Physics news on Phys.org
The natural (and somewhat unsatisfactory) answer is that it depends on the nature of the bug. If it is a bug in the tail of some distribution it probably is not going to matter much ... unless what you study is exactly that tail.

This is not restricted to HEP. It occurs in other fields too. Just this year it turned out there was a significant bug in the software used for functional MRI studies ...
 
Analyses rarely rely on a single piece of code. A second method is used to cross-check the result of the main method. Sometimes a third method is used. Some analyses even have two teams working completely independent for a while, and comparing their results afterwards.
The worst bugs are those that lead to small deviations. If they lead to large deviations anywhere (and they usually do), they are easy to spot.
 
This is not always true, and often bugs (which are sometimes, but not always) can lead to incorrect results being published.

This is true of experimental and theoretical work.

In some fortunate cases, mistakes are found. For example, someone repeats a calculation and does not find agreement.

You often find these bugs are large enough to warrant an erratum.

The fact that some erratum exist, probably means there are mistakes in published results which are not found...
 
Oh, for sure they exist. I found one example myself when I checked code used in a previous publication. We checked its influence, it was something like 1/100 of the statistical uncertainty - small enough to ignore it (and also so small that the cross-checks didn't catch it). The follow-up analysis with a larger dataset had the bug fixed of course.

Errata in HEP are rare, while at the same time the analyses get checked over and over again. The rate of relevant bugs has to be very low.
 
There are surely errors in Geant (I say this because each new version has corrected errors found in previous versions), and that's pretty much the only program of its scope and kind. The experiments try and mitigate this by
  • Looking at known distributions to ensure that any undiscovered errors are small
  • Using data driven backgrounds whenever feasible
  • Reporting which version was used in the publication so if a serious error were found, the community would know how seriously to take a result
 

Similar threads

Replies
3
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 8 ·
Replies
8
Views
4K
  • · Replies 24 ·
Replies
24
Views
6K
  • · Replies 9 ·
Replies
9
Views
5K
  • · Replies 9 ·
Replies
9
Views
5K
  • · Replies 11 ·
Replies
11
Views
2K