The Significance Filter: How selective publishing biases results

Click For Summary
SUMMARY

The discussion centers on the paper "The Significance Filter" which analyzes over a million z-scores from Medline to highlight the biases introduced by selective publishing and p-hacking in research. The authors demonstrate that the expected normal distribution is skewed due to the prevalence of null result suppression, leading to a misrepresentation of statistical significance. Key findings include the sharp edge at z=1.96, indicating a bias in published results, and the implications for meta-analyses that rely solely on published data. The paper emphasizes the need for a more accurate representation of data to avoid misleading conclusions in scientific research.

PREREQUISITES
  • Understanding of z-scores and their significance in statistical analysis
  • Familiarity with p-hacking and its impact on research validity
  • Knowledge of meta-analysis techniques and their reliance on published data
  • Basic comprehension of statistical distributions and their implications in experimental design
NEXT STEPS
  • Research the implications of p-hacking on scientific integrity and publication practices
  • Explore methods for improving transparency in research, such as pre-registration of studies
  • Learn about alternative statistical methods to address biases in published results
  • Investigate the role of open data in enhancing the reliability of meta-analyses
USEFUL FOR

Researchers, statisticians, and academic publishers interested in understanding the effects of selective publishing and improving the integrity of scientific research.

Messages
37,402
Reaction score
14,234
On arXiv: The Significance Filter, the Winner's Curse and the Need to Shrink

It's well known that it is difficult to publish null results in various fields. The authors try to quantify that by analyzing over a million z-scores from Medline, a database for medical publications. A very striking result is figure 1, red lines and numbers were added by me:

zscores.png


The infamous two-sided p<0.05 is at z=1.96. We see a really sharp edge, if the bin edge were at 1.96 instead of 2 it would look even more pronounced.
Should we expect a normal distribution? No. That would be the shape if nothing depends on anything ever. Things do depend on each other, so the tails are larger. It's perfectly fine to start a study where earlier results suggest you'll end up with a z-score around 2 for example. But we certainly should expect a smooth distribution - in a well-designed experiment the difference between z=1.9 and z=2.1 is purely random chance.

If 10 groups take measurements of something that's uncorrelated, one of them finds z>2 by chance and only that one gets published we get a strong bias in the results. p-hacking leads to a similar result as selective publishing. Things look important where the combined measurements would clearly show that there is nothing relevant. Even worse, meta-analyses will only take the published results, calculating an average that's completely disconnected from reality.

The authors quantify some problems that arise from selective publishing.
 
  • Like
Likes   Reactions: jim mcnamara, jrmichler, BWV and 8 others
Physics news on Phys.org
mfb said:
Should we expect a normal distribution? No. That would be the shape if nothing depends on anything ever. Things do depend on each other, so the tails are larger.
Did not read the paper, but its presumably T-stats in the various studies examined? which can have a larger tail, that are then converted to Z-scores in the paper
 

Similar threads

  • · Replies 4 ·
Replies
4
Views
4K
  • · Replies 22 ·
Replies
22
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 2 ·
Replies
2
Views
4K
  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 101 ·
4
Replies
101
Views
15K
  • · Replies 13 ·
Replies
13
Views
3K