Our registered report in Advances in Methods and Practices in Psychological Science experimentally tests how statistical significance and hypothesis-consistency influence clinical psychologists' decisions to submit, read, and cite research. Across four within-subjects experiments with 303 researchers, we find a consistent preference for statistically significant abstracts, but no effect of hypothesis-consistency. Deliberation does not attenuate the bias.
| 📄 Paper (AMPPS) | 📝 Preregistration | 💻 GitHub |
Over 60 years ago, Sterling
Previous experimental studies produced mixed evidence—some found preferences for positive results
We address three gaps: (A) within-subjects experiments on publication bias in psychology, (B) experimental research on non-reception (reading and citing), and (C) the role of decision-making processes, drawing on Dual Process Theory
We conducted four online experiments with clinical psychology researchers ($n = 303$ total, ~75 per experiment). In each experiment, participants evaluated 16 fictitious abstracts (from 16 pairs) using a two-response paradigm:
Experiments 1 and 2 assessed publishability (statistical significance and hypothesis-consistency, respectively). Experiments 3 and 4 assessed reading and citation likelihood.
We contacted 33,924 clinical psychology researchers globally; 303 completed the study (1.0% completion rate). Participants were 58% female, 40% male; 27% pre-doctoral, 29% post-doctoral, and 44% professors, from 40+ countries.
The 64 abstract pairs were carefully constructed to be identical except for the experimentally varied result section (significant vs. non-significant, or hypothesis-consistent vs. hypothesis-inconsistent). Participants were explicitly told to assume all studies had sufficient statistical power.
Across all experiments involving statistical significance, non-significant abstracts were consistently rated lower:
Whether an abstract’s results were consistent or inconsistent with its hypothesis had no detectable effect on any outcome:
In most experiments, initial intuitive evaluations were not revised after deliberation. FOR did not systematically mediate response changes. One exception: in Experiment 1, non-significant abstracts lowered FOR ($b = -0.16$, $p = .006$), but higher FOR unexpectedly predicted more positive revision—contrary to our hypothesis.
For metascience: Our within-subjects design provides experimental evidence that statistical significance—but not hypothesis-consistency—is associated with researchers’ evaluations of abstracts for publication, reading, and citation. Effect sizes were small (1–2% of variance), reflecting the multifactorial nature of these decisions, but the pattern was consistent across four experiments.
For decision-making research: The two-response paradigm revealed that deliberation does not attenuate the significance bias. This suggests that the preference for significant results operates as a persistent heuristic rather than a correctable first impression.
For clinical psychology: If non-significant findings are less likely to be read and cited
Limitations: This is a controlled vignette study; how strongly these patterns translate into real-world publication outcomes with actual manuscripts and real incentives remains to be tested in field experiments.
If you find this work useful for your research, please consider citing our paper:
@article{schiekiera2025publication,
title={Publication Bias in Academic Decision Making
in Clinical Psychology},
author={Schiekiera, Louis and Eichel, Kristina and Sachse,
Jacqueline and M{\"u}ller, Sophie P. and He{\ss}elmann,
Felicitas and Niemeyer, Helen},
journal={Advances in Methods and Practices in Psychological Science},
year={2025}
}
📄 Read the paper here: https://doi.org/10.1177/25152459251338393
📝 View the preregistration on OSF: https://osf.io/6tpm7
💻 View the code on GitHub: https://github.com/schiekiera/metascience_experiment_history