The University of Southampton
University of Southampton Institutional Repository

Investigating the replicability of the social and behavioral sciences

Investigating the replicability of the social and behavioral sciences
Investigating the replicability of the social and behavioral sciences
Pursuing replicability — independent evidence for previous claims — is important for creating generalizable knowledge1,2. Here we attempted replications of 274 claims of positive results from 164 quantitative papers published from 2009 to 2018 in 54 journals in the social and behavioural sciences. Replications were high powered on average to detect the original effect size (median of 99.6%), used original materials when relevant and available, and were peer reviewed in advance through a standardized internal protocol. Replications showed statistically significant results in the original pattern for 151 of 274 claims (55.1% (95% confidence interval (CI) 49.2–60.9%)) and for 80.8 of 164 papers (49.3% (95% CI 43.8–54.7%)), weighed for replicating multiple claims per paper. We observed modest variation in replication rates across disciplines (42.5–63.1%), although some estimates had high uncertainty. The median Pearson’s r effect size was 0.25 (95% CI 0.21–0.27) for original studies and 0.10 (95% CI 0.09–0.13) for replication studies, an 82.4% (95% CI 67.8–88.2%) reduction in shared variance. Thirteen methods for evaluating replication success provided estimates ranging from 28.6% to 74.8% (median of 49.3%). Some decline in effect size and significance is expected based on power to detect original effects and regression to the mean because we replicated only positive results. We observe that challenges for replicability extend across social–behavioural sciences, illustrating the importance of identifying conditions that promote or inhibit replicability3,4.
0028-0836
143-150
Tyner, Andrew H.
9f1abfac-65f7-468b-bc18-5ced3b20a6ee
Abatayo, Anna Lou
af5d1ec5-f3a1-4b2a-9fee-2d740cb95423
Daley, Mason
fd9b08d5-e7e9-49d3-ab5e-bd55681dc636
Bokhove, Christian
7fc17e5b-9a94-48f3-a387-2ccf60d2d5d8
et al.
Tyner, Andrew H.
9f1abfac-65f7-468b-bc18-5ced3b20a6ee
Abatayo, Anna Lou
af5d1ec5-f3a1-4b2a-9fee-2d740cb95423
Daley, Mason
fd9b08d5-e7e9-49d3-ab5e-bd55681dc636
Bokhove, Christian
7fc17e5b-9a94-48f3-a387-2ccf60d2d5d8

Tyner, Andrew H., Abatayo, Anna Lou and Daley, Mason , et al. (2026) Investigating the replicability of the social and behavioral sciences. Nature, 652 (8108), 143-150. (doi:10.1038/s41586-025-10078-y).

Record type: Article

Abstract

Pursuing replicability — independent evidence for previous claims — is important for creating generalizable knowledge1,2. Here we attempted replications of 274 claims of positive results from 164 quantitative papers published from 2009 to 2018 in 54 journals in the social and behavioural sciences. Replications were high powered on average to detect the original effect size (median of 99.6%), used original materials when relevant and available, and were peer reviewed in advance through a standardized internal protocol. Replications showed statistically significant results in the original pattern for 151 of 274 claims (55.1% (95% confidence interval (CI) 49.2–60.9%)) and for 80.8 of 164 papers (49.3% (95% CI 43.8–54.7%)), weighed for replicating multiple claims per paper. We observed modest variation in replication rates across disciplines (42.5–63.1%), although some estimates had high uncertainty. The median Pearson’s r effect size was 0.25 (95% CI 0.21–0.27) for original studies and 0.10 (95% CI 0.09–0.13) for replication studies, an 82.4% (95% CI 67.8–88.2%) reduction in shared variance. Thirteen methods for evaluating replication success provided estimates ranging from 28.6% to 74.8% (median of 49.3%). Some decline in effect size and significance is expected based on power to detect original effects and regression to the mean because we replicated only positive results. We observe that challenges for replicability extend across social–behavioural sciences, illustrating the importance of identifying conditions that promote or inhibit replicability3,4.

Text
replicability.fullmanuscript - Accepted Manuscript
Restricted to Repository staff only until 1 October 2026.
Available under License Other.
Request a copy

More information

Accepted/In Press date: 19 November 2025
e-pub ahead of print date: 1 April 2026
Published date: 1 April 2026

Identifiers

Local EPrints ID: 510560
URI: http://eprints.soton.ac.uk/id/eprint/510560
ISSN: 0028-0836
PURE UUID: 4519b471-fe0c-4b88-8907-906c65d9e62c
ORCID for Christian Bokhove: ORCID iD orcid.org/0000-0002-4860-8723

Catalogue record

Date deposited: 13 Apr 2026 16:52
Last modified: 14 Apr 2026 01:48

Export record

Altmetrics

Contributors

Author: Andrew H. Tyner
Author: Anna Lou Abatayo
Author: Mason Daley
Corporate Author: et al.

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×