Tracking replicability as a method of post-publication open evaluation

被引:77
作者
Hartshorne, Joshua K. [1 ]
Schachner, Adena [1 ]
机构
[1] Harvard Univ, Dept Psychol, Cambridge, MA 02138 USA
来源
FRONTIERS IN COMPUTATIONAL NEUROSCIENCE | 2012年 / 6卷
关键词
replication; replicability; post-publication evaluation; open evaluation; DISEASE SUSCEPTIBILITY GENE; STATISTICAL POWER ANALYSIS; PSYCHOLOGICAL-RESEARCH; EMPIRICAL-ASSESSMENT; CLINICAL-RESEARCH; BIAS; METAANALYSIS; REPLICATION; ASSOCIATION; DISCRIMINATION;
D O I
10.3389/fncom.2012.00008
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Recent reports have suggested that many published results are unreliable. To increase the reliability and accuracy of published papers, multiple changes have been proposed, such as changes in statistical methods. We support such reforms. However, we believe that the incentive structure of scientific publishing must change for such reforms to be successful. Under the current system, the quality of individual scientists is judged on the basis of their number of publications and citations, with journals similarly judged via numbers of citations. Neither of these measures takes into account the replicability of the published findings, as false or controversial results are often particularly widely cited. We propose tracking replications as a means of post-publication evaluation, both to help researchers identify reliable findings and to incentivize the publication of reliable results. Tracking replications requires a database linking published studies that replicate one another. As any such database is limited by the number of replication attempts published, we propose establishing an open-access journal dedicated to publishing replication attempts. Data quality of both the database and the affiliated journal would be ensured through a combination of crowd-sourcing and peer review. As reports in the database are aggregated, ultimately it will be possible to calculate replicability scores, which may be used alongside citation counts to evaluate the quality of work published in individual journals. In this paper, we lay out a detailed description of how this system could be implemented, including mechanisms for compiling the information, ensuring data quality, and incentivizing the research community to participate.
引用
收藏
页数:14
相关论文
共 82 条
  • [1] Adamic LA, 2008, P 17 INT C WORLD WID
  • [2] A cautionary note on the robustness of latent class models for estimating diagnostic error without a gold standard
    Albert, PS
    Dodd, LE
    [J]. BIOMETRICS, 2004, 60 (02) : 427 - 435
  • [3] [Anonymous], 2009, Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing
  • [4] REPEATED SIGNIFICANCE TESTS ON ACCUMULATING DATA
    ARMITAGE, P
    MCPHERSO.CK
    ROWE, BC
    [J]. JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES A-GENERAL, 1969, 132 : 235 - &
  • [5] Mixed-effects modeling with crossed random effects for subjects and items
    Baayen, R. H.
    Davidson, D. J.
    Bates, D. M.
    [J]. JOURNAL OF MEMORY AND LANGUAGE, 2008, 59 (04) : 390 - 412
  • [6] Bederson B. B., 2010, P GRAPH INT OTT, P39
  • [7] Statistical power and effect sizes of clinical neuropsychology research
    Bezeau, S
    Graves, R
    [J]. JOURNAL OF CLINICAL AND EXPERIMENTAL NEUROPSYCHOLOGY, 2001, 23 (03) : 399 - 406
  • [8] False-positive results in cancer epidemiology: A plea for epistemological modesty
    Boffetta, Paolo
    McLaughlin, Joseph K.
    La Vecchia, Carlo
    Tarone, Robert E.
    Lipworth, Loren
    Blot, William J.
    [J]. JNCI-JOURNAL OF THE NATIONAL CANCER INSTITUTE, 2008, 100 (14): : 988 - 995
  • [9] Positive-outcome bias and other limitations in the outcome of research abstracts submitted to a scientific meeting
    Callaham, ML
    Wears, RL
    Weber, EJ
    Barton, C
    Young, G
    [J]. JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 1998, 280 (03): : 254 - 257
  • [10] STATISTICAL POWER ANALYSIS OF APPLIED PSYCHOLOGICAL-RESEARCH
    CHASE, LJ
    CHASE, RB
    [J]. JOURNAL OF APPLIED PSYCHOLOGY, 1976, 61 (02) : 234 - 237