Blurring the distinctions between p's and α's in psychological research

被引:71
作者
Hubbard, R [1 ]
机构
[1] Drake Univ, Coll Business & Publ Adm, Des Moines, IA 50311 USA
关键词
Fisher; hybrid statistical model; inductive behavior; inductive inference; Neyman-Pearson;
D O I
10.1177/0959354304043638
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Confusion over the reporting and interpretation of results of commonly employed classical statistical tests is recorded in a sample of 1,645 papers from 12 psychology journals for the period 1990 through 2002. The confusion arises because researchers mistakenly believe that their interpretation is guided by a single unified theory of statistical inference. But this is not so: classical statistical testing is a nameless amalgamation of the rival and often contradictory approaches developed by Ronald Fisher, on the one hand, and Jerzy Neyman and Egon Pearson, on the other. In particular, there is extensive failure to acknowledge the incompatibility of Fisher's evidential p value with the Type I error rate, alpha, of Neyman-Pearson statistical orthodoxy. The distinction between evidence (p's) and errors (alpha's) is not trivial. Rather, it reveals the basic differences underlying Fisher's ideas on significance testing and inductive inference, and Neyman-Pearson views on hypothesis testing and inductive behavior. So complete is this misunderstanding over measures of evidence versus error that it is not viewed as even being a problem among the vast majority of researchers and other relevant parties. These include the APA Task Force on Statistical Inference, and those writing the guidelines concerning statistical testing mandated in APA Publication Manuals. The result is that, despite supplanting Fisher's significance-testing paradigm some fifty years or so ago, recognizable applications of Neyman-Pearson theory are few and far between in psychology's empirical literature. On the other hand, Fisher's influence is ubiquitous.
引用
收藏
页码:295 / 327
页数:33
相关论文
共 103 条
[21]   SEQUENTIAL TRIALS, SEQUENTIAL ANALYSIS AND LIKELIHOOD PRINCIPLE [J].
CORNFIELD, J .
AMERICAN STATISTICIAN, 1966, 20 (02) :18-23
[22]   On the logic and purpose of significance testing [J].
Cortina, JM ;
Dunlap, WP .
PSYCHOLOGICAL METHODS, 1997, 2 (02) :161-172
[23]   A primer on the understanding, use, and calculation of confidence intervals that are based on central and noncentral distributions [J].
Cumming, G ;
Finch, S .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 2001, 61 (04) :532-574
[24]  
Daniel L.G., 1998, Research in the Schools, V5
[25]   MISUSE OF STATISTICAL TESTS IN 3 DECADES OF PSYCHOTHERAPY-RESEARCH [J].
DAR, R ;
SERLIN, RC ;
OMER, H .
JOURNAL OF CONSULTING AND CLINICAL PSYCHOLOGY, 1994, 62 (01) :75-82
[26]   Scientific versus statistical inference [J].
Dixon, P ;
O'Reilly, T .
CANADIAN JOURNAL OF EXPERIMENTAL PSYCHOLOGY-REVUE CANADIENNE DE PSYCHOLOGIE EXPERIMENTALE, 1999, 53 (02) :133-149
[27]   SIGNIFICANCE TESTS DIE HARD - THE AMAZING PERSISTENCE OF A PROBABILISTIC MISCONCEPTION [J].
FALK, R ;
GREENBAUM, CW .
THEORY & PSYCHOLOGY, 1995, 5 (01) :75-98
[28]   The fifth edition of the APA Publication Manual:: Why its statistics recommendations are so controversial [J].
Fidler, F .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 2002, 62 (05) :749-770
[29]   Reporting of statistical inference in the Journal of Applied Psychology:: Little evidence of reform [J].
Finch, S ;
Cumming, G ;
Thomason, N .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 2001, 61 (02) :181-210
[30]  
FISHER R, 1955, J ROY STAT SOC B, V17, P69