Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 4 de 4
Filter
Add more filters











Database
Language
Publication year range
1.
Perspect Psychol Sci ; 6(3): 291-8, 2011 May.
Article in English | MEDLINE | ID: mdl-26168519

ABSTRACT

Statistical inference in psychology has traditionally relied heavily on p-value significance testing. This approach to drawing conclusions from data, however, has been widely criticized, and two types of remedies have been advocated. The first proposal is to supplement p values with complementary measures of evidence, such as effect sizes. The second is to replace inference with Bayesian measures of evidence, such as the Bayes factor. The authors provide a practical comparison of p values, effect sizes, and default Bayes factors as measures of statistical evidence, using 855 recently published t tests in psychology. The comparison yields two main results. First, although p values and default Bayes factors almost always agree about what hypothesis is better supported by the data, the measures often disagree about the strength of this support; for 70% of the data sets for which the p value falls between .01 and .05, the default Bayes factor indicates that the evidence is only anecdotal. Second, effect sizes can provide additional evidence to p values and default Bayes factors. The authors conclude that the Bayesian approach is comparatively prudent, preventing researchers from overestimating the evidence in favor of an effect.

2.
Psychol Methods ; 15(2): 172-81, 2010 Jun.
Article in English | MEDLINE | ID: mdl-20515238

ABSTRACT

The purpose of the recently proposed prep statistic is to estimate the probability of concurrence, that is, the probability that a replicate experiment yields an effect of the same sign (Killeen, 2005a). The influential journal Psychological Science endorses prep and recommends its use over that of traditional methods. Here we show that prep overestimates the probability of concurrence. This is because prep was derived under the assumption that all effect sizes in the population are equally likely a priori. In many situations, however, it is advisable also to entertain a null hypothesis of no or approximately no effect. We show how the posterior probability of the null hypothesis is sensitive to a priori considerations and to the evidence provided by the data; and the higher the posterior probability of the null hypothesis, the smaller the probability of concurrence. When the null hypothesis and the alternative hypothesis are equally likely a priori, prep may overestimate the probability of concurrence by 30% and more. We conclude that prep provides an upper bound on the probability of concurrence, a bound that brings with it the danger of having researchers believe that their experimental effects are much more reliable than they actually are.


Subject(s)
Models, Psychological , Models, Statistical , Probability , Psychology, Experimental/statistics & numerical data , Bayes Theorem , Bias , Data Interpretation, Statistical , Humans , Mathematical Computing , Reproducibility of Results
3.
Psychon Bull Rev ; 16(2): 225-37, 2009 Apr.
Article in English | MEDLINE | ID: mdl-19293088

ABSTRACT

Progress in science often comes from discovering invariances in relationships among variables; these invariances often correspond to null hypotheses. As is commonly known, it is not possible to state evidence for the null hypothesis in conventional significance testing. Here we highlight a Bayes factor alternative to the conventional t test that will allow researchers to express preference for either the null hypothesis or the alternative. The Bayes factor has a natural and straightforward interpretation, is based on reasonable assumptions, and has better properties than other methods of inference that have been advocated in the psychological literature. To facilitate use of the Bayes factor, we provide an easy-to-use, Web-based program that performs the necessary calculations.


Subject(s)
Analysis of Variance , Bayes Theorem , Data Interpretation, Statistical , Mathematical Computing , Psychology, Experimental/statistics & numerical data , Software , Humans , Likelihood Functions , Probability
4.
Psychon Bull Rev ; 16(2): 424-9, 2009 Apr.
Article in English | MEDLINE | ID: mdl-19293117

ABSTRACT

The probability of "replication," P(rep), has been proposed as a means of identifying replicable and reliable effects in the psychological sciences. We conduct a basic test of P(rep) that reveals that it misestimates the true probability of replication, especially for small effects. We show how these general problems with P(rep) play out in practice, when it is applied to predict the replicability of observed effects over a series of experiments. Our results show that, over any plausible series of experiments, the true probabilities of replication will be very different from those predicted by P(rep). We discuss some basic problems in the formulation of P(rep) that are responsible for its poor performance, and conclude that P(rep) is not a useful statistic for psychological science.


Subject(s)
Data Interpretation, Statistical , Probability , Psychology, Experimental/statistics & numerical data , Statistics as Topic , Bias , Humans , Reproducibility of Results
SELECTION OF CITATIONS
SEARCH DETAIL