These slides were presented on November 22 2016 during the Annual Julius Symposium, organised by the Julius Center for Health Sciences and Primary Care, University Medical Hospital Utrecht.
Only a few months ago, the American Statistical Association authoritatively issued an official statement on significance and p-values (American Statistician, 2016, 70:2, 129-133), claiming that the p-value is: “commonly misused and misinterpreted.”
In this presentation I focus on the principles of the ASA statement.
2. About
• statistician by training
• phd (2016): diagnostic research in absence gold standard
(JC)
• post-doc: biostatistics / epidemiological methods (JC)
12. … the p-value fails
“arguably significant” (P = 0.07)
“direction heading to significance” (P = 0.10)
“flirting with conventional levels of significance” (P > 0.1)
“marginally significant” (P ≥ 0.1)
convenient sample from: https://mchankins.wordpress.com/2013/04/21/still-not-significant-2/
listing 509 expressions for non-significant results at α = .05 level (24 October 2016)
13. + 23!!! supplementary files
Wasserstein & Lazar (2016) The ASA's Statement on p-Values:
Context, Process, and Purpose, The American Statistician, 70:2, 129-133
14. A few quotes (1)
“The ASA has not previously taken positions on specific
matters of statistical practice.”
nb. founded in 1839
“Nothing in the ASA statement is new.”
from the ASA Statement
15. A few quotes (2)
“… process was lengthier and more controversial than
anticipated.”
“… the statement articulates in non-technical terms a few select
principles that could improve the conduct or interpretation of
quantitative science, according to widespread consensus in the
statistical community."
from the ASA Statement
18. Why do we need a statement?
‘“It’s science’s dirtiest secret: The ‘scientific method’ of testing
hypotheses by statistical analysis stands on a flimsy
foundation.”’
Quoting Siegfried (2010), Odds Are, It’s Wrong: Science Fails to Face the Shortcomings of Statistics, Science News, 177, 26.
from the ASA Statement: Wasserstein & Lazar (2016) The ASA's Statement on p-Values:
Context, Process, and Purpose, The American Statistician, 70:2, 129-133
19. OK, but why now?
“… highly visible discussions over the last few years”
“The statistical community has been deeply concerned about
issues of reproducibility and replicability …”
from the ASA statement
24. P-value increasingly central in reporting
From: Chavalarias et al. JAMA. 2016;315(11):1141-1148, doi:10.1001/jama.2016.1952
Using text-mining >1.6 million abstracts
25. In the large (‘big’) data era
“With a combination of large datasets, confounding, flexibility in
analytical choices …, and superimposed selective reporting
bias, using a P < 0.05 threshold to declare “success,” ….
means next to nothing.”
From ASA supplementary material, response by Ioannidis.
26. To summarise: why?
• p-values and the P < .05 rule are at the core of inference in
today’s science (social, biomedical, …)
• there is growing concern that these inference are often wrong
• perhaps, if we understand p-values better, we’ll be less
often wrong
28. The statement: 6 principles
1. P-values can indicate how incompatible the data are with a specified
statistical model.
2. P-values do not measure the probability that the studied hypothesis is
true, or the probability that the data were produced by random chance
alone.
3. Scientific conclusions and business or policy decisions should not be
based only on whether a p-value passes a specific threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of an
effect or the importance of a result.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis.
from the ASA statement
29. Statistical model?
• every method of statistical inference relies on a web of
assumptions which together can be viewed as a ‘statistical
model’
• the tested hypothesis is one of these assumptions. Often a
‘zero-effect’ called ‘null hypothesis’
30. About assumptions
the calculation of p-values always relies on assumptions
besides the hypothesis tested. It is easy to ignore/forget those
assumptions while analysing.
Your assumptions are your windows on the world.
Scrub them off every once in a while, or the light
won't come in.
Alan Alda
31. The statement: 6 principles
1. P-values can indicate how incompatible the data are with a specified
statistical model.
2. P-values do not measure the probability that the studied hypothesis
is true, or the probability that the data were produced by random
chance alone.
3. Scientific conclusions and business or policy decisions should not be
based only on whether a p-value passes a specific threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of an
effect or the importance of a result.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis.
from the ASA statement
32. From a probability point of view
p-value*: P(Data|Hypothesis)
is not: P(Hypothesis|Data)
*Somewhat simplified, correct notation would be: P(T(X) ≥ x | Hypothesis)
33. Does it matter?
P(Death|Handgun)
= 5% to 20%*
P(Handgun|Death)
= 0.028%**
* from New York Times (http://www.nytimes.com article published: 2008/04/03/)
** from CBS StatLine (concerning deaths and registered gun crimes in 2015 in the Netherlands)
34. If there only was a way…
P(Data|Hypothesis)
P(Hypothesis|Data)
36. The statement: 6 principles
1. P-values can indicate how incompatible the data are with a specified
statistical model.
2. P-values do not measure the probability that the studied hypothesis is
true, or the probability that the data were produced by random chance
alone.
3. Scientific conclusions and business or policy decisions should not be
based only on whether a p-value passes a specific threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of an
effect or the importance of a result.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis.
from the ASA statement
37. On bright-line rules
“Practices that reduce data analysis or scientific
inference to mechanical “bright-line” rules (such as “p <
0.05”) for justifying scientific claims or conclusions can
lead to erroneous beliefs and poor decision making. A
conclusion does not immediately become “true” on
one side of the divide and “false” on the other.”
from the ASA statement
38. If p ~ .05
D Colquhoun (2014). An investigation of the false discovery rate and the misinterpretation of p-values. R.Soc.opensci.1:140216.
“If you want to avoid making a fool of yourself very often, do not
regard anything greater than p < 0.001 as a demonstration that
you have discovered something”
40. The statement: 6 principles
1. P-values can indicate how incompatible the data are with a specified
statistical model.
2. P-values do not measure the probability that the studied hypothesis is
true, or the probability that the data were produced by random chance
alone.
3. Scientific conclusions and business or policy decisions should not be
based only on whether a p-value passes a specific threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of an
effect or the importance of a result.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis.
from the ASA statement
41. The issue of pre-specified hypotheses
From: http://compare-trials.org/ accessed on November 20 2016
42. Ed Yong (2012). Replication studies: Bad copy, Nature. Data credits to: D Fanelli.
43. Why is this enormous positivity?
If you torture the data long enough,
it will confess to anything
Ronald Coase
besides journal editors requirement for p < .05
44. Multiple (potential) comparisons
aka
- p-hacking
- data fishing
- data dredging
- multiple testing
- multiplicity
- significance chasing
- significance questing
- selective inference
- etc.
45. Selective reporting
“Whenever a researcher chooses what to present based on
statistical results, valid interpretation of those results is
severely compromised if the reader is not informed of the choice
and its basis. Researchers should disclose the number of
hypotheses explored during the study, all data collection
decisions, all statistical analyses conducted, and all p-
values computed. Valid scientific conclusions based on p-
values and related statistics cannot be drawn without at least
knowing how many and which analyses were conducted, and
how those analyses (including p-values) were selected for
reporting.”
from the ASA statement
46. The statement: 6 principles
1. P-values can indicate how incompatible the data are with a specified
statistical model.
2. P-values do not measure the probability that the studied hypothesis is
true, or the probability that the data were produced by random chance
alone.
3. Scientific conclusions and business or policy decisions should not be
based only on whether a p-value passes a specific threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of
an effect or the importance of a result.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis.
from the ASA statement
47. About effect size
• statistical significance does not imply practical importance
• to understand practical importance we need information on
the effect size
• Is the p-value a good measure for effect size?
48. Dance of the p-values
https://www.youtube.com/watch?v=5OL1RqHrZQ8&t=10s
Credits to Professor Geoff Cumming
49. The statement: 6 principles
1. P-values can indicate how incompatible the data are with a specified
statistical model.
2. P-values do not measure the probability that the studied hypothesis
is true, or the probability that the data were produced by random
chance alone.
3. Scientific conclusions and business or policy decisions should not
be based only on whether a p-value passes a specific threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of an
effect or the importance of a result.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis.
from the ASA Statement
50. P-values in isolation
“Researchers should recognize that a p-value without context
or other evidence provides limited information. For example, a
p-value near 0.05 taken by itself offers only weak evidence
against the null hypothesis. Likewise, a relatively large p-value
does not imply evidence in favour of the null hypothesis; many
other hypotheses may be equally or more consistent with the
observed data. For these reasons, data analysis should not
end with the calculation of a p-value when other approaches
are appropriate and feasible.”
from the ASA statement
51. The statement: 6 principles
1. P-values can indicate how incompatible the data are with a specified
statistical model.
2. P-values do not measure the probability that the studied hypothesis
is true, or the probability that the data were produced by random
chance alone.
3. Scientific conclusions and business or policy decisions should not
be based only on whether a p-value passes a specific threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of an
effect or the importance of a result.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis.
from the ASA statement
52. Agreement reached?
“you can believe me that had it been any stronger, then all but
one of the statisticians would have resigned.”
“If only the rest could have agreed with me, we would have a
much stronger statement.”
from SlideShare, by Stephen Senn: P Values and the art of herding cats (accessed on Oct 30 2016)
Stephen Senn, involved in the ASA statement
53. From a practical point of view
if you work with p-values (derived from the 6 ASA principles):
1. think carefully about the underlying assumptions
2. avoid statements about the truth of the tested hypothesis
3. avoid strong statements about effect based solely on p < .
05 or absence of effect based solely on p > .05
4. report no. and sequence of analyses; avoid data torture
5. avoid statements about effect size based on p-value
6. if feasible, use additional information from other inferential
tools
58. Rational for Bayesian inference
the posterior distribution (θ|D) is “more informative” than the
likelihood (D|θ)
However:
“Proponents of the “Bayesian revolution” should be wary of
chasing het another chimera: an apparently universal inference
procedure. A better path would be to promote both an
understanding of various devices in the “statistical toolbox” and
informed judgment to select among these.”
Gigerenzer and Marewski (2015), Surrogate Science: The Idol of a Universal Method for Scientific Inference. Journal of Management
60. The words of the pioneer
No scientific worker has a fixed level of
significance at which from year to year, and in
all circumstances, he rejects hypotheses; he rather
gives his mind to each particular case in the light of
his evidence and his ideas.
Ronald Fisher
61. Many initiatives to improve science…
see: http://www.scienceintransition.nl/english
62. and reduce waste
~ 85% of all health research is being avoidably “wasted”
see also: http://blogs.bmj.com/bmj/2016/01/14/paul-glasziou-and-iain-chalmers-is-85-of-health-research-really-wasted/,
and: Lancet’s 2014 series on increasing value, reducing waste (incl video’s etc.): http://www.thelancet.com/series/research
63. Conclusion
• statistical inference is inherently difficult; we should avoid
making a fool of ourselves too often
• p-values can be useful tools for inference; most often, p-
values should not be the ‘star of the inference show’
• bright line rules such as p < .05 give a false sense of
scientific objectivity
• like to play around with data? Me too! Think twice before you
publish such explorations; if you do, be honest and
transparent in reporting
64. Some random thoughts
• inference is thought as a primarily mathematical or
computational problem, it should not.
• we should ban the term “significant” from scientific output
for describing effects that are accompanied with p < .05.
• in applied statistics education, we should invest more time
in discussing various forms of inference (e.g., Bayesian
inference) and their merits and pitfalls
66. Points for discussion
• is there a need for changing the way we do inference?
• if so, how and what do we change?
• education?
• journals?
• should we downplay the role of p < .05 in scientific output?