A p-value is the probability — under the assumption that the null hypothesis is true — of observing data at least as extreme as the actual sample. Small p-values mean the observed data would be unlikely if were true, providing evidence against .
Convention: reject if (commonly ). The threshold is the Type I error rate you accept.
Common misconceptions (drilled by every stats professor):
- is not "the probability that is true."
- is not "the probability the result is due to chance."
- A small doesn't mean a large effect — only an unlikely-under- effect. With huge samples, trivially small effects become "statistically significant."
- is not proof that is true — only insufficient evidence to reject it.
The American Statistical Association (2016) explicitly warned against treating p-values as binary "significant / not" decisions; report effect sizes and confidence intervals alongside.