Bonferroni correction

In statistics, the Bonferroni correction is one of several methods used to counteract the problem of multiple comparisons.

Background

The Bonferroni correction is named after Italian mathematician Carlo Emilio Bonferroni for its use of Bonferroni inequalities.[1] Its development is often credited to Olive Jean Dunn, who described the procedure's application to confidence intervals.[2][3]

Statistical hypothesis testing is based on rejecting the null hypothesis if the likelihood of the observed data under the null hypotheses is low. If multiple hypotheses are tested, the chance of a rare event increases, and therefore, the likelihood of incorrectly rejecting a null hypothesis (i.e., making a Type I error) increases.[4]

The Bonferroni correction compensates for that increase by testing each individual hypothesis at a significance level of , where is the desired overall alpha level and is the number of hypotheses.[5] For example, if a trial is testing hypotheses with a desired , then the Bonferroni correction would test each individual hypothesis at .

Definition

Let be a family of hypotheses and their corresponding p-values. Let be the total number of null hypotheses and the number of true null hypotheses. The familywise error rate (FWER) is the probability of rejecting at least one true , that is, of making at least one type I error. The Bonferroni correction rejects the null hypothesis for each , thereby controlling the FWER at . Proof of this control follows from Boole's inequality, as follows:

This control does not require any assumptions about dependence among the p-values or about how many of the null hypotheses are true.[6]

Extensions

Generalization

Rather than testing each hypothesis at the level, the hypotheses may be tested at any other combination of levels that add up to , provided that the level of each test is determined before looking at the data.[7] For example, for two hypothesis tests, an overall of .05 could be maintained by conducting one test at .04 and the other at .01.

Confidence intervals

The Bonferroni correction can be used to adjust confidence intervals. If one establishes confidence intervals, and wishes to have an overall confidence level of , each individual confidence interval can be adjusted to the level of .[2][3]

Alternatives

There are alternative ways to control the familywise error rate. For example, the Holm–Bonferroni method and the Šidák correction are universally more powerful procedures than the Bonferroni correction, meaning that they are always at least as powerful. Unlike the Bonferroni procedure, these methods do not control the expected number of Type I errors per family (the per-family Type I error rate).[8]

Criticism

With respect to FWER control, the Bonferroni correction can be conservative if there are a large number of tests and/or the test statistics are positively correlated.

The correction comes at the cost of increasing the probability of producing false negatives, i.e., reducing statistical power.[9]

There is not a definitive consensus on how to define a family in all cases, and adjusted test results may vary depending on the number of tests included in the family of hypotheses.

Note that these criticisms apply to FWER control in general, and are not specific to the Bonferroni correction.

References

  1. Bonferroni, C. E., Teoria statistica delle classi e calcolo delle probabilità, Pubblicazioni del R Istituto Superiore di Scienze Economiche e Commerciali di Firenze 1936
  2. 1 2 Dunn, Olive Jean (1958). "Estimation of the Means for Dependent Variables". Annals of Mathematical Statistics. 29 (4): 1095–1111. doi:10.1214/aoms/1177706374. JSTOR 2237135.
  3. 1 2 Dunn, Olive Jean (1961). "Multiple Comparisons Among Means" (PDF). Journal of the American Statistical Association. 56 (293): 52–64. doi:10.1080/01621459.1961.10482090.
  4. Mittelhammer, Ron C.; Judge, George G.; Miller, Douglas J. (2000). Econometric Foundations. Cambridge University Press. pp. 73–74. ISBN 0-521-62394-4.
  5. Miller, Rupert G. (1966). Simultaneous Statistical Inference. Springer.
  6. Goeman, Jelle J.; Solari, Aldo (2014). "Multiple Hypothesis Testing in Genomics". Statistics in Medicine. 33 (11). doi:10.1002/sim.6082.
  7. Neuwald, AF; Green, P (1994). "Detecting patterns in protein sequences". J. Mol. Biol. 239 (5): 698–712. doi:10.1006/jmbi.1994.1407. PMID 8014990.
  8. Frane, Andrew (2015). "Are per-family Type I error rates relevant in social and behavioral science?". Journal of Modern Applied Statistical Methods. 14 (1): 12–23.
  9. Nakagawa, Shinichi (2004). "A farewell to Bonferroni: the problems of low statistical power and publication bias". Behavioral Ecology. 15 (6): 1044–1045. doi:10.1093/beheco/arh107.

Further reading

  • Dunnett, C. W. (1955). "A multiple comparisons procedure for comparing several treatments with a control". Journal of the American Statistical Association. 50 (272): 1096–1121. doi:10.1080/01621459.1955.10501294.
  • Dunnett, C. W. (1964). "New tables for multiple comparisons with a control". Biometrics. 20 (3): 482–491. doi:10.2307/2528490. JSTOR 2528490.
  • Shaffer, J. P. (1995). "Multiple Hypothesis Testing". Annual Review of Psychology. 46: 561–584. doi:10.1146/annurev.ps.46.020195.003021.
  • Strassburger, K.; Bretz, Frank (2008). "Compatible simultaneous lower confidence bounds for the Holm procedure and other Bonferroni-based closed tests". Statistics in Medicine. 27 (24): 4914–4927. doi:10.1002/sim.3338.
  • Šidák, Z. (1967). "Rectangular confidence regions for the means of multivariate normal distributions". Journal of the American Statistical Association. 62 (318): 626–633. doi:10.1080/01621459.1967.10482935.
  • Hochberg, Yosef (1988). "A Sharper Bonferroni Procedure for Multiple Tests of Significance" (PDF). Biometrika. 75 (4): 800–802. doi:10.1093/biomet/75.4.800.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.