期刊论文详细信息
Frontiers in Applied Mathematics and Statistics
New metrics for multiple testing with correlated outcomes
Applied Mathematics and Statistics
Tyler J. VanderWeele1  Maya B. Mathur2 
[1] Department of Epidemiology, Harvard T. H. Chan School of Public Health, Boston, MA, United States;Quantitative Sciences Unit and Department of Pediatrics, Stanford University, Palo Alto, CA, United States;
关键词: multiplicity;    Type I error;    bootstrap;    resampling;    familywise error rate;    multiverse;   
DOI  :  10.3389/fams.2023.1151314
 received in 2023-01-25, accepted in 2023-04-10,  发布年份 2023
来源: Frontiers
PDF
【 摘 要 】

When investigators test multiple outcomes or fit different model specifications to the same dataset, as in multiverse analyses, the resulting test statistics may be correlated. We propose new multiple-testing metrics that compare the observed number of hypothesis test rejections (θ^) at an unpenalized α-level to the distribution of rejections that would be expected if all tested null hypotheses held (the “global null”). Specifically, we propose reporting a “null interval” for the number of α-level rejections expected to occur in 95% of samples under the global null, the difference between θ^ and the upper limit of the null interval (the “excess hits”), and a one-sided joint test based on θ^ of the global null. For estimation, we describe resampling algorithms that asymptotically recover the sampling distribution under the global null. These methods accommodate arbitrarily correlated test statistics and do not require high-dimensional analyses, though they also accommodate such analyses. In a simulation study, we assess properties of the proposed metrics under varying correlation structures as well as their power for outcome-wide inference relative to existing methods for controlling familywise error rate. We recommend reporting our proposed metrics along with appropriate measures of effect size for all tests. We provide an R package, NRejections. Ultimately, existing procedures for multiple hypothesis testing typically penalize inference in each test, which is useful to temper interpretation of individual findings; yet on their own, these procedures do not fully characterize global evidence strength across the multiple tests. Our new metrics help remedy this limitation.

【 授权许可】

Unknown   
Copyright © 2023 Mathur and VanderWeele.

【 预 览 】
附件列表
Files Size Format View
RO202310103415134ZK.pdf 3694KB PDF download
FEART_feart-2023-1169258_wc_tfx17.tif 99KB Image download
【 图 表 】

FEART_feart-2023-1169258_wc_tfx17.tif

  文献评价指标  
  下载次数:6次 浏览次数:1次