RESEARCH

Recent Projects

The consequences of optional stopping on the research literature.

Open Access Link

ABSTRACT: Optional stopping is the practice of repeatedly analyzing data during the data collection process with the intention of terminating collection once statistical significance is observed. While there are ways to appropriately incorporate interim data analyses, the misuse and lack of transparent reporting of this practice has garnered criticism as a questionable research practice, resulting in inflated Type I error rates. The present simulation study examines optional stopping from the perspective of the overall research literature, where some researchers engage in this practice. Across varying contexts and severities of optional stopping, we examined consequences relevant to replicable research literatures, including effect size bias (partitioned into publication bias and bias unique to optional stopping), heterogeneity across studies, and error rates (including Type S errors). Results demonstrated that optional stopping within a research literature can lead to bias, inaccurate estimates of heterogeneity across studies, and increased Type I and Type S error rates. However, more importantly, patterns emerged that help to demonstrate when and how optional stopping exerts specific consequences on the research literature. We hope these results further clarify the way optional stopping stands in the way of a more cumulative, replicable research literature, and we provide recommendations for researchers in light of this.

Inference in randomized pretest-posttest studies under missing data: Influence of MAR sub-patterns on statistical power and precision.

ABSTRACT: Missing data are common in multiwave studies, including the (two wave) randomized pretest posttest (RPP) design. The commonly assumed missing-at-random (MAR) assumption specifies that the missing data is related only to observed variables. However, limited prior research suggests potential utility in examining missing data patterns within this broad category. This may be particularly important for statistical power and precision: understanding how different forms of MAR affect these quantities could build a foundation for more customized sample size planning. Thus, I investigated how different sub-patterns of the MAR mechanism impact power, precision, bias, and coverage rates when estimating the average treatment effect (ATE) in an RPP design. The present simulation examined how variations in missing data sub-patterns – which link pretest score and/or group membership to dropout at posttest in different ways – affect ATE estimates under listwise deletion and multiple imputation. Results demonstrated that different MAR sub-patterns led to differential impacts on power and precision for both the ATE and within-group regression estimates, although the within-group regression was impacted substantially more. Notably, the patterns associated with the greatest power loss for the ATE did not align with those affecting within-group regression, underscoring the need for more tailored approaches to sample size planning under missing data. These findings support the need for a more nuanced understanding of the MAR assumption, highlighting the importance of developing methods to consider MAR sub-patterns in sample size planning and study design.