General Consideration
1. Trial Objective and Context
The foundation of any clinical trial lies in a clear understanding of its scientific objective. It is crucial to define whether the trial aims to predict outcomes or explain mechanisms. Specifically, the trial should clarify whether it intends to determine if an intervention is better, worse, or equivalent, and if so, for which specific patient subgroups this conclusion applies.
Equally important is identifying the primary audience for the trial’s conclusions. This could be:
In designing the trial, practical limitations must be considered, such as:
2. Data and Endpoint Considerations
Careful planning is required regarding the data to be collected, ensuring:
A clear mapping must exist between the collected data and:
The endpoint type, which could be:
Additionally, the chosen estimator, such as a hazard ratio for survival data, must align appropriately with both the data type and the trial objectives.
3. Hypothesis Construction
The trial should explicitly define its primary hypothesis, which can take several forms:
If the trial involves multiple tests or comparisons, it is essential to:
Define how overall trial success will be determined, using:
Implement a strategy to control the Type I error rate, such as:
4. Trial Design Considerations
Selecting an appropriate trial design is central to obtaining reliable and interpretable results. Possible designs include:
The design impacts both:
If randomization is required, considerations include:
5. Estimator, Effect Size, and Estimand Considerations
A well-defined estimand strategy ensures clarity in interpreting treatment effects, considering:
Whether the estimand reflects:
The estimator selected must:
Approaches for handling:
6. Statistical Model and Test Considerations
The statistical model forms the backbone of data analysis, and its selection should:
Align with the endpoint, trial design, and estimand strategy.
Be chosen for its efficiency, ensuring maximum power with minimal sample size.
Be compared against alternatives through:
Additional considerations include:
7. Advanced Trial Aspects
For complex trials, advanced methodologies may be considered, such as:
Adaptive Designs, where pre-planned adaptations (e.g., sample size adjustments, early stopping) are permitted based on interim data. Key factors include:
Bayesian Approaches, which incorporate prior knowledge through:
Engagement with regulatory authorities and stakeholders is critical, ensuring alignment with their:
8. Sample Size Considerations
An adequate sample size is essential to ensure:
If practical constraints limit sample size:
1. “Simplifying” Data and/or Model
Dichotomania and Loss of Efficiency Dichotomizing continuous variables (turning them into “yes/no”, “responder/non-responder”, etc.) severely reduces statistical efficiency:
Pitman Efficiency Curve
Responder Analysis Pitfalls Responder analyses are especially problematic because:
They inherently involve dichotomization based on arbitrary thresholds.
Thresholds often lack a clear clinical or biological justification.
When based on change from baseline variables, responder analyses can produce misleading scenarios, where:
As highlighted by Senn and others in the literature, responder analyses frequently:
Best Practice Principle
Always use the data as close to its original form as possible:
Avoid “simplifying” data just to make results easier to explain to non-technical stakeholders — this shortcut sacrifices scientific integrity.
2. Not Justifying or Understanding Design
Trial designs must be evaluated for:
Correct interpretation of:
Justification of equivalence or non-inferiority margins is critical and must be:
3. Post-hoc Design Changes or Rationalization
Avoid making unplanned design or analysis changes after seeing the data:
Particularly, avoid:
Instead:
4. Covariates: Use Them Properly
If you collect baseline covariates (e.g., age, gender, smoking status, biomarkers), use them in the analysis, especially:
Proper covariate adjustment:
This is supported by:
Caveat: Only pre-randomization covariates should be included:
5. Not Involving Other Stakeholders (Early and Often)
Early involvement and feedback from:
This collaboration increases the likelihood of trial success and reduces avoidable mistakes.
Quote from Ronald Fisher
“To call in the statistician after the experiment is done may be no more than asking him to perform a postmortem examination: he may be able to say what the experiment died of.”
Ronald Fisher, a pioneer in modern statistics, emphasizes that:
Overall Message
Clinical trial success requires:
✔ Rigorous design based on sound statistical principles. ✔ Avoidance of unnecessary simplifications that reduce power and efficiency. ✔ Pre-specification to minimize bias and misleading findings. ✔ Early, active collaboration with all stakeholders. ✔ Proper interpretation of statistical results aligned with scientific objectives.
Failing to respect these principles increases the risk of invalid, inefficient, or irreproducible trial outcomes.
Background on Power
Explanation: This refers to converting a continuous outcome into a binary one (e.g., “responder” vs. “non-responder”).
Impact on Power:
Example: Turning a pain score from 0-10 into “improved” vs. “not improved” reduces sensitivity.
Bottom Line: Responder analysis is (nearly) always a poor choice if a continuous measure is available.
Explanation: Adding relevant covariates (baseline characteristics) into the statistical model.
Impact on Power:
Important Caveats:
Example: Adjusting for baseline age or disease severity can boost efficiency.
Explanation: Trial includes planned interim analyses with the possibility of stopping early for efficacy or futility.
Impact on Power and Sample Size:
Example: Trial designed to stop at interim if strong positive effect is seen—reduces average trial size over many repetitions.
Explanation: The structure or level at which randomization is applied.
Impact on Power:
Example:
Explanation: The specific statistical tests and estimators used in the analysis.
Impact on Power:
Example:
Recommendation: Compare performance of candidate tests under plausible scenarios during planning.
Summary Takeaways
The practice of milestone prediction in clinical trials is multifaceted, blending statistical rigor with strategic foresight. It’s about more than just adhering to a schedule; it’s about adapting to realities on the ground and ensuring that a trial can meet its objectives without wasting resources. Effective milestone management helps maintain the integrity of the trial process, ensuring that therapeutic potentials are accurately assessed while upholding the highest standards of safety and efficacy.
In clinical trial management, understanding both enrollment dynamics and event occurrence—including dropouts, cures, or any factors preventing subjects from experiencing key events—is crucial. Given the commonality of delays, with approximately 80% of trials experiencing slowdowns and about 85% failing to reach recruitment goals, the need for robust milestone prediction is evident. This prediction involves assessing practical elements such as enrollment strategies and resource allocation, which are essential to maintaining trial timelines and efficiency.
Key Challenges and Strategies in Milestone Prediction
Enrollment and Event Tracking: The primary milestones in most trials involve patient enrollment and tracking event occurrences, like patient survival or endpoint achievement. In event-driven studies, such as those focusing on survival, predicting when the study might conclude or when interim analyses might be needed is paramount.
Handling Practical Challenges: Addressing practical issues involves predicting enrollment timelines and identifying potential delays early. If enrollment lags, strategies might include opening new trial sites or closing underperforming ones. Proactive resource management, such as reallocating resources to more critical areas, becomes possible with accurate milestone forecasting.
Data Availability and Prediction Management:
Frequency and Timing of Predictions:
Special Considerations for Survival Studies:
External Predictions: Employing external experts for milestone predictions can reduce bias and provide access to a broader range of methodologies. External predictors, less influenced by internal trial dynamics, might offer a clearer, unbiased perspective.
Additional Considerations
Enrollment prediction is a crucial aspect of clinical trial planning and management, serving as a foundational metric for assessing a trial’s timeline and resource allocation. It encompasses predicting both the rate and completeness of participant recruitment over the course of the study. This process not only impacts the financial and logistical aspects of a trial but also its scientific validity, as timely enrollment ensures that the trial can achieve its intended statistical power and objectives.
Implementing effective enrollment predictions requires a multi-faceted approach: - Data integration: Combining data from multiple sources, including historical trial data, current site performance, and external factors. - Continuous monitoring: Regularly updating predictions based on new data to stay responsive to changing conditions. - Stakeholder communication: Using prediction data to maintain open dialogue with sponsors and adjust expectations and strategies as needed.
Initial and Mid-Trial Predictions
Enrollment predictions typically begin with estimating how long it will take to recruit the full sample size needed to meet the study’s power requirements. This involves assessing: - Demographics: The availability and willingness of the target population to participate. - Competing studies: Other ongoing trials that could affect participant availability. - Site capabilities: Each site’s ability to recruit and manage participants.
Mid-trial predictions evaluate whether enrollment is on track to meet planned timelines. Adjustments might be needed if the trial is progressing faster or slower than expected.
Challenges with Early and Later Phase Trials
Site-Specific Predictions
More sophisticated approaches to enrollment prediction involve modeling each recruitment site or region separately. This allows for: - Detailed tracking: Identifying which sites are underperforming. - Resource reallocation: Shifting resources to more effective sites or boosting those that are lagging. - Adaptive strategies: Adjusting recruitment tactics based on real-time data.
Simple Statistical Models These include linear or polynomial models that provide a basic forecast based on past recruitment rates.
Piecewise Parametric Models These models identify changes in recruitment pace, such as an initial slow start followed by a faster rate, allowing for more nuanced predictions.
Simulation-Based Modeling Simulation offers a flexible and dynamic approach to modeling recruitment. It allows for:
Bayesian Models These incorporate prior data and expert opinions to refine predictions, adapting as new data becomes available during the trial.
Machine Learning Approaches While not covered in detail here, machine learning methods can analyze complex datasets to predict recruitment outcomes, potentially uncovering hidden patterns that affect enrollment.
In survival trials, the occurrence of key events such as death or disease progression is fundamental to determining the trial’s timeline and outcomes. The predictive modeling of these events is complex due to the multifaceted nature of survival data, which can include various competing risks and time-dependent factors.
Event-Driven Endpoints: In many clinical trials, especially those concerning life-threatening conditions, the trial’s endpoint is driven by the accumulation of specific events among participants (e.g., death, disease progression). The number of events directly impacts the trial’s power and its ability to provide statistically meaningful results. Without a sufficient number of events, the trial cannot conclude or make robust inferences.
Challenges in Event Prediction
Modeling Techniques for Event Prediction
Parametric Models: These models, such as the exponential or Weibull models, assume a specific distribution for the time until an event occurs. They are straightforward but often too simplistic for complex survival data.
Piecewise Parametric Models: These improve on simple parametric models by allowing different parameters in different phases of the study, accommodating varying hazard rates across the trial’s duration.
Simulation-Based Models: Simulations provide a flexible and dynamic approach to understanding how different factors might impact event rates. This method is particularly useful in survival trials where complex interactions between patient characteristics and treatment effects need to be considered.
Practical Implementation of Event Prediction
Exponential Models: Assume constant hazard rates throughout the trial period. This is simplistic but can serve as a baseline for understanding baseline event rates.
Piecewise Exponential Models: Offer more flexibility by dividing the trial into segments, each with its own hazard rate, better modeling the natural progression of disease or treatment effects over time.
Two-Parameter Models: These models account for the duration a participant has been in the study, adjusting the event probability based on this tenure. They are useful in long-term studies where the risk of an event may increase or decrease over time.
Model Selection and Evaluation: Employing information criteria like AIC (Akaike Information Criterion) helps in selecting the best-fitting model amongst various candidates. Advanced techniques might also dynamically allocate change points to adapt the model to observed data patterns more accurately.
Senn, S. S. (2021). Statistical issues in drug development (3rd ed.). John Wiley & Sons.
Harrell, F. (2025). Biostatistics for Biomedical Research. Vanderbilt Department of Biostatistics. Available at: https://hbiostat.org/bbr/
Lakens, D. (2016). Improving your statistical inferences. Available at: https://lakens.github.io/statistical_inferences
Meinert, C. L. (2012). Clinical trials: Design, conduct and analysis (Vol. 39). OUP USA.
Friedman, L. M., Furberg, C. D., DeMets, D. L., Reboussin, D. M., & Granger, C. B. (2015). Fundamentals of clinical trials. Springer.
American Medical Association. (2023). Guidelines for conducting clinical trials. Journal of Clinical Research, 29(3), 211–230.
International Council for Harmonisation of Technical Requirements for Pharmaceuticals for Human Use (ICH). (1998). Statistical principles for clinical trials (E9). Available at: https://database.ich.org/sites/default/files/E9_Guideline.pdf
International Council for Harmonisation of Technical Requirements for Pharmaceuticals for Human Use (ICH). (2019). Addendum on estimands and sensitivity analysis in clinical trials to the guideline on statistical principles for clinical trials (E9[R1]). Available at: https://database.ich.org/sites/default/files/E9-R1_Step4_Guideline_2019_1203.pdf
Kahan, B. C., Hindley, J., Edwards, M., Cro, S., & Morris, T. P. (2024). The estimands framework: A primer on the ICH E9 (R1) addendum. BMJ, 384.
Senn, S. S. (2002). Cross-over trials in clinical research (2nd ed.). John Wiley & Sons.
International Council for Harmonisation of Technical Requirements for Pharmaceuticals for Human Use (ICH). (2021). General considerations for clinical studies (E8[R1]). Available at: https://database.ich.org/sites/default/files/ICH_E8-R1_Guideline_Step4_2021_1006.pdf
International Council for Harmonisation of Technical Requirements for Pharmaceuticals for Human Use (ICH). ICH Guidelines. Available at: https://www.ich.org/page/ich-guidelines
U.S. Food and Drug Administration (FDA). Clinical Trials Guidance Documents. Available at: https://www.fda.gov/science-research/clinical-trials-and-human-subject-protection/clinical-trials-guidance-documents
European Medicines Agency (EMA). Clinical Trials Regulation. Available at: https://www.ema.europa.eu/en/human-regulatory-overview/research-and-development/clinical-trials-human-medicines/clinical-trials-regulation
SPIRIT-CONSORT Group. SPIRIT-CONSORT Guidelines. Available at: https://www.consort-spirit.org/
van Smeden, M. (2022). A very short list of common pitfalls in research design, data analysis, and reporting. PRiMER: Peer-Reviewed Reports in Medical Education Research, 6, 26.
Sydes, M. R., & Langley, R. E. (2010). Potential pitfalls in the design and reporting of clinical trials. The Lancet Oncology, 11(7), 694–700.
Zlowodzki, M., Jönsson, A., & Bhandari, M. (2005). Common pitfalls in the conduct of clinical research. Medical Principles and Practice, 15(1), 1–8.
Dahly, D. (2024). Sorry, what was the question again? Available at: https://statsepi.substack.com/p/sorry-what-was-the-question-again
Altman, D. BMJ Statistics Notes. Available at: https://www.bmj.com/specialties/statistics-notes
Covariates/Randomization Issues
Senn, S. (2006). Change from baseline and analysis of covariance revisited. Statistics in Medicine, 25(24), 4334–4344.
Senn, S. (2019). The well-adjusted statistician: Analysis of covariance explained. Applied Clinical Trials.
Van Breukelen, G. J. (2006). ANCOVA versus change from baseline had more power in randomized studies and more bias in nonrandomized studies. Journal of Clinical Epidemiology, 59(9), 920–925.
Food and Drug Administration (FDA). (2023). Adjusting for covariates in randomized clinical trials for drugs and biological products: Guidance for industry. Available at: https://www.fda.gov/media/148910/download
Bland, J. M., & Altman, D. G. (2011). Comparisons against baseline within randomised groups are often used and can be highly misleading. Trials, 12, Article 264.
Senn, S. (2013). Seven myths of randomisation in clinical trials. Statistics in Medicine, 32(9), 1439–1450.
Recruitment, Adherence, and Missing Data
Brown, J. T., & Smith, R. A. (2022). Patient recruitment strategies in clinical trials: Challenges and solutions. Journal of Clinical Trials, 15(1), 45–67.
Sterne, J. A., White, I. R., Carlin, J. B., Spratt, M., Royston, P., Kenward, M. G., Wood, A. M., & Carpenter, J. R. (2009). Multiple imputation for missing data in epidemiological and clinical research: Potential and pitfalls. BMJ, 338, Article b2393.
Hughes, R. A., Heron, J., Sterne, J. A., & Tilling, K. (2019). Accounting for missing data in statistical analyses: Multiple imputation is not always the answer. International Journal of Epidemiology, 48(4), 1294–1304.
Sacristán, J. A., Aguarón, A., Avendaño-Solá, C., Garrido, P., Carrión, J., Gutiérrez, A., Kroes, R., & Flores, A. (2016). Patient involvement in clinical research: Why, when, and how. Patient Preference and Adherence, 10, 631–640.
P-values: Misinterpretation and P-hacking
Wasserstein, R. L., & Lazar, N. A. (2016). The ASA statement on p-values: Context, process, and purpose. The American Statistician, 70(2), 129–133.
Altman, D. G., & Bland, J. M. (1995). Statistics notes: Absence of evidence is not evidence of absence. BMJ, 311(7003), 485.
Smith, G. D., & Ebrahim, S. (2002). Data dredging, bias, or confounding: They can all get you into the BMJ and the Friday papers. BMJ, 325(7378), 1437–1438.
Stefan, A. M., & Schönbrodt, F. D. (2023). Big little lies: A compendium and simulation of p-hacking strategies. Royal Society Open Science, 10(2), 220346.
Stefan, A. M., & Schönbrodt, F. D. (2023). Gazing into the abyss of p-hacking: A Shiny app for p-hacking simulation. Available at: https://shiny.psy.lmu.de/felix/ShinyPHack/
Food and Drug Administration (FDA), 2019. Adaptive Designs for Clinical Trials of Drugs and Biologics Guidance for Industry. Download at: https://www.fda.gov/media/78495/download A Practical Adaptive & Novel Designs and Analysis toolkit (PANDA). Find at: https://panda.shef.ac.uk/ Jennison, C., & Turnbull, B. W. (1999). Group sequential methods with applications to clinical trials. CRC Press. Wassmer, G., & Brannath, W. (2016). Group sequential and confirmatory adaptive designs in clinical trials. Cham: Springer International Publishing. Bauer, P., Bretz, F., Dragalin, V., König, F. and Wassmer, G., 2016. Twenty‐five years of confirmatory adaptive designs: opportunities and pitfalls. Statistics in Medicine, 35(3), pp.325-347. Pallmann, P., Bedding, A.W., Choodari-Oskooei, B., Dimairo, M., Flight, L., Hampson, L.V., Holmes, J., Mander, A.P., Odondi, L.O., Sydes, M.R. and Villar, S.S., 2018. Adaptive designs in clinical trials: why use them, and how to run and report them. BMC medicine, 16(1), pp.1-15. Arnold, B.F., Hogan, D.R., Colford, J.M. and Hubbard, A.E., 2011. Simulation methods to estimate design power: an overview for applied research. BMC medical research methodology, 11, pp.1-10. Kimko, H.C. and Duffull, S.B., 2002. Simulation for designing clinical trials. Marcel Dekker Incorporated. Landau, S. and Stahl, D., 2013. Sample size and power calculations for medical studies by simulation when closed form expressions are not available. Statistical methods in medical research, 22(3), pp.324-345. Eng, J., 2004. Sample size estimation: a glimpse beyond simple formulas. Radiology, 230(3), pp.606-612. Meyer, K.M., Mooij, W.M., Vos, M., Hol, W.G. and van der Putten, W.H., 2009. The power of simulating experiments. Ecological Modelling, 220(19), pp.2594-2597. PowerSimulate: https://shiny.jdl-svr.lat/PowerSimulate/ Dallow, N, Best, N, Montague, TH., 2018. Better decision making in drug development through adoption of formal prior elicitation. Pharmaceutical Statistics. 17: 301– 316. Oakley, J. E. and O’Hagan, A. SHELF: The Sheffield Elicitation Framework (version 2.0), School of Mathematics and Statistics, University of Sheffield, 2010 http://tonyohagan.co.uk/shelf.
Background
Enrollment
Events