Phase II Trials - Design Considerations
Phase II clinical trials primarily aim to evaluate the treatment effects in a relatively small number of patients to decide whether the treatment should be studied in a larger scale comparative trial. They play a crucial role in drug development, as the outcomes determine whether to proceed with Phase III trials. The multistage designs for Phase II clinical trials proposed by Gehan, Fleming, Simon, and Ensign are described and compared. Gehan and Simon’s designs include two stages, Fleming’s design may have two or more stages, and Ensign’s three-stage design combines Gehan’s first stage with Simon’s two stages.
Kramar, A., Potvin, D., & Hill, C. (1996). Multistage designs for phase II clinical trials: statistical issues in cancer research. British Journal of Cancer, 74(8), 1317–1320. doi:10.1038/bjc.1996.537
The response to treatment will be summarized as either success or failure. In oncology, success is often defined as a complete response or objective response, which includes partial responses.
The treatment effect is evaluated by the parameter \(\pi\), which is the true proportion of success in the given population. If this true proportion of success is less than or equal to a predefined value \(p_0\) (known as the maximum ineffectiveness rate), the treatment is considered insufficiently effective. If the true proportion of success is greater than or equal to a predefined value \(p_1\) (known as the minimum effectiveness rate), the treatment is considered sufficiently effective for further study in Phase III trials.
Most statistical methods for Phase II trials are developed to include patients in at least two stages. In the simplest case, \(r_1\) patients are included in the first stage, and \(r_1\) successes are observed. Based on the value of \(r_1\), the trial either stops or continues to enroll \(n_2\) patients into the second stage, where \(r_2\) successes are observed. Across \(n_1 + n_2\) patients, the cumulative number of successes is \(R_2 = r_1 + r_2\). Depending on the observed cumulative number of successes \(R_k\), the trial continues to accumulate patients. The general procedure for each stage \(k\) is as follows:
At the end of the last stage, there is a cutoff point. If the total number of successes is at least equal to this boundary, a conclusion in favor of efficacy is drawn; if the total number of successes is below this boundary, the treatment is concluded to be ineffective.
The primary purpose of Phase II clinical trials is to evaluate the treatment effects in a relatively small group of patients to determine whether the treatment should be studied further in large-scale comparative trials. They play a critical role in the drug development process, as the results decide whether to proceed to Phase III trials. The designs for Phase II clinical trials proposed by Gehan, Fleming, Simon, and Ensign are described and compared. Gehan and Simon’s designs involve two stages, Fleming’s design may include two or more stages, and Ensign’s three-stage design combines elements from Gehan’s first stage with Simon’s two stages.
Simon (1989) proposed two designs, each with two stages. Simon’s Optimum design minimizes the average number of patients receiving an ineffective treatment, while his Minimax design minimizes the maximum sample size required. In these designs, unlike Fleming’s design, the number of patients in each stage is not specified by the investigator, but rather is the result of a minimization constraint.
In Simon’s Ensign design, on the other hand, the sample size is not defined by the investigator but is determined by a minimization criterion. In Simon’s Optimum design, the probability of early termination of the trial is high under the assumption that the treatment is ineffective, and the average number of patients is minimized. For extremely rare diseases, and therefore low accrual rates, Simoa’s Minimax design will be superior to the Optimum design because it limits the maximum duration of the study.
Reference: Simon, R. (1989). Optimal two-stage designs for phase II clinical trials. Controlled Clinical Trials, 10(1), 1-10. doi:10.1016/0197-2456(89)90015-9
Phase II studies in cancer treatment are non-controlled trials used to obtain a preliminary estimate of the anti-tumor effect of a treatment. - Phase I trials provide information about the maximum tolerated dose, which is crucial because most cancer treatments must be administered at high doses to achieve the best effect. However, Phase I trials usually treat only 3 to 6 patients per dose level, and the diagnoses of cancer vary among patients. Therefore, these trials offer little or no information about anti-tumor activity. The proportion of patients experiencing tumor reduction of at least 50% is the primary endpoint for most Phase II trials, although the duration of such responses is also of interest. These uncontrolled trials do not establish the “efficacy” or the role of the drug in treating the disease. The purpose of Phase II trials for new anticancer drugs is to determine whether the drug has sufficient activity against a particular type of tumor to justify further development. Further development might mean combining the drug with other drugs, assessing it in patients with milder conditions, or initiating Phase III studies to compare survival outcomes with those of standard treatments.
The combined protocol of Phase II trials aims to determine whether a treatment has enough promise to warrant a major comparative clinical assessment against standard treatments. The designs developed here are based on testing the null hypothesis \(H_0: P \le P_0\), i.e., the true response rate is less than some uninteresting level \(P_0\). If the null hypothesis is true, we require the probability of concluding that the drug has enough promise to warrant further clinical trials to be less than \(\alpha\). We also require that, if a specified alternative hypothesis \(P \ge P_1\)—where the true response rate is at least some desirable target level \(P_1\)—is true, the probability of rejecting the drug for further study should be less than \(\beta\). In addition to these constraints, we aim to minimize the number of patients receiving a low-activity treatment. For practical considerations of managing multicenter clinical trials, we focus on two-stage designs.
Simon’s Design is a two-stage design with an early stopping rule for futility, testing a null hypothesis of a “poor” response against an alternative of a “good” response. If the numbers of patients studied in the first and second stages are denoted by \(n_1\) and \(n_2\) respectively, then the expected sample size is \(EN = n_1 + (1 - \text{PET}) n_2\), where PET represents the probability of early termination after the first stage. The decision of whether or not to terminate after the first stage will be based on the number of responses observed among those \(n_1\) patients. The expected sample size \(EN\) and the probability of early termination depend on the true probability of response \(p\). The experiment will be terminated at the end of the first stage and the drug rejected if \(r_1\) or fewer responses are observed. This occurs with probability \(PET = B(r_1; p, n_1)\), where \(B\) denotes the cumulative binomial distribution. The drug will be rejected at the end of the second stage if \(r\) or fewer responses are observed. Thus, the probability of rejecting a drug with a success probability \(p\) is: \[ B(r_1; p, n_1) + \sum_{x = r_1 + 1}^{\min[n_1, r]} b(x; p, n_1) B(r - x; p, n_2), \] where \(b\) denotes the binomial probability mass function.
For specified values of \(p_0\), \(p_1\), \(\alpha\), and \(\beta\), optimal designs have been determined by enumeration using exact binomial probabilities. For each value of total sample size \(n\) and each value of \(n_1\) in the range \(1, n - 1\), we determined the integer values of \(r_1\) and \(r\), which satisfied the two constraints and minimized the expected sample size when \(p = p_0\). This was found by searching over the range \(r_1 \in (0, n_1)\). For each value of \(r_1\), we determined the maximum value of \(r\) that satisfied the type 2 error constraint. We then examined whether that set of parameters \((n, n_1, r_1, r)\) satisfied the type 1 error constraint. If it did, then we compared the expected sample size to the minimum achieved by previous feasible designs and continued the search over \(r_1\). Keeping \(n\) fixed, we searched over the range of \(n_1\) to find the optimal two-stage design for that maximum sample size \(n\). The search over \(n\) ranged from a lower value of about: \[ \bar{p}(1 - \bar{p}) \left[\frac{z_{1 - \alpha} + z_{1 - \beta}}{p_1 - p_0}\right]^2, \] where \(\bar{p} = \left(p_0 + p_1\right) / 2\). We checked below this starting point to ensure that we had determined the smallest maximum sample size \(n\) for which there was a nontrivial \((n_1, n_2 > 0)\) two-stage design that satisfied the error probability constraints. The enumeration procedure searched upwards from this minimum value of \(n\) until it was clear that the optimum had been determined. The minimum expected sample size for fixed \(n\) is not a unimodal function of \(n\) because of the discreteness of the underlying binomial distributions. Nevertheless, eventually, as \(n\) increased, the value of the local minima increased, and it was clear that a global minimum had been found.
Parameter | ph2simon(pu, pa, ep1, ep2, nmax=100) |
---|---|
pu | unacceptable response rate |
pa | response rate that is desirable |
ep1 | threshold for the probability of declaring drug desirable under p0 |
ep2 | threshold for the probability of rejecting the drug under p1 |
knitr::include_graphics("./02_Plots/Simon_twoStage.PNG")
Figure: Simon’s 2-Stage Design
## `ph2simon(pu, pa, ep1, ep2, nmax=100)`
# library("clinfun")
Simon <- ph2simon(0.2, 0.35, 0.05, 0.05, nmax=150)
Simon
##
## Simon 2-stage Phase II design
##
## Unacceptable response rate: 0.2
## Desirable response rate: 0.35
## Error rates: alpha = 0.05 ; beta = 0.05
##
## r1 n1 r n EN(p0) PET(p0) qLo qHi
## Minimax 15 68 25 95 75.44 0.7244 0.582 1.000
## Admissible 11 53 26 99 69.88 0.6330 0.306 0.582
## Admissible 11 51 27 104 67.68 0.6852 0.109 0.306
## Optimal 11 50 28 109 67.07 0.7107 0.000 0.109
plot(Simon)
Figure: Simon’s 2-Stage Design
In safety and efficacy studies, it is meaningful to determine if an increase in dose leads to an increase (or decrease) in response. The statistical analysis for such a situation is referred to as dose-response or trend analysis. Here, the goal is to observe a trend, not merely differences between groups. Typically, patients in a dose-response study are randomized into K + 1 treatment groups (a placebo group and K increasing doses of the drug). The response variables may be binary, ordinal, or continuous. In some instances, trend tests are sensitive and can reveal a mild trend where pairwise comparisons would not find significant differences.
For illustration, assume the response is continuous and we aim to determine if there is a trend across the K + 1 population means. A one-sided hypothesis testing framework for investigating an increasing trend could be: \[ H_0: \mu_0 = \mu_1 = \cdots = \mu_K \] versus \[ H_1: \mu_0 \leq \mu_1 \leq \cdots \leq \mu_K \] with at least one strict inequality.
A one-sided hypothesis testing framework for investigating a decreasing trend could be: \[ H_0: \mu_0 = \mu_1 = \cdots = \mu_K \] versus \[ H_1: \mu_0 \geq \mu_1 \geq \cdots \geq \mu_K \] with at least one
strict inequality.
A two-sided hypothesis testing framework for investigating a trend could be: \[ H_0: \mu_0 = \mu_1 = \cdots = \mu_K \] versus \[ H_1: \mu_0 \leq \mu_1 \leq \cdots \leq \mu_K \] or \[ \mu_0 \geq \mu_1 \geq \cdots \geq \mu_K \] with at least one strict inequality.
Jonckheere-Terpstra (JT) trend test
For continuous responses, an appropriate test is the Jonckheere-Terpstra (JT) trend test, which is based on a sum of Mann-Whitney-Wilcoxon tests: \[ JT = \sum_{k=0}^{K-1}\sum_{k'=1}^{K} MWW_{kk'} \] where \(MWW_{kk'}\) is the Mann-Whitney-Wilcoxon test comparing group \(k\) to group \(k'\), with \(0 \leq k < k' \leq K\). This test compares each pair of groups and sums up the results to look for trends.
Example of constructing the JT statistic: Suppose there are four dose groups in a study (placebo, low dose, mid-dose, high dose). Then the JT trend test is the sum of six Mann-Whitney-Wilcoxon test statistics.
Next, assume that the \(K+1\) groups have a homogeneous population variance, \(\sigma^2\). The population variance is estimated by the pooled sample variance, \(s^2\), with \(d\) degrees of freedom: \[ s^2 = \frac{1}{d} \sum_{k=0}^{K} \sum_{i=1}^{n_k} (Y_{ki} - \bar{Y}_k)^2, \quad d = \sum_{k=0}^{K}(n_k - 1) \] Letting \(c_k = 2k - K\), the numerator reduces to: \[ \sum_{k=0}^{K} c_k \bar{Y}_k \] The trend statistic is then: \[ T = \left( \sum_{k=0}^{K} c_k \bar{Y}_k \right) / \left( \sqrt{s^2 \sum_{k=0}^{K} \frac{c_k^2}{n_k^2}} \right) \]
The JT trend test works well for binary and ordinal data, as well as for continuous data. Another trend test for binary data is the Cochran-Armitage (CA) trend test. The difference between the JT and CA trend tests is that for the CA test, the actual dose levels can be specified. In other words, instead of designating dose levels as low, mid, or high, the actual numerical dose levels can be used in the CA trend test, such as 20 mg, 60 mg, 180 mg.
*** Constructing JT trend tests;
proc format;
value groupfmt 0='Placebo' 1='20 mg' 2='60 mg' 3='180 mg';
value reactfmt 0='F' 1='S';
run;
data contin;
input group subject response;
cards;
0 01 27
0 02 28
0 03 27
0 04 31
0 05 34
0 06 32
1 01 31
1 02 35
1 03 34
1 04 32
1 05 31
1 06 33
2 01 32
2 02 33
2 03 30
2 04 34
2 05 37
2 06 36
3 01 40
3 02 39
3 03 41
3 04 38
3 05 42
3 06 43
;
run;
proc glm
data=contin;
class group;
model response=group;
contrast 'Trend Test' group -1.5 -0.5 0.5 1.5;
title "Parametric Trend Test for Continuous Data";
run;
proc freq
data=contin;
tables group*response/jt;
title "Jonckheere-Terpstra Trend Test for Continuous Data";
run;
data binary;
set contin;
if group=0 then dose=0;
if group=1 then dose=20;
if group=2 then dose=60;
if group=3 then dose=180;
if response<32 then react=0;
if response>=32 then react=1;
format react reactfmt.;
run;
proc freq
data=binary;
tables react*group/jt trend;
exact jt trend;
title "Jonckheere-Terpstra and Cochran-Armitage Trend Tests for Binary Data";
title2 "Ordinal Scores";
run;
proc freq
data=binary;
tables react*dose/jt trend;
exact jt trend;
title "Jonckheere-Terpstra and Cochran-Armitage Trend Tests for Binary Data";
title2 "Dose Scores";
run;
# library("clinfun")
set.seed(1234)
g <- rep(1:5, rep(10,5))
x <- rnorm(50)
jonckheere.test(x+0.3*g, g)
##
## Jonckheere-Terpstra test
##
## data:
## JT = 629, p-value = 0.02734
## alternative hypothesis: two.sided
x[1:2] <- mean(x[1:2]) # tied data
jonckheere.test(x+0.3*g, g)
##
## Jonckheere-Terpstra test
##
## data:
## JT = 639, p-value = 0.01741
## alternative hypothesis: two.sided
Perform a Cochran Armitage test for trend in binomial proportions across the levels of a single variable. This test is appropriate only when one variable has two levels and the other variable is ordinal. The two-level variable represents the response, and the other represents an explanatory variable with ordered levels. The null hypothesis is the hypothesis of no trend, which means that the binomial proportion is the same for all levels of the explanatory variable.
# library("DescTools")
dose <- matrix(c(10,9,10,7, 0,1,0,3), byrow=TRUE, nrow=2, dimnames=list(resp=0:1, dose=0:3))
Desc(dose)
## ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────
## dose (matrix, array)
##
## Summary:
## n: 40, rows: 2, columns: 4
##
## Pearson's Chi-squared test:
## X-squared = 6.6667, df = 3, p-value = 0.08332
## Log likelihood ratio (G-test) test of independence:
## G = 7.2877, X-squared df = 3, p-value = 0.06327
## Mantel-Haenszel Chi-squared:
## X-squared = 3.4667, df = 1, p-value = 0.06262
##
## Warning message:
## Exp. counts < 5: Chi-squared approx. may be incorrect!!
##
##
## Contingency Coeff. 0.378
## Cramer's V 0.408
## Kendall Tau-b 0.272
##
##
## dose 0 1 2 3 Sum
## resp
##
## 0 freq 10 9 10 7 36
## perc 25.0% 22.5% 25.0% 17.5% 90.0%
## p.row 27.8% 25.0% 27.8% 19.4% .
## p.col 100.0% 90.0% 100.0% 70.0% .
##
## 1 freq 0 1 0 3 4
## perc 0.0% 2.5% 0.0% 7.5% 10.0%
## p.row 0.0% 25.0% 0.0% 75.0% .
## p.col 0.0% 10.0% 0.0% 30.0% .
##
## Sum freq 10 10 10 10 40
## perc 25.0% 25.0% 25.0% 25.0% 100.0%
## p.row . . . . .
## p.col . . . . .
##
CochranArmitageTest(dose)
##
## Cochran-Armitage test for trend
##
## data: dose
## Z = -1.8856, dim = 4, p-value = 0.05935
## alternative hypothesis: two.sided
### Test independence using permutation test
# library("coin")
lungtumor <- data.frame(dose = rep(c(0, 1, 2), c(40, 50, 48)),
tumor = c(rep(c(0, 1), c(38, 2)),
rep(c(0, 1), c(43, 7)),
rep(c(0, 1), c(33, 15))))
independence_test(tumor ~ dose, data = lungtumor, teststat = "quad")
##
## Asymptotic General Independence Test
##
## data: tumor by dose
## chi-squared = 10.638, df = 1, p-value = 0.001108
## Test propotion
tab <- table(lungtumor$dose, lungtumor$tumor)
CochranArmitageTest(tab)
##
## Cochran-Armitage test for trend
##
## data: tab
## Z = -3.2735, dim = 3, p-value = 0.001062
## alternative hypothesis: two.sided
## similar to
prop.trend.test(tab[,1], apply(tab,1, sum))
##
## Chi-squared Test for Trend in Proportions
##
## data: tab[, 1] out of apply(tab, 1, sum) ,
## using scores: 1 2 3
## X-squared = 10.716, df = 1, p-value = 0.001062
MCP-Mod combines the principles of Multiple Comparisons Procedure (MCP) with modeling techniques to address the challenges in dose-response studies. This approach, introduced by Bretz et al. in 2005, enhances the flexibility in modeling dose estimation while maintaining robustness against model mis-specification associated with traditional MCP methods.
MCP-Mod is an innovative statistical methodology that has gained popularity due to its ability to produce robust statistical evidence in Phase II clinical trials, particularly concerning dose selection. Recognized and approved by both the FDA and EMA as fit-for-purpose (FFP), MCP-Mod serves as a two-step approach for analyzing dose-finding data in Phase II trials, aiming to:
1. MCP-Step (Multiple Comparisons Procedure Step): - Objective: To assess the presence of a dose-response signal effectively. - Method: Utilizes a trend test derived from a set of pre-specified candidate models. - Steps Involved: - Model Specification: Define a study population that accurately represents the underlying dose-response relationship. - Candidate Models: Pre-specify candidate dose-response models based on existing data, focusing on critical metrics such as Type I error rate, the power to detect significant dose-response, and the ability to identify the minimal effective dose. - Dose Determination and Sample Size Calculation: Establish doses and calculate sample sizes needed to meet specified performance characteristics, ensuring that the trial is neither underpowered nor excessively large.
2. Mod-Step (Modeling Step): - Objective: To find the optimal dose for confirmatory trials using a more detailed modeling approach. - Method: Involves parametric modeling or model averaging based on the best-fit model from the MCP step. - Steps Involved: - Model Selection: After conducting the MCP-step, select the most suitable model using criteria like the Akaike Information Criterion (AIC) or Bayesian Information Criterion (BIC). - Dose-Response Modeling: Perform detailed modeling of the dose-response curve using the selected model to pinpoint the optimal dosages for further trials.
Phase II clinical trials are pivotal in determining whether a new treatment or procedure has sufficient efficacy and safety profiles to proceed to further stages of drug development. These trials are critical as they often decide if a drug candidate can advance to Phase III, where more extensive testing is conducted on larger populations.
Phase IIa Trials: These trials primarily assess the feasibility of a treatment or intervention and establish preliminary evidence regarding its safety and efficacy. The focus is on understanding the biological activity of the treatment and assessing whether it performs as expected in a human population.
Phase IIb Trials: At this stage, the focus shifts to refining the understanding of the drug’s efficacy, often involving dose-ranging studies to determine the optimal dose that maximizes therapeutic benefits while minimizing adverse effects.
The two-stage design in Phase II trials is an adaptive approach that allows for early termination of the trial if the treatment proves to be ineffective or excessively toxic in the initial phase. This design is advantageous because it minimizes the exposure of patients to potentially ineffective or harmful treatments and conserves resources that could be better utilized on more promising therapies.
Key Features of the Two-Stage Design:
The method outlined by Bryant and Day (1995) serves as a classic example of how two-stage designs can be effectively implemented. Their approach allows for a comprehensive assessment of both response and toxicity, integrating these critical aspects into a unified framework that supports decision-making in clinical development.
Steps Involved in the Two-Stage Design:
Benefits of the Two-Stage Design:
Resource Efficiency: Limits the number of participants and resources expended on treatments that show early signs of being ineffective or harmful.
Ethical Considerations: Reduces patient exposure to potentially ineffective or unsafe treatments.
Data-Driven Decisions: Allows data from the initial cohort to inform adjustments and refinements, increasing the likelihood of successful outcomes in subsequent phases.