library(gsDesign2) library(simtrial) library(dplyr) library(gt) library(doFuture) library(tibble) set.seed(2025)
The vignette Simulate Fixed Designs with Ease via sim_fixed_n presents fixed design simulations using a single function call, sim_fixed_n()
. It offers a simple and straightforward process for running simulations quickly.
If users are interested in the following aspects, we recommend simulating from scratch rather than directly using sim_fixed_n()
:
The process for simulating from scratch is outlined in Steps 1 to 5 below.
The sim_pw_surv()
function allows for the simulation of a clinical trial with essentially arbitrary patterns of enrollment, failure rates, and censoring. To implement sim_pw_surv()
, you need to specify 5 design characteristics to simulate time-to-event data:
n
).stratum
).block
). The sim_pw_surv()
function uses fixed block randomization.enroll_rate
). The sim_pw_surv()
function supports piecewise enrollment, allowing the enrollment rate to be piecewise constant.fail_rate
) or time-to-event rate. The sim_pw_surv()
function uses a piecewise exponential distribution for the failure rate, which makes it easy to define a distribution with changing failure rates over time. Specifically, in the $j$-th interval, the rate is denoted as $\lambda_j \geq 0$. We require that at least one interval has $\lambda_j > 0$. There are two methods for defining the failure rate:fail_rate
using gsDesign2::define_fail_rate
, and then convert it to the required format using to_sim_pw_surv()
. An example is provided in Scenario a).dropout_Rate
). The sim_pw_surv()
function accepts piecewise constant dropout rates, which may vary by treatment group. The configuration for dropout should be specified by treatment group, stratum, and time period, and setting up the dropout rate follows the same approach as the failure rate.We begin with the simplest implementation of sim_pw_surv()
. The following lines of code will generate 500 subjects using equal randomization and an unstratified design.
n_sim <- 100 n <- 500 stratum <- data.frame(stratum = "All", p = 1) block <- rep(c("experimental", "control"), 2) enroll_rate <- define_enroll_rate(rate = 12, duration = n / 12) fail_rate <- define_fail_rate(duration = c(6, Inf), fail_rate = log(2) / 10, hr = c(1, 0.7), dropout_rate = 0.0001) uncut_data_a <- sim_pw_surv(n = n, stratum = stratum, block = block, enroll_rate = enroll_rate, fail_rate = to_sim_pw_surv(fail_rate)$fail_rate, dropout_rate = to_sim_pw_surv(fail_rate)$dropout_rate)
The output of sim_pw_surv()
is subject-level observations, including stratum, enrollment time for the observation, treatment group the observation is randomized to, failure time, dropout time, calendar time of enrollment plot the minimum of failure time and dropout time ( cte
), and an failure and dropout indicator (fail = 1
is a failure, fail = 0
is a dropout).
uncut_data_a |> head() |> gt() |> tab_header("An Overview of Simulated TTE data")
The dropout rate can differ between groups. For instance, in open-label studies, the control group may experience a higher dropout rate. The follow lines of code assumes the control group has a dropout rate of 0.002 for the first 10 months, which then decreases to 0.001 thereafter. In contrast, the experimental group has a constant dropout rate of 0.001 throughout the study.
differential_dropout_rate <- data.frame( stratum = rep("All", 3), period = c(1, 2, 1), treatment = c("control", "control", "experimental"), duration = c(10, Inf, Inf), rate = c(0.002, 0.001, 0.001)) uncut_data_b <- sim_pw_surv(n = n, stratum = stratum, block = block, enroll_rate = enroll_rate, fail_rate = to_sim_pw_surv(fail_rate)$fail_rate, dropout_rate = differential_dropout_rate)
The following code assumes there are two strata (biomarker-positive and biomarker-negative) with equal prevalence of 0.5 for each. In the control arm, the median survival time is 10 months for biomarker-positive subjects and 8 months for biomarker-negative subjects. For both strata, the hazard ratio is 1 for the first 3 months, after which it decreases to 0.6 for biomarker-positive subjects and 0.8 for biomarker-negative subjects. The dropout rate is contently 0.001 for both strata over time.
stratified_enroll_rate <- data.frame( stratum = c("Biomarker positive", "Biomarker negative"), rate = c(12, 12), duration = c(1, 1)) stratified_fail_rate <- data.frame( stratum = c(rep("Biomarker positive", 3), rep("Biomarker negative", 3)), period = c(1, 1, 2, 1, 1, 2), treatment = rep(c("control", "experimental", "experimental"), 2), duration = c(Inf, 3, Inf, Inf, 3, Inf), rate = c(# failure rate of biomarker positive subjects: control arm, exp arm period 1, exp arm period 2 log(2) / 10, log(2) /10, log(2) / 10 * 0.6, # failure rate of biomarker negative subjects: control arm, exp arm period 1, exp arm period 2 log(2) / 8, log(2) /8, log(2) / 8 * 0.8) ) stratified_dropout_rate <- data.frame( stratum = rep(c("Biomarker positive", "Biomarker negative"), each = 2), period = c(1, 1, 1, 1), treatment = c("control", "experimental", "control", "experimental"), duration = rep(Inf, 4), rate = rep(0.001, 4) ) uncut_data_c <- sim_pw_surv(n = n, stratum = data.frame(stratum = c("Biomarker positive", "Biomarker negative"), p = c(0.5, 0.5)), block = block, enroll_rate = stratified_enroll_rate, fail_rate = stratified_fail_rate, dropout_rate = stratified_dropout_rate )
Suppose you wish to have 3 arms: control, low-dose and high-dose. The following code assumes the control arm has a median survival time of 10 months, the low-dose arm has a median survival time of 12 months, and the high-dose arm has a median survival time of 14 months. The hazard ratio for the low-dose arm is 0.8, and the hazard ratio for the high-dose arm is 0.6. The dropout rate is 0.001 for all arms. Block size is 7 with 3:2:2 randomization.
We begin by setting up enrollment, failure and dropout rates.
enroll_rate <- define_enroll_rate(rate = 12, duration = n / 12) three_arm_fail_rate <- data.frame( stratum = "All", period = c(1, 1, 2, 1, 2), treatment = c("control", "low-dose", "low-dose", "high-dose", "high-dose"), duration = c(Inf, 3, Inf, 3, Inf), rate = c(# failure rate of control arm: period 1, period 2 log(2) / 10, # failure rate of low-dose arm: period 1, period 2 log(2) / c(10, 10 / .8), # failure rate of high-dose arm: period 1, period 2 log(2) / c(10, 10 / .6))) three_arm_dropout_rate <- data.frame( stratum = "All", period = c(1, 1, 1), treatment = c("control", "low-dose", "high-dose"), duration = rep(Inf, 3), rate = rep(0.001, 3)) uncut_data_d <- sim_pw_surv(n = n, stratum = data.frame(stratum = "All"), block = c(rep("control", 3), rep("low-dose", 2), rep("high-dose", 2)), enroll_rate = enroll_rate, fail_rate = three_arm_fail_rate, dropout_rate = three_arm_dropout_rate)
For illustration purposes, we will focus on scenario b) for the following discussion.
uncut_data <- uncut_data_b
The get_analysis_date()
derives analysis date for interim/final analysis given multiple conditions, see the help page of get_analysis_date()
at the pkgdown website.
Users can cut for analysis at the 24th month and there are 300 events, whichever arrives later. This is equivalent to timing_type = 4
in sim_fixed_n()
.
cut_date_a <- get_analysis_date(data = uncut_data, planned_calendar_time = 24, target_event_overall = 300)
Users can also cut by the maximum of targeted 300 event and minimum follow-up 12 months. This is equivalent to timing_type = 5
in sim_fixed_n()
.
cut_date_b <- get_analysis_date(data = uncut_data, min_followup = 12, target_event_overall = 300)
Users can cut data when there are 300 events, with maximum time extension to reach targeted events of 24 months. This is not enabled in timing_type
of sim_fixed_n()
.
cut_date_c <- get_analysis_date(data = uncut_data, max_extension_for_target_event = 12, target_event_overall = 300)
Users can cut data after 12 months followup when 80\% of the patients are enrolled in the overall population as below. This is not enabled in timing_type
of sim_fixed_n()
.
cut_date_d <- get_analysis_date(data = uncut_data, min_n_overall = 100 * 0.8, min_followup = 12)
More examples are available in the reference page of get_analysis_date()
For illustration purposes, we will focus on scenario d) for the following discussion.
cut_date <- cut_date_d cat("The cutoff date is ", round(cut_date, 2)) cut_data <- uncut_data |> cut_data_by_date(cut_date) cut_data |> head() |> gt() |> tab_header(paste0("An Overview of TTE data Cut at ", round(cut_date, 2), "Months"))
The simtrial package provides many options for testing methods, including (weighted) logrank tests, RMST test, milestone test, and MaxComboi test, see the [Section "Compute p-values/test statistics" at the pkgdown reference page] (https://merck.github.io/simtrial/reference/index.html#compute-p-values-test-statistics).
The following code lists all possible tests available in simtrial. Users can select one of the tests listed above or combine the testing results to make comparisons across tests. For demonstration purposes, we will aggregate all tests together.
# Logrank test sim_res_lr <- cut_data |> wlr(weight = fh(rho = 0, gamma = 0)) # weighted logrank test by Fleming-Harrington weights sim_res_fh <- cut_data |> wlr(weight = fh(rho = 0, gamma = 0.5)) # Modestly weighted logrank test sim_res_mb <- cut_data |> wlr(weight = mb(delay = Inf, w_max = 2)) # Weighted logrank test by Xu 2017's early zero weights sim_res_xu <- cut_data |> wlr(weight = early_zero(early_period = 3)) # RMST test sim_res_rmst <- cut_data |> rmst(tau = 10) # Milestone test sim_res_ms <- cut_data |> milestone(ms_time = 10) # Maxcombo tests comboing multiple weighted logrank test with Fleming-Harrington weights sim_res_mc <- cut_data |> maxcombo(rho = c(0, 0), gamma = c(0, 0.5))
The output of the tests mentioned above are lists including:
sim_res_rmst$method
.sim_res_rmst$parameter
.sim_res_rmst$estimate
and sim_res_rmst$se
.sim_res_rmst$z
. Please note that the Z-score is not provided for the MaxCombo test; instead, the p-value is reported (sim_res_mc$p_value
).sim_res <- tribble( ~Method, ~Parameter, ~Z, ~Estimate, ~SE, ~`P value`, sim_res_lr$method, sim_res_lr$parameter, sim_res_lr$z, sim_res_lr$estimate, sim_res_lr$se, pnorm(-sim_res_lr$z), sim_res_fh$method, sim_res_fh$parameter, sim_res_fh$z, sim_res_fh$estimate, sim_res_fh$se, pnorm(-sim_res_fh$z), sim_res_mb$method, sim_res_mb$parameter, sim_res_mb$z, sim_res_mb$estimate, sim_res_mb$se, pnorm(-sim_res_mb$z), sim_res_xu$method, sim_res_xu$parameter, sim_res_xu$z, sim_res_xu$estimate, sim_res_xu$se, pnorm(-sim_res_xu$z), sim_res_rmst$method, sim_res_rmst$parameter|> as.character(), sim_res_rmst$z, sim_res_rmst$estimate, sim_res_rmst$se, pnorm(-sim_res_rmst$z), sim_res_ms$method, sim_res_ms$parameter |> as.character(), sim_res_ms$z, sim_res_ms$estimate, sim_res_ms$se, pnorm(-sim_res_ms$z), sim_res_mc$method, sim_res_mc$parameter, NA, NA, NA, sim_res_mc$p_value ) sim_res |> gt() |> tab_header("One Simulation Results")
We will now merge Steps 1 to 3 into a single function named one_sim()
, which facilitates a single simulation run. The construction of one_sim()
involves copying all the lines of code from Steps 1 to 3.
one_sim <- function(sim_id = 1, # arguments from Step 1: design characteristic n, stratum, enroll_rate, fail_rate, dropout_rate, block, # arguments from Step 2; cutting method min_n_overall, min_followup, # arguments from Step 3; testing method fh, mb, xu, rmst, ms, mc ) { # Step 1: simulate a time-to-event data uncut_data <- sim_pw_surv( n = n, stratum = stratum, block = block, enroll_rate = enroll_rate, fail_rate = fail_rate, dropout_rate = dropout_rate) ## Step 2: Cut data cut_date <- get_analysis_date(min_n_overall = min_n_overall, min_followup = min_followup, data = uncut_data) cut_data <- uncut_data |> cut_data_by_date(cut_date) # Step 3: Run tests sim_res_lr <- cut_data |> wlr(weight = fh(rho = 0, gamma = 0)) sim_res_fh <- cut_data |> wlr(weight = fh(rho = fh$rho, gamma = fh$gamma)) sim_res_mb <- cut_data |> wlr(weight = mb(delay = mb$delay, w_max = mb$w_max)) sim_res_xu <- cut_data |> wlr(weight = early_zero(early_period = xu$early_period)) sim_res_rmst <- cut_data |> rmst(tau = rmst$tau) sim_res_ms <- cut_data |> milestone(ms_time = ms$ms_time) sim_res_mc <- cut_data |> maxcombo(rho = mc$rho, gamma = mc$gamma) sim_res <- tribble( ~`Sim ID`, ~Method, ~Parameter, ~Z, ~Estimate, ~SE, ~`P value`, sim_id, sim_res_lr$method, sim_res_lr$parameter, sim_res_lr$z, sim_res_lr$estimate, sim_res_lr$se, pnorm(-sim_res_lr$z), sim_id, sim_res_fh$method, sim_res_fh$parameter, sim_res_fh$z, sim_res_fh$estimate, sim_res_fh$se, pnorm(-sim_res_fh$z), sim_id, sim_res_mb$method, sim_res_mb$parameter, sim_res_mb$z, sim_res_mb$estimate, sim_res_mb$se, pnorm(-sim_res_mb$z), sim_id, sim_res_xu$method, sim_res_xu$parameter, sim_res_xu$z, sim_res_xu$estimate, sim_res_xu$se, pnorm(-sim_res_xu$z), sim_id, sim_res_rmst$method, sim_res_rmst$parameter|> as.character(), sim_res_rmst$z, sim_res_rmst$estimate, sim_res_rmst$se, pnorm(-sim_res_rmst$z), sim_id, sim_res_ms$method, sim_res_ms$parameter |> as.character(), sim_res_ms$z, sim_res_ms$estimate, sim_res_ms$se, pnorm(-sim_res_ms$z), sim_id, sim_res_mc$method, sim_res_mc$parameter, NA, NA, NA, sim_res_mc$p_value ) return(sim_res) }
After that, we will execute one_sim()
multiple times using parallel computation. The following lines of code uses 2 workers to run 100 simulations.
set.seed(2025) plan("multisession", workers = 2) ans <- foreach( sim_id = seq_len(n_sim), .errorhandling = "stop", .options.future = list(seed = TRUE) ) %dofuture% { ans_new <- one_sim( sim_id = sim_id, # arguments from Step 1: design characteristic n = n, stratum = stratum, enroll_rate = enroll_rate, fail_rate = to_sim_pw_surv(fail_rate)$fail_rate, dropout_rate = differential_dropout_rate, block = block, # arguments from Step 2; cutting method min_n_overall = 500 * 0.8, min_followup = 12, # arguments from Step 3; testing method fh = list(rho = 0, gamma = 0.5), mb = list(delay = Inf, w_max = 2), xu = list(early_period = 3), rmst = list(tau = 10), ms = list(ms_time = 10), mc = list(rho = c(0, 0), gamma = c(0, 0.5)) ) ans_new } ans <- data.table::rbindlist(ans) plan("sequential")
The output from the parallel computation resembles the output of sim_fix_n()
described in the vignette Simulate Fixed Designs with Ease via sim_fixed_n. Each row in the output corresponds to the simulation results for each testing method per each repeation.
ans |> head() |> gt() |> tab_header("Overview Each Simulation results")
Using the 100 parallel simulations provided above, users can summarize the simulated power and compare it across different testing methods with some data manipulation using dplyr
. Please note that the power calculation for the MaxCombo test differs from the other tests, as it does not report a Z-score.
ans_non_mc <- ans |> filter(Method != "MaxCombo") |> group_by(Method, Parameter) %>% summarise(Power = mean(Z > -qnorm(0.025))) |> ungroup() ans_mc <- ans |> filter(Method == "MaxCombo") |> summarize(Power = mean(`P value` < 0.025), Method = "MaxCombo", Parameter = "FH(0, 0) + FH(0, 0.5)") ans_non_mc |> union(ans_mc) |> gt() |> tab_header("Summary from 100 simulations")
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.