glm_studt | R Documentation |
Program for fitting a GLM equipped with a regularized student-t prior on the regression coefficients, parametrized using the normal-inverse-gamma distribution. The 'regularization' refers to the fact that the inverse-gamma scale is has a finite upper bound that it smoothly approaches. This method was not used in the simulation study but was used in the data analysis. Specifically, it corresponds to 'PedRESC2'.
glm_studt(
y,
x_standardized,
family = "binomial",
beta_scale,
dof = 1,
slab_dof = Inf,
slab_scale = 15,
mu_sd = 5,
only_prior = F,
mc_warmup = 1000,
mc_iter_after_warmup = 1000,
mc_chains = 1,
mc_thin = 1,
mc_stepsize = 0.1,
mc_adapt_delta = 0.9,
mc_max_treedepth = 15,
return_as_CmdStanMCMC = FALSE,
seed = sample.int(.Machine$integer.max, 1),
slab_precision = NULL
)
y |
(vector) outcomes corresponding to the type of glm desired. This should match whatever datatype is expected by the stan program. |
x_standardized |
(matrix) matrix of numeric values with number of rows equal to the length of y and number of columns equal to p+q. It is assumed without verification that each column is standardized to whatever scale the prior expects - in Boonstra and Barbaro, all predictors are marginally generated to have mean zero and unit variance, so no standardization is conducted. In practice, all data should be standardized to have a common scale before model fitting. If regression coefficients on the natural scale are desired, they be easily obtained through unstandardizing. |
family |
(character) Similar to argument in |
beta_scale |
(pos. real) constants indicating the prior scale of the student-t prior. |
dof |
(pos. integer) degrees of freedom for the student-t prior |
slab_dof |
see |
slab_scale |
(pos. real) these control the slab-part of the regularized
horseshoe. Specifically, in the notation of Boonstra and Barbaro,
d^2~InverseGamma( |
mu_sd |
(pos. real) the prior standard deviation for the intercept parameter mu |
only_prior |
(logical) should all data be ignored, sampling only from the prior? |
mc_warmup |
number of MCMC warm-up iterations |
mc_iter_after_warmup |
number of MCMC iterations after warm-up |
mc_chains |
number of MCMC chains |
mc_thin |
every nth draw to keep |
mc_stepsize |
positive stepsize |
mc_adapt_delta |
between 0 and 1 |
mc_max_treedepth |
max tree depth |
return_as_CmdStanMCMC |
(logical) should the function return the CmdStanMCMC object asis or should a summary of CmdStanMCMC be returned as a regular list |
seed |
seed for the underlying STAN model to allow for reproducibility |
slab_precision |
(pos. real) the slab-part of the regularized horseshoe,
this is equivalent to (1/d)^2 in the notation of Boonstra and Barbaro. If
specified, it is assumed that you want a fixed slab component and will take
precedence over any provided values of |
list
object containing the draws and other information.
data(historical)
foo = glm_studt(y = historical$y_hist,
x_standardized = historical[,2:5],
family = "binomial",
beta_scale = 0.0231,
dof = 1,
mu_sd = 5,
only_prior = 0,
mc_warmup = 200,
mc_iter_after_warmup = 200,
mc_chains = 2,
mc_thin = 1,
mc_stepsize = 0.1,
mc_adapt_delta = 0.99,
mc_max_treedepth = 15);
data(current)
foo = glm_studt(y = current$y_curr,
x_standardized = current[,2:11],
family = "binomial",
beta_scale = 0.0231,
dof = 1,
mu_sd = 5,
only_prior = 0,
mc_warmup = 200,
mc_iter_after_warmup = 200,
mc_chains = 2,
mc_thin = 1,
mc_stepsize = 0.1,
mc_adapt_delta = 0.99,
mc_max_treedepth = 15);
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.