version r utils::packageVersion("SVEMnet")
SVEMnet
implements Self-Validated Ensemble Models (SVEM, Lemkus et al. 2021) and the SVEM whole model test (Karl 2024) using Elastic Net regression via the glmnet
package Friedman et al. (2010). This vignette provides an overview of the package’s functionality and usage.
The motivation to create the SVEMnet
package was primarily to have a personal sandbox to explore SVEM performance in different scenarios and with various modifications to its structure. As noted in the documentation, I used GPT o1-preview
to help form the code structure of the package and to code the Roxygen structure of the documentation. I have subsequently used more recent versions for auditing. The SVEM significance test R code comes from the supplementary material of Karl (2024). I wrote that code by hand and validated each step (not including the creation of the SVEM predictions) against corresponding results in JMP (the supplementary material of Karl (2024) provides the matching JSL script). For the SVEMnet()
code, assuming only a single value of alpha for glmnet
is being tested, the heart of the SVEM code is simply
#partial code for illustration of the SVEM loop coef_matrix <- matrix(NA, nrow = nBoot, ncol = p + 1) for (i in 1:nBoot) { U <- runif(n) w_train <- -log(U) w_valid <- -log(1 - U) #match glmnet normalization of training weight vector w_train <- w_train * (n / sum(w_train)) w_valid <- w_valid * (n / sum(w_valid)) glmnet( X, y_numeric, alpha = alpha, weights = w_train, intercept = TRUE, standardize = standardize, maxit = 1e6, nlambda = 500 ) predict(fit, newx = X) val_errors <- colSums(w_valid * (y_numeric - pred_valid)^2) k_values <- fit$df n_obs <- length(y_numeric) aic_values <- n_obs * log(val_errors / n_obs) + 2 * k_values # Choose lambda if (objective == "wSSE") { idx_min <- which.min(val_errors) lambda_opt <- fit$lambda[idx_min] val_error <- val_errors[idx_min] } else if (objective == "wAIC") { idx_min <- which.min(aic_values) lambda_opt <- fit$lambda[idx_min] val_error <- aic_values[idx_min] } coef_matrix[i, ] <- as.vector(coef(fit, s = lambda_opt)) }
However, to get this to a stable implementation that includes error and warning handling and structure to pass to S3 methods for predict()
, coef()
, plot()
, etc, it was only practical for me to utilize help from GPT o1-preview. I simply would not have taken the time to add that structure otherwise, and my implementation would have been inferior. I reviewed any of the code that was generated from this tool before integrating it, and corrected its occasional mistakes. If someone would like to create a purely human-written set of code for a similar purpose, let me know and I will be happy to add links to your package and a description to the SVEMnet
documentation.
Later revisions make use of later versions of GPT for code auditing, stress testing, and simulaiton. Many of the later entries in this vignette were written with GPT (code, analysis, summary).
library(SVEMnet) # Example data data <- iris svem_model <- SVEMnet(Sepal.Length ~ ., data = data, relaxed=FALSE,glmnet_alpha=c(1),nBoot = 50) coef(svem_model)
Generate a plot of actual versus predicted values:
plot(svem_model)
Predict outcomes for new data using the predict()
function:
predictions <- predict(svem_model, data) print(predictions)
This is the serial version of the significance test. It is slower but the code is less complicated to read than the faster parallel version.
test_result <- svem_significance_test(Sepal.Length ~ ., data = data) print(test_result) plot(test_result) SVEM Significance Test p-value: [1] 0
knitr::include_graphics("figures/whole_model_test.png")
Note that there is a parallelized version that runs much faster
test_result <- svem_significance_test_parallel(Sepal.Length ~ ., data = data) print(test_result) plot(test_result) SVEM Significance Test p-value: [1] 0
# Simulate data set.seed(1) n <- 25 X1 <- runif(n) X2 <- runif(n) X3 <- runif(n) X4 <- runif(n) X5 <- runif(n) #y only depends on X1 and X2 y <- 1 + X1 + X2 + X1 * X2 + X1^2 + rnorm(n) data <- data.frame(y, X1, X2, X3, X4, X5) # Perform the SVEM significance test test_result <- svem_significance_test_parallel( y ~ (X1 + X2 + X3)^2 + I(X1^2) + I(X2^2) + I(X3^2), data = data ) # View the p-value print(test_result) SVEM Significance Test p-value: [1] 0.009399093 test_result2 <- svem_significance_test_parallel( y ~ (X1 + X2 )^2 + I(X1^2) + I(X2^2), data = data ) # View the p-value print(test_result2) SVEM Significance Test p-value: [1] 0.006475736 #note that the response does not depend on X4 or X5 test_result3 <- svem_significance_test_parallel( y ~ (X4 + X5)^2 + I(X4^2) + I(X5^2), data = data ) # View the p-value print(test_result3) SVEM Significance Test p-value: [1] 0.8968502 # Plot the Mahalanobis distances plot(test_result,test_result2,test_result3)
knitr::include_graphics("figures/whole_model_2.png")
Newly added wrapper for cv.glmnet() to compare performance of SVEM to glmnet's native CV implementation.
Simulations show improved behavior from a relaxed grid search that allows the model to apply a lighter penalty to parameteres retained from the initial elastic net fit. This option tends to hurt RMSE on holdout data for cross validated glmnet, but the SVEM bootstraps average over the addtional variability introduced by this option and produce smaller RMSE on holdout data.
Lemkus, T., Gotwalt, C., Ramsey, P., & Weese, M. L. (2021). Self-Validated Ensemble Models for Elastic Net Regression.
Chemometrics and Intelligent Laboratory Systems, 219, 104439.
DOI: 10.1016/j.chemolab.2021.104439
Karl, A. T. (2024). A Randomized Permutation Whole-Model Test for SVEM.
Chemometrics and Intelligent Laboratory Systems, 249, 105122.
DOI: 10.1016/j.chemolab.2024.105122
Friedman, J. H., Hastie, T., & Tibshirani, R. (2010). Regularization Paths for Generalized Linear Models via Coordinate Descent.
Journal of Statistical Software, 33(1), 1–22.
DOI: 10.18637/jss.v033.i01
Gotwalt, C., & Ramsey, P. (2018). Model Validation Strategies for Designed Experiments Using Bootstrapping Techniques With Applications to Biopharmaceuticals.
JMP Discovery Conference.
Link
Ramsey, P., Gaudard, M., & Levin, W. (2021). Accelerating Innovation with Space-Filling Mixture Designs, Neural Networks, and SVEM.
JMP Discovery Conference.
Link
Ramsey, P., & Gotwalt, C. (2018). Model Validation Strategies for Designed Experiments Using Bootstrapping Techniques With Applications to Biopharmaceuticals.
JMP Discovery Summit Europe.
Link
Ramsey, P., Levin, W., Lemkus, T., & Gotwalt, C. (2021). SVEM: A Paradigm Shift in Design and Analysis of Experiments.
JMP Discovery Summit Europe.
Link
Ramsey, P., & McNeill, P. (2023). CMC, SVEM, Neural Networks, DOE, and Complexity: It's All About Prediction.
JMP Discovery Conference.
Karl, A., Wisnowski, J., & Rushing, H. (2022). JMP Pro 17 Remedies for Practical Struggles with Mixture Experiments.
JMP Discovery Conference.
Link
Xu, L., Gotwalt, C., Hong, Y., King, C. B., & Meeker, W. Q. (2020). Applications of the Fractional-Random-Weight Bootstrap.
The American Statistician, 74(4), 345–358.
Link
Karl, A. T. (2024). SVEMnet: Self-Validated Ensemble Models with Elastic Net Regression.
R package
JMP Help Documentation Overview of Self-Validated Ensemble Models.
Link
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.