knitr::opts_chunk$set( echo=TRUE, collapse=TRUE, comment="#>", fig.path="man/figures/" )

This package implements the ExclusiveLasso
penalty of Zhou, Jin, and Hoi (2010)
and Campbell and Allen (2017) for generalized linear models.
[\text{arg min}{\beta} \frac{1}{2n} \|y - X\beta\|_2^2 + \lambda \sum{g \in \mathcal{G}} \frac{\|\beta_g\|_1^2}{2}]
This penalty is the "converse" of the group lasso, encouraging selection of a single variable in each group. See Campbell and Allen (2017) for a thorough discussion of this estimator and its properties.
The package provides efficient inexact proximal gradient and coordinate descent
schemes to solve exclusive lasso problems. The interface is similar to that of
the popular glmnet
,
ncvreg
, and grpreg
packages.
The current working version of the package can be installed from Github:
library(devtools) install_github("DataSlingers/ExclusiveLasso")
We begin by simulating a small data set with simple structure:
library(ExclusiveLasso) n <- 200 p <- 500 groups <- rep(1:10, times=50) beta <- numeric(p); beta[1:10] <- 3 X <- matrix(rnorm(n * p), ncol=p) y <- X %*% beta + rnorm(n)
We fit the exclusive lasso to this data set, using a user-specified group structure:
exfit <- exclusive_lasso(X, y, groups) print(exfit) plot(exfit)
As we can see, for this very simple problem, the exclusive lasso picked out the true variables (though the standard lasso would have done as well here).
The cv.exclusive_lasso
function can be used to select the tuning parameter $\lambda$,
though as Campbell and Allen (2017) note, standard cross-validation does not perform
particularly well for this problem, and model selection according to BIC / EBIC
with a group-thresholding step yields superior results. To facilitate model selection
by BIC / EBIC, an unbiased estimate of the degrees of freedom is calculated.
In addition to standard linear regression, the ExclusiveLasso
package also implements
logistic and Poisson regression. See the package vignette for details.
Department of Statistics, Rice University
Department of Statistics, Rice University
Departments of Statistics, CS,and ECE, Rice University
Jan and Dan Duncan Neurological Research Institute Baylor College of Medicine and Texas Children's Hospital
MW was supported by NSF Graduate Research Fellowship No.1450681.
FC was supported by NSF Graduate Research Fellowship No.0940902.
GA acknowledges funding from NSF/DMS-1264058 and NSF/DMS-1209017.
Campbell, Frederick, and Genevera I. Allen. 2017. "Within-Group Variable Selection Through the Exclusive Lasso." Electronic Journal of Statistics 11(2): 4220–4257. doi:10.1214/17-EJS1317.
Zhou, Yang, Rong Jin, and Steven C.H. Hoi. 2010. "Exclusive Lasso for Multi-Task Feature Selection." In AISTATS 2010: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, edited by Yee Whye Teh and Mike Titterington. JMLR. http://proceedings.mlr.press/v9/zhou10a.html.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.