Added DOI for JSS 2023 paper and corrected some typos in documentation
(nfold
-> nfolds
) and vignette.
Removed unneeded legacy fortran code, leaving only coxnet. Fixed up Matrix as() sequences
Relatively minor changes to bugs in survival functions and bigGlm, and some improved failure messages.
Most of the Fortran code has been replaced by C++ by James Yang, leading to speedups in all cases. The exception is the Cox routine for right censored data, which is still under development.
Some of the Fortran in glmnet has been replaced by C++, written by the
newest member of our team, James Yang.
the wls
routines (dense and sparse), that are the engines under the glmnet.path
function when we use programmable families, are now written in C++, and lead
to speedups of around 8x.
the family of elnet routines (sparse/dense, covariance/naive) for
glmnet(...,family="gaussian")
are all in C++, and lead to speedups
around 4x.
A new feature added, as well as some minor fixes to documentation.
The exclude argument has come to life. Users can now pass a function
that can take arguments x, y and weights, or a subset of these, for
filtering variables. Details in documentation and vignette.
Prediction with single newx
observation failed before. This is
fixed.
Labeling of predictions from cv.glmnet
improved.
Fixed a bug in mortran/fortran that caused program to loop ad infinitum
Fixed some bugs in the coxpath function to do with sparse X.
when some penalty factors are zero, and X is sparse, we should not
call GLM to get the start
apply does not work as intended with sparse X, so we now use matrix
multiplies instead in computing lambda_max
* added documentation for cv.glmnet
to explain implications of
supplying lambda
Expanded scope for the Cox model.
We now allow (start, stop) data in
addition to the original right-censored all start at zero option.
Allow for strata as in survival::coxph
Allow for sparse X matrix with Cox models (was not available before)
Provide method for survival::survfit
Vignettes are revised and reorganized.
Additional index information stored on cv.glmnet
objects, and
included when printed.
concordance
function from package survival
Major revision with added functionality. Any GLM family can be used
now with glmnet
, not just the built-in families. By passing a
"family" object as the family argument (rather than a character
string), one gets access to all families supported by glm
. This
development was programmed by our newest member of the glmnet
team,
Kenneth Tay.
Bug fixes
Intercept=FALSE
with "Gaussian" is fixed. The dev.ratio
comes out
correctly now. The mortran code was changed directly in 4
places. look for "standard". Thanks to Kenneth Tay.Bug fixes
confusion.glmnet
was sometimes not returning a list because of
apply collapsing structurecv.mrelnet
and cv.multnet
dropping dimensions inappropriatelystorePB
to avoid segfault. Thanks Tomas Kalibera!assess.glmnet
and cousins to be more helpful!lambda.interp
to avoid edge cases (thanks
David Keplinger)Minor fix to correct Depends in the DESCRIPTION to R (>= 3.6.0)
This is a major revision with much added functionality, listed
roughly in order of importance. An additional vignette called relax
is supplied to describe the usage.
relax
argument added to glmnet
. This causes the
models in the path to be refit without regularization. The
resulting object inherits from class glmnet
, and has an
additional component, itself a glmnet object, which is the relaxed
fit.relax
argument to cv.glmnet
. This allows selection from a
mixture of the relaxed fit and the regular fit. The mixture is
governed by an argument gamma
with a default of 5 values between
0 and 1. predict
, coef
and plot
methods for relaxed
and cv.relaxed
objects.print
method for relaxed
object, and new print
methods for
cv.glmnet
and cv.relaxed
objects.trace.it=TRUE
to glmnet
and cv.glmnet
. This can also be set
for the session via glmnet.control
. assess.glmnet
, roc.glmnet
and
confusion.glmnet
for displaying the performance of models.makeX
for building the x
matrix for input to glmnet
. Main
functionality is one-hot-encoding of factor variables, treatment
of NA
and creating sparse inputs.bigGlm
for fitting the GLMs of glmnet
unpenalized.In addition to these new features, some of the code in glmnet
has
been tidied up, especially related to CV.
coxnet.deviance
to do with input
pred
, as well as saturated loglike
(missing) and weightscoxgrad
function for computing the gradientcv.glmnet
, for cases when wierd
things happeninst/mortran
inst/mortran
-Wall
warningsnewoffset
created problems all over - fixed theseexact=TRUE
calls to coef
and
predict
. See help file for more detailsy
blows up elnet
; error trap includedlambda.interp
which was returning NaN
under degenerate
circumstances.Surv
objectpredict
and coef
with exact=TRUE
. The
user is strongly encouraged to supply the original x
and y
values, as well as any other data such as weights that were used in
the original fit.lognet
when some weights are zero and x
is sparsepredict.glmnet
, predict.multnet
and predict.coxnet
,
when s=
argument is used with a vector of values. It was not doing
the matrix multiply correctlyintercept
optionglmnet.control
for setting systems parameterscoxnet
exact=TRUE
option for prediction and coef functionsmgaussian
family for multivariate responsegrouped
option for multinomial familynewx
and make dgCmatrix
if sparselognet
added a classnames component to the objectpredict.lognet(type="class")
now returns a character vector/matrixpredict.glmnet
: fixed bug with type="nonzero"
glmnet
: Now x can inherit from sparseMatrix
rather than the very
specific dgCMatrix
, and this will trigger sparse mode for glmnetglmnet.Rd
(lambda.min
) : changed value to 0.01 if nobs < nvars
,
(lambda
) added warnings to avoid single value,
(lambda.min
): renamed it lambda.min.ratio
glmnet
(lambda.min
) : changed value to 0.01 if nobs < nvars
(HessianExact
) : changed the sense (it was wrong),
(lambda.min
): renamed it lambda.min.ratio
. This allows it to be
called lambda.min
in a call though predict.cv.glmnet
(new function) : makes predictions directly from
the saved glmnet
object on the cv object coef.cv.glmnet
(new function) : as abovepredict.cv.glmnet.Rd
: help functions for the abovecv.glmnet
: insert drop(y)
to avoid 1 column matrices; now include a glmnet.fit
object for later predictionsnonzeroCoef
: added a special case for a single variable in x
; it was dying on thisdeviance.glmnet
: includeddeviance.glmnet.Rd
: includedglmnet_1.4
.Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.