Nothing
A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization algorithm. MADGRAD is a 'best-of-both-worlds' optimizer with the generalization performance of stochastic gradient descent and at least as fast convergence as that of Adam, often faster. A drop-in optim_madgrad() implementation is provided based on Defazio et al (2020) <arxiv:2101.11075>.
Package details |
|
---|---|
Author | Daniel Falbel [aut, cre, cph], RStudio [cph], MADGRAD original implementation authors. [cph] |
Maintainer | Daniel Falbel <daniel@rstudio.com> |
License | MIT + file LICENSE |
Version | 0.1.0 |
Package repository | View on CRAN |
Installation |
Install the latest version of this package by entering the following in R:
|
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.