Nothing
runParallel
in Gradstepx()
for higher-order derivatives, update all step...
functions using ith
is a character in Grad
, extract the gradient directly if the order is 2f(x)
takes more than 0.002 s|x| / max(stencil)
, throw a warningcontrol
or method.args
for Grad
with automatic step selectiongradstep()
explicitly!numDeriv
and check compatibility with 10 top!Grad
takes all the arguments of GenD
and Jacobian
, and vice versatodor::todor_package()
, lintr::lint_package()
, R CMD check --as-cran
, and goodpractice::gp(checks = all_checks()[!grepl("^lintr", all_checks())])
step.M
step...
functionsstep.K()
FUN(x)
is finite but FUN(x+h)
is not in all SSS routinesdiagnostics
and report
arguments; the iteration information is always saved, but not printedmax.rel.error
for all step-selection methodsv
argument for numDeriv
compatibilitycheckDimensions
could not handle character h
passed for auto-selection...
did non propagate properly to step...
functionsHessian()
with the arguments for methods "Richardson"
and "simple"
from numDeriv
Grad(sin, 1:4)
f'''
with a rule of thumbx
)step.M()
Hessian()
that supports central differences (for the moment) and arbitrary accuracyGrad()
and Jacobian()
that call the workhorse, GenD()
, for compatibility with numDeriv
step.SW()
gradstep()
solveVandermonde()
to solve ill-conditioned problems that arise in weight calculationstep.DV()
step.CR()
and its modificationGrad()
preserves the names of x
and FUN(x)
, which prevents errors in cases where names are requiredmclapply()
on *nix systems onlyAny scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.