The function sets, shows, or clears an option that is specific to the optimizer in the back-end.
An MxModel object or NULL
The name of the option.
The value of the option.
If TRUE then reset all options to their defaults.
mxOption is used to set, clear, or query an option (given in the ‘key’ argument) in the back-end optimizer. Valid option keys are listed below.
Use value = NULL to remove an existing option. Leaving value blank will return the current value of the option specified by ‘key’.
To reset all options to their default values, use ‘reset = TRUE’. When reset = TRUE, ‘key’ and ‘value’ are ignored.
If the ‘model’ argument is set to NULL, the default optimizer option (i.e those applying to all models by default) will be set.
To see the defaults, use
Before the model is submitted to the back-end, all keys and values are converted into strings using the as.character function.
The “Default optimizer” option can only be set globally (i.e., with
model=NULL), and not locally (i.e., specifically to a given MxModel). Although the checkpointing options may be set globally, OpenMx's behavior is only affected by locally set checkpointing options (that is, global checkpointing options are ignored at runtime).
Options “Gradient step size”, “Gradient iterations”, and “Function precision” have on-load global defaults of
"Auto". If value
"Auto" is in effect for any of these three options at runtime, then OpenMx selects a reasonable numerical value in its place. These automated numerical values are intended to (1) adjust for the limited precision of the algorithm for computing multivariate-normal probability integrals, and (2) calculate accurate numeric derivatives at the optimizer's solution. If the user replaces
"Auto" with a valid numerical value, then OpenMx uses that value as-is.
Some options only affect certain optimizers. Currently, options “Gradient algorithm” and “Gradient iterations” are ignored by all optimizers other than SLSQP. Also, currently option “Analytic Gradients” only affects SLSQP and NPSOL. Option “Gradient step size” is used slightly differently by SLSQP and CSOLNP, and is ignored by NPSOL (see
mxComputeGradientDescent() for details). If the MxModel containsMxConstraints, OpenMx's interface to NPSOL scales “Feasibility tolerance” by multiplying it by 0.02 and dividing it by 0.05 before passing it to NPSOL; if there are no MxConstraints, OpenMx passes NPSOL a hardcoded value of 1e-5. Note that NPSOL's criterion for returning a status code of 0 versus 1 for a given solution depends partly on “Optimality tolerance”.
For a block of
n ordinal variables, the maximum number of integration points that OpenMx may use to calculate multivariate-normal probability integrals is given by
mvnMaxPointsA + mvnMaxPointsB*n + mvnMaxPointsC*n*n
The maximum number of major iterations (the option “Major iterations”) for optimization for NPSOL can be specified either by using a numeric value (such as 50, 1000, etc) or by specifying a user-defined function. The user-defined function should accept two arguments as input, the number of parameters and the number of constraints, and return a numeric value as output.
|Calculate Hessian||[Yes | No]||calculate the Hessian explicitly after optimization.|
|Standard Errors||[Yes | No]||return standard error estimates from the explicitly calculate hessian.|
|Default optimizer||[NPSOL | SLSQP | CSOLNP]||the gradient-descent optimizer to use|
|Number of Threads||[0|1|2|...|10|...]||number of threads used for optimization. Default value is taken from the environment variable OMP_NUM_THREADS or, if that is not set, 2.|
|Feasibility tolerance||r||the maximum acceptable absolute violations in linear and nonlinear constraints.|
|Optimality tolerance||r||the accuracy with which the final iterate approximates a solution to the optimization problem; roughly, the number of reliable significant figures that the fitfunction value should have at the solution.|
|Gradient algorithm||see list||finite difference method, either 'forward' or 'central'.|
|Gradient iterations||1:4||the number of Richardson extrapolation iterations|
|Gradient step size||r||amount of change made to free parameters when numerically calculating gradient|
|Analytic Gradients||[Yes | No]||should the optimizer use analytic gradients (if available)?|
|loglikelihoodScale||i||factor by which the loglikelihood is scaled.|
|Parallel diagnostics||[Yes | No]||whether to issue diagnostic messages about use of multiple threads|
|Nudge zero starts||[TRUE | FALSE]||Should OpenMx "nudge"" starting values of zero to 0.1 at runtime?|
|Nolist||this option suppresses printing of the options|
|Print level||i||the value of i controls the amount of printout produced by the major iterations|
|Minor print level||i||the value of i controls the amount of printout produced by the minor iterations|
|Print file||i||for i > 0 a full log is sent to the file with logical unit number i.|
|Summary file||i||for i > 0 a brief log will be output to file i.|
|Function precision||r||a measure of accuracy with which the fitfunction and constraint functions can be computed.|
|Infinite bound size||r||if r > 0 defines the "infinite" bound bigbnd.|
|Major iterations||i or a function||the maximum number of major iterations before termination.|
|Verify level||[-1:3 | Yes | No]||see NPSOL manual.|
|Line search tolerance||r||controls the accuracy with which a step is taken.|
|Derivative level||[0-3]||see NPSOL manual.|
|Hessian||[Yes | No]||return the Hessian (Yes) or the transformed Hessian (No).|
|Step Limit||r||maximum change in free parameters at first step of linesearch.|
|Always Checkpoint||[Yes | No]||whether to checkpoint all models during optimization.|
|Checkpoint Directory||path||the directory into which checkpoint files are written.|
|Checkpoint Prefix||string||the string prefix to add to all checkpoint filenames.|
|Checkpoint Fullpath||path||overrides the directory and prefix (useful to output to /dev/fd/2)|
|Checkpoint Units||see list||the type of units for checkpointing: 'minutes', 'iterations', or 'evaluations'.|
|Checkpoint Count||i||the number of units between checkpoint intervals.|
Model transformation options
|Error Checking||[Yes | No]||whether model consistency checks are performed in the OpenMx front-end|
|No Sort Data||character vector of model names for which FIML data sorting is not performed|
|RAM Inverse Optimization||[Yes | No]||whether to enable solve(I - A) optimization|
|RAM Max Depth||i||the maximum depth to be used when solve(I - A) optimization is enabled|
Multivariate normal integration parameters
|maxOrdinalPerBlock||i||maximum number of ordinal variables to evaluate together|
|mvnMaxPointsA||i||base number of integration points|
|mvnMaxPointsB||i||number of integration points per row|
|mvnMaxPointsC||i||number of integration points per rows^2|
|mvnAbsEps||i||absolute error tolerance|
|mvnRelEps||i||relative error tolerance|
If a model is provided, it is returned with the optimizer option either set or cleared. If value is empty, the current value is returned.
The OpenMx User's guide can be found at http://openmx.ssri.psu.edu/documentation.
mxModel(), as almost all uses of
mxOption() are via an mxModel whose options are set or cleared. See
mxComputeGradientDescent() for details on how different optimizers are affected by different options.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
# set the Numbder of Threads (cores to use) mxOption(NULL, "Number of Threads", parallel::detectCores() - 1) testModel <- mxModel(model = "testModel") # make a model to use for example testModel$options # show the model options (none yet) options()$mxOptions # list all mxOptions (global settings) testModel <- mxOption(testModel, "Function precision", 1e-5) # set precision testModel <- mxOption(testModel, "Function precision", NULL) # clear precision # N.B. This is model-specific precision (defaults to global setting) # may optimize for speed # at cost of not getting standard errors testModel <- mxOption(testModel, "Calculate Hessian", "No") testModel <- mxOption(testModel, "Standard Errors" , "No") testModel$options # see the list of options you set
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.