.parse_evaluation_settings | R Documentation |
Internal function for parsing settings related to model evaluation
.parse_evaluation_settings(
config = NULL,
data,
parallel,
outcome_type,
hpo_metric,
development_batch_id,
vimp_aggregation_method,
vimp_aggregation_rank_threshold,
prep_cluster_method,
prep_cluster_linkage_method,
prep_cluster_cut_method,
prep_cluster_similarity_threshold,
prep_cluster_similarity_metric,
evaluate_top_level_only = waiver(),
skip_evaluation_elements = waiver(),
ensemble_method = waiver(),
evaluation_metric = waiver(),
sample_limit = waiver(),
detail_level = waiver(),
estimation_type = waiver(),
aggregate_results = waiver(),
confidence_level = waiver(),
bootstrap_ci_method = waiver(),
feature_cluster_method = waiver(),
feature_cluster_cut_method = waiver(),
feature_linkage_method = waiver(),
feature_similarity_metric = waiver(),
feature_similarity_threshold = waiver(),
sample_cluster_method = waiver(),
sample_linkage_method = waiver(),
sample_similarity_metric = waiver(),
eval_aggregation_method = waiver(),
eval_aggregation_rank_threshold = waiver(),
eval_icc_type = waiver(),
stratification_method = waiver(),
stratification_threshold = waiver(),
time_max = waiver(),
evaluation_times = waiver(),
dynamic_model_loading = waiver(),
parallel_evaluation = waiver(),
...
)
config |
A list of settings, e.g. from an xml file. |
data |
Data set as loaded using the |
parallel |
Logical value that whether familiar uses parallelisation. If
|
outcome_type |
Type of outcome found in the data set. |
hpo_metric |
Metric defined for hyperparameter optimisation. |
development_batch_id |
Identifiers of batches used for model development.
These identifiers are used to determine the cohorts used to determine a
setting for |
vimp_aggregation_method |
Method for variable importance aggregation that was used for feature selection. |
vimp_aggregation_rank_threshold |
Rank threshold for variable importance aggregation used during feature selection. |
prep_cluster_method |
Cluster method used during pre-processing. |
prep_cluster_linkage_method |
Cluster linkage method used during pre-processing. |
prep_cluster_cut_method |
Cluster cut method used during pre-processing. |
prep_cluster_similarity_threshold |
Cluster similarity threshold used during pre-processing. |
prep_cluster_similarity_metric |
Cluster similarity metric used during pre-processing. |
evaluate_top_level_only |
(optional) Flag that signals that only evaluation at the most global experiment level is required. Consider a cross-validation experiment with additional external validation. The global experiment level consists of data that are used for development, internal validation and external validation. The next lower experiment level are the individual cross-validation iterations. When the flag is Setting the flag to |
skip_evaluation_elements |
(optional) Specifies which evaluation steps,
if any, should be skipped as part of the evaluation process. Defaults to
|
ensemble_method |
(optional) Method for ensembling predictions from models for the same sample. Available methods are:
This parameter is only used if |
evaluation_metric |
(optional) One or more metrics for assessing model performance. See the vignette on performance metrics for the available metrics. Confidence intervals (or rather credibility intervals) are computed for each
metric during evaluation. This is done using bootstraps, the number of which
depends on the value of If unset, the metric in the |
sample_limit |
(optional) Set the upper limit of the number of samples that are used during evaluation steps. Cannot be less than 20. This setting can be specified per data element by providing a parameter
value in a named list with data elements, e.g.
This parameter can be set for the following data elements:
|
detail_level |
(optional) Sets the level at which results are computed and aggregated.
Note that each level of detail has a different interpretation for bootstrap
confidence intervals. For
A non-default |
estimation_type |
(optional) Sets the type of estimation that should be possible. This has the following options:
As with |
aggregate_results |
(optional) Flag that signifies whether results
should be aggregated during evaluation. If The default value is equal to As with |
confidence_level |
(optional) Numeric value for the level at which
confidence intervals are determined. In the case bootstraps are used to
determine the confidence intervals bootstrap estimation, The default value is |
bootstrap_ci_method |
(optional) Method used to determine bootstrap confidence intervals (Efron and Hastie, 2016). The following methods are implemented:
Note that the standard method is not implemented because this method is often not suitable due to non-normal distributions. The bias-corrected and accelerated (BCa) method is not implemented yet. |
feature_cluster_method |
(optional) Method used to perform clustering
of features. The same methods as for the The value for the |
feature_cluster_cut_method |
(optional) Method used to divide features
into separate clusters. The available methods are the same as for the
The value for the |
feature_linkage_method |
(optional) Method used for agglomerative
clustering with The value for the |
feature_similarity_metric |
(optional) Metric to determine pairwise
similarity between features. Similarity is computed in the same manner as
for clustering, and The value used for the |
feature_similarity_threshold |
(optional) The threshold level for
pair-wise similarity that is required to form feature clusters with the
By default, the value for the Unlike for |
sample_cluster_method |
(optional) The method used to perform
clustering based on distance between samples. These are the same methods as
for the The value for the |
sample_linkage_method |
(optional) The method used for agglomerative
clustering in The value for the |
sample_similarity_metric |
(optional) Metric to determine pairwise
similarity between samples. Similarity is computed in the same manner as for
clustering, but
The underlying feature data for numerical features is scaled to the
Regardless of metric, all categorical features are handled as for the Gower's distance: distance is 0 if the values in a pair of samples match, and 1 if they do not. |
eval_aggregation_method |
(optional) Method for aggregating variable importances for the purpose of evaluation. Variable importances are determined during feature selection steps and after training the model. Both types are evaluated, but feature selection variable importance is only evaluated at run-time. See the documentation for the |
eval_aggregation_rank_threshold |
(optional) The threshold used to define the subset of highly important features during evaluation. See the documentation for the |
eval_icc_type |
(optional) String indicating the type of intraclass
correlation coefficient ( |
stratification_method |
(optional) Method for determining the stratification threshold for creating survival groups. The actual, model-dependent, threshold value is obtained from the development data, and can afterwards be used to perform stratification on validation data. The following stratification methods are available:
One or more stratification methods can be selected simultaneously. This parameter is only relevant for |
stratification_threshold |
(optional) Numeric value(s) signifying the
sample quantiles for stratification using the The default value is This parameter is only relevant for |
time_max |
(optional) Time point which is used as the benchmark for e.g. cumulative risks generated by random forest, or the cutoff for Uno's concordance index. If This parameter is only relevant for |
evaluation_times |
(optional) One or more time points that are used for assessing calibration in survival problems. This is done as expected and observed survival probabilities depend on time. If unset, This parameter is only relevant for |
dynamic_model_loading |
(optional) Enables dynamic loading of models
during the evaluation process, if |
parallel_evaluation |
(optional) Enable parallel processing for
hyperparameter optimisation. Defaults to
|
... |
Unused arguments. |
List of parameters related to model evaluation.
Davison, A. C. & Hinkley, D. V. Bootstrap methods and their application. (Cambridge University Press, 1997).
Efron, B. & Hastie, T. Computer Age Statistical Inference. (Cambridge University Press, 2016).
Lausen, B. & Schumacher, M. Maximally Selected Rank Statistics. Biometrics 48, 73 (1992).
Hothorn, T. & Lausen, B. On the exact distribution of maximally selected rank statistics. Comput. Stat. Data Anal. 43, 121–137 (2003).
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.