Description Usage Arguments Value See Also
This function is a very parallel version of benchmark using batchtools.
Experiments are created in the provided registry for each combination of
learners, tasks and resamplings. The experiments are then stored in a registry and the
runs can be started via submitJobs. A job is one train/test split
of the outer resampling. In case of nested resampling (e.g. with makeTuneWrapper),
each job is a full run of inner resampling, which can be parallelized in a second step
with ParallelMap. For details on the usage and support backends have
a look at the batchtools tutorial page:
https://github.com/mllg/batchtools.
The general workflow with batchmark looks like this:
Create an ExperimentRegistry using makeExperimentRegistry.
Call batchmark(...) which defines jobs for all learners and tasks in an expand.grid fashion.
Submit jobs using submitJobs.
Babysit the computation, wait for all jobs to finish using waitForJobs.
Call reduceBatchmarkResult() to reduce results into a BenchmarkResult.
If you want to use this with OpenML datasets you can generate tasks from a vector
of dataset IDs easily with
tasks = lapply(data.ids, function(x) convertOMLDataSetToMlr(getOMLDataSet(x))).
1 2 |
learners |
[(list of) |
tasks |
[(list of) |
resamplings |
[(list of) |
measures |
[(list of) |
models |
[ |
reg |
[ |
[data.table]. Generated job ids are stored in the column “job.id”.
Other benchmark: BenchmarkResult,
benchmark,
convertBMRToRankMatrix,
friedmanPostHocTestBMR,
friedmanTestBMR,
generateCritDifferencesData,
getBMRAggrPerformances,
getBMRFeatSelResults,
getBMRFilteredFeatures,
getBMRLearnerIds,
getBMRLearnerShortNames,
getBMRLearners,
getBMRMeasureIds,
getBMRMeasures, getBMRModels,
getBMRPerformances,
getBMRPredictions,
getBMRTaskDescs,
getBMRTaskIds,
getBMRTuneResults,
plotBMRBoxplots,
plotBMRRanksAsBarChart,
plotBMRSummary,
plotCritDifferences,
reduceBatchmarkResults
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.