This creates a BenchmarkResult from a batchtools::ExperimentRegistry. To setup the benchmark have a look at batchmark.
reduceBatchmarkResults(ids = NULL, keep.pred = TRUE,
show.info = getMlrOption("show.info"),
reg = batchtools::getDefaultRegistry())
(data.frame or integer) A base::data.frame (or data.table::data.table) with a column named “job.id”. Alternatively, you may also pass a vector of integerish job ids. If not set, defaults to all successfully terminated jobs (return value of batchtools::findDone.
(logical(1)
)
Keep the prediction data in the pred
slot of the result object.
If you do many experiments (on larger data sets) these objects might unnecessarily increase
object size / mem usage, if you do not really need them.
In this case you can set this argument to FALSE
.
Default is TRUE
.
(logical(1)
)
Print verbose output on console?
Default is set via configureMlr.
(batchtools::ExperimentRegistry) Registry, created by batchtools::makeExperimentRegistry. If not explicitly passed, uses the last created registry.
Other benchmark: BenchmarkResult
,
batchmark
, benchmark
,
convertBMRToRankMatrix
,
friedmanPostHocTestBMR
,
friedmanTestBMR
,
generateCritDifferencesData
,
getBMRAggrPerformances
,
getBMRFeatSelResults
,
getBMRFilteredFeatures
,
getBMRLearnerIds
,
getBMRLearnerShortNames
,
getBMRLearners
,
getBMRMeasureIds
,
getBMRMeasures
, getBMRModels
,
getBMRPerformances
,
getBMRPredictions
,
getBMRTaskDescs
,
getBMRTaskIds
,
getBMRTuneResults
,
plotBMRBoxplots
,
plotBMRRanksAsBarChart
,
plotBMRSummary
,
plotCritDifferences