Construct your own performance measure, used after resampling. Note that individual training / test set performance values will be set to `NA`, you only calculate an aggregated value. If you can define a function that makes sense for every single training / test set, implement your own [Measure].
makeCustomResampledMeasure(measure.id, aggregation.id, minimize = TRUE,
properties = character(0L), fun, extra.args = list(), best = NULL,
worst = NULL, measure.name = measure.id,
aggregation.name = aggregation.id, note = "")
(`character(1)`) Short name of measure.
(`character(1)`) Short name of aggregation.
(`logical(1)`) Should the measure be minimized? Default is `TRUE`.
([character]) Set of measure properties. For a list of values see [Measure]. Default is `character(0)`.
(`function(task, group, pred, extra.args)`) Calculates performance value from [ResamplePrediction] object. For rare cases you can also use the task, the grouping or the extra arguments `extra.args`.
The task.
Grouping of resampling iterations. This encodes whether specific iterations 'belong together' (e.g. repeated CV).
Prediction object.
See below.
([list]) List of extra arguments which will always be passed to `fun`. Default is empty list.
(`numeric(1)`) Best obtainable value for measure. Default is -`Inf` or `Inf`, depending on `minimize`.
(`numeric(1)`) Worst obtainable value for measure. Default is `Inf` or -`Inf`, depending on `minimize`.
(`character(1)`) Long name of measure. Default is `measure.id`.
(`character(1)`) Long name of the aggregation. Default is `aggregation.id`.
([character]) Description and additional notes for the measure. Default is “”.
Other performance: ConfusionMatrix
,
calculateConfusionMatrix
,
calculateROCMeasures
,
estimateRelativeOverfitting
,
makeCostMeasure
, makeMeasure
,
measures
, performance
,
setAggregation
,
setMeasurePars