This is the abstract base class for learner objects like LearnerClassif and LearnerRegr.
Learners are build around the three following key parts:
Methods $train()
and $predict()
which call internal methods or private methods $.train()
/$.predict()
).
A paradox::ParamSet which stores meta-information about available hyperparameters, and also stores hyperparameter settings.
Meta-information about the requirements and capabilities of the learner.
The fitted model stored in field $model
, available after calling $train()
.
Predefined learners are stored in the dictionary mlr_learners,
e.g. classif.rpart
or regr.rpart
.
More classification and regression learners are implemented in the add-on package mlr3learners. Learners for survival analysis (or more general, for probabilistic regression) can be found in mlr3proba. Unsupervised cluster algorithms are implemented in mlr3cluster. The dictionary mlr_learners gets automatically populated with the new learners as soon as the respective packages are loaded.
More (experimental) learners can be found in the GitHub repository: https://github.com/mlr-org/mlr3extralearners. A guide on how to extend mlr3 with custom learners can be found in the mlr3book.
To combine the learner with preprocessing operations like factor encoding, mlr3pipelines is recommended.
Hyperparameters stored in the param_set
can be tuned with mlr3tuning.
Specific learner implementations are free to implement additional getters to ease the access of certain parts of the model in the inherited subclasses.
For the following operations, extractors are standardized:
importance(...)
: Returns the feature importance score as numeric vector.
The higher the score, the more important the variable.
The returned vector is named with feature names and sorted in decreasing order.
Note that the model might omit features it has not used at all.
The learner must be tagged with property "importance"
.
To filter variables using the importance scores, see package mlr3filters.
selected_features(...)
: Returns a subset of selected features as character()
.
The learner must be tagged with property "selected_features"
.
oob_error(...)
: Returns the out-of-bag error of the model as numeric(1)
.
The learner must be tagged with property "oob_error"
.
loglik(...)
: Extracts the log-likelihood (c.f. stats::logLik()
).
This can be used in measures like mlr_measures_aic or mlr_measures_bic.
internal_valid_scores
: Returns the internal validation score(s) of the model as a named list()
.
Only available for Learner
s with the "validation"
property.
If the learner is not trained yet, this returns NULL
.
internal_tuned_values
: Returns the internally tuned hyperparameters of the model as a named list()
.
Only available for Learner
s with the "internal_tuning"
property.
If the learner is not trained yet, this returns NULL
.
All information about hyperparameters is stored in the slot param_set
which is a paradox::ParamSet.
The printer gives an overview about the ids of available hyperparameters, their storage type, lower and upper bounds,
possible levels (for factors), default values and assigned values.
To set hyperparameters, assign a named list to the subslot values
:
lrn = lrn("classif.rpart")
lrn$param_set$values = list(minsplit = 3, cp = 0.01)
Note that this operation replaces all previously set hyperparameter values.
If you only intend to change one specific hyperparameter value and leave the others as-is, you can use the helper function mlr3misc::insert_named()
:
lrn$param_set$values = mlr3misc::insert_named(lrn$param_set$values, list(cp = 0.001))
If the learner has additional hyperparameters which are not encoded in the ParamSet, you can easily extend the learner.
Here, we add a factor hyperparameter with id "foo"
and possible levels "a"
and "b"
:
lrn$param_set$add(paradox::ParamFct$new("foo", levels = c("a", "b")))
Some Learners, such as XGBoost
, other boosting algorithms, or deep learning models (mlr3torch
),
utilize validation data during the training to prevent overfitting or to log the validation performance.
It is possible to configure learners to be able to receive such an independent validation set during training.
To do so, one must:
annotate the learner with the "validation"
property
implement the active binding $internal_valid_scores
(see section Optional Extractors), as well as the
private method $.extract_internal_valid_scores()
which returns the (final) internal validation scores from the
model of the Learner
and returns them as a named list()
of numeric(1)
.
If the model is not trained yet, this method should return NULL
.
Add the validate
parameter, which can be either NULL
, a ratio in $(0, 1)$, "test"
, or "predefined"
:
NULL
: no validation
ratio
: only proportion 1 - ratio
of the task is used for training and ratio
is used for validation.
"test"
means that the "test"
task is used.
Warning: This can lead to biased performance estimation.
This option is only available if the learner is being trained via resample()
, benchmark()
or functions that
internally use them, e.g. tune()
of mlr3tuning or batchmark()
of mlr3batchmark.
This is especially useful for hyperparameter tuning, where one might e.g. want to use the same validation data
for early stopping and model evaluation.
"predefined"
means that the task's (manually set) $internal_valid_task
is used.
See the Task
documentation for more information.
For an example how to do this, see LearnerClassifDebug
.
Note that in .train()
, the $internal_valid_task
will only be present if the $validate
field of the Learner
is set to a non-NULL
value.
Some learners such as XGBoost
or cv.glmnet
can internally tune hyperparameters.
XGBoost, for example, can tune the number of boosting rounds based on the validation performance.
CV Glmnet, on the other hand, can tune the regularization parameter based on an internal cross-validation.
Internal tuning can therefore rely on the internal validation data, but does not necessarily do so.
In order to be able to combine this internal hyperparamer tuning with the standard hyperparameter optimization implemented via mlr3tuning, one most:
annotate the learner with the "internal_tuning"
property
implement the active binding $internal_tuned_values
(see section Optional Extractors) as well as the
private method $.extract_internal_tuned_values()
which extracts the internally tuned values from the Learner
's
model and returns them as a named list()
.
If the model is not trained yet, this method should return NULL
.
Have at least one parameter tagged with "internal_tuning"
, which requires to also provide a in_tune_fn
and
disable_tune_fn
, and should also include a default aggr
egation function.
For an example how to do this, see LearnerClassifDebug
.
Some Learner
s have models that cannot be serialized as they e.g. contain external pointers.
In order to still be able to save them, use them with parallelization or callr encapsulation it is necessary
to implement how they should be (un)-marshaled. See marshaling
for how to do this.
id
(character(1)
)
Identifier of the object.
Used in tables, plot and text output.
label
(character(1)
)
Label for this object.
Can be used in tables, plot and text output instead of the ID.
state
(NULL
| named list()
)
Current (internal) state of the learner.
Contains all information gathered during train()
and predict()
.
It is not recommended to access elements from state
directly.
This is an internal data structure which may change in the future.
task_type
(character(1)
)
Task type, e.g. "classif"
or "regr"
.
For a complete list of possible task types (depending on the loaded packages),
see mlr_reflections$task_types$type
.
predict_types
(character()
)
Stores the possible predict types the learner is capable of.
A complete list of candidate predict types, grouped by task type, is stored in mlr_reflections$learner_predict_types
.
feature_types
(character()
)
Stores the feature types the learner can handle, e.g. "logical"
, "numeric"
, or "factor"
.
A complete list of candidate feature types, grouped by task type, is stored in mlr_reflections$task_feature_types
.
properties
(character()
)
Stores a set of properties/capabilities the learner has.
A complete list of candidate properties, grouped by task type, is stored in mlr_reflections$learner_properties
.
data_formats
(character()
)
Supported data format, e.g. "data.table"
or "Matrix"
.
packages
(character(1)
)
Set of required packages.
These packages are loaded, but not attached.
predict_sets
(character()
)
During resample()
/benchmark()
, a Learner can predict on multiple sets.
Per default, a learner only predicts observations in the test set (predict_sets == "test"
).
To change this behavior, set predict_sets
to a non-empty subset of {"train", "test", "internal_valid"}
.
The "train"
predict set contains the train ids from the resampling. This means that if a learner does validation and
sets $validate
to a ratio (creating the validation data from the training data), the train predictions
will include the predictions for the validation data.
Each set yields a separate Prediction object.
Those can be combined via getters in ResampleResult/BenchmarkResult, or Measures can be configured
to operate on specific subsets of the calculated prediction sets.
parallel_predict
(logical(1)
)
If set to TRUE
, use future to calculate predictions in parallel (default: FALSE
).
The row ids of the task
will be split into future::nbrOfWorkers()
chunks,
and predictions are evaluated according to the active future::plan()
.
This currently only works for methods Learner$predict()
and Learner$predict_newdata()
,
and has no effect during resample()
or benchmark()
where you have other means
to parallelize.
timeout
(named numeric(2)
)
Timeout for the learner's train and predict steps, in seconds.
This works differently for different encapsulation methods, see
mlr3misc::encapsulate()
.
Default is c(train = Inf, predict = Inf)
.
Also see the section on error handling the mlr3book:
https://mlr3book.mlr-org.com/chapters/chapter10/advanced_technical_aspects_of_mlr3.html#sec-error-handling
man
(character(1)
)
String in the format [pkg]::[topic]
pointing to a manual page for this object.
Defaults to NA
, but can be set by child classes.
model
(any)
The fitted model. Only available after $train()
has been called.
timings
(named numeric(2)
)
Elapsed time in seconds for the steps "train"
and "predict"
.
Measured via mlr3misc::encapsulate()
.
log
(data.table::data.table()
)
Returns the output (including warning and errors) as table with columns
"stage"
("train" or "predict"),
"class"
("output", "warning", or "error"), and
"msg"
(character()
).
warnings
(character()
)
Logged warnings as vector.
errors
(character()
)
Logged errors as vector.
hash
(character(1)
)
Hash (unique identifier) for this object.
phash
(character(1)
)
Hash (unique identifier) for this partial object, excluding some components
which are varied systematically during tuning (parameter values).
predict_type
(character(1)
)
Stores the currently active predict type, e.g. "response"
.
Must be an element of $predict_types
.
param_set
(paradox::ParamSet)
Set of hyperparameters.
encapsulate
(named character()
)
Controls how to execute the code in internal train and predict methods.
Must be a named character vector with names "train"
and "predict"
.
Possible values are "none"
, "try"
, "evaluate"
(requires package evaluate) and "callr"
(requires package callr).
See mlr3misc::encapsulate()
for more details.
fallback
(Learner)
Learner which is fitted to impute predictions in case that either the model fitting or the prediction of the top learner is not successful.
Requires encapsulation, otherwise errors are not caught and the execution is terminated before the fallback learner kicks in.
If you have not set encapsulation manually before, setting the fallback learner automatically
activates encapsulation using the evaluate package.
Also see the section on error handling the mlr3book:
https://mlr3book.mlr-org.com/chapters/chapter10/advanced_technical_aspects_of_mlr3.html#sec-error-handling
hotstart_stack
(HotstartStack)
.
Stores HotstartStack
.
new()
Creates a new instance of this R6 class.
Note that this object is typically constructed via a derived classes, e.g. LearnerClassif or LearnerRegr.
Learner$new(
id,
task_type,
param_set = ps(),
predict_types = character(),
feature_types = character(),
properties = character(),
data_formats = "data.table",
packages = character(),
label = NA_character_,
man = NA_character_
)
id
(character(1)
)
Identifier for the new instance.
task_type
(character(1)
)
Type of task, e.g. "regr"
or "classif"
.
Must be an element of mlr_reflections$task_types$type.
param_set
(paradox::ParamSet)
Set of hyperparameters.
predict_types
(character()
)
Supported predict types. Must be a subset of mlr_reflections$learner_predict_types
.
feature_types
(character()
)
Feature types the learner operates on. Must be a subset of mlr_reflections$task_feature_types
.
properties
(character()
)
Set of properties of the Learner.
Must be a subset of mlr_reflections$learner_properties
.
The following properties are currently standardized and understood by learners in mlr3:
"missings"
: The learner can handle missing values in the data.
"weights"
: The learner supports observation weights.
"importance"
: The learner supports extraction of importance scores, i.e. comes with an $importance()
extractor function (see section on optional extractors in Learner).
"selected_features"
: The learner supports extraction of the set of selected features, i.e. comes with a $selected_features()
extractor function (see section on optional extractors in Learner).
"oob_error"
: The learner supports extraction of estimated out of bag error, i.e. comes with a oob_error()
extractor function (see section on optional extractors in Learner).
"validation"
: The learner can use a validation task during training.
"internal_tuning"
: The learner is able to internally optimize hyperparameters (those are also tagged with "internal_tuning"
).
"marshal"
: To save learners with this property, you need to call $marshal()
first.
If a learner is in a marshaled state, you call first need to call $unmarshal()
to use its model, e.g. for prediction.
data_formats
(character()
)
Set of supported data formats which can be processed during $train()
and $predict()
,
e.g. "data.table"
.
packages
(character()
)
Set of required packages.
A warning is signaled by the constructor if at least one of the packages is not installed,
but loaded (not attached) later on-demand via requireNamespace()
.
label
(character(1)
)
Label for the new instance.
man
(character(1)
)
String in the format [pkg]::[topic]
pointing to a manual page for this object.
The referenced help package can be opened via method $help()
.
...
(ignored).
...
(ignored).
train()
Train the learner on a set of observations of the provided task
.
Mutates the learner by reference, i.e. stores the model alongside other information in field $state
.
Learner$train(task, row_ids = NULL)
task
(Task).
row_ids
(integer()
)
Vector of training indices as subset of task$row_ids
.
For a simple split into training and test set, see partition()
.
Returns the object itself, but modified by reference.
You need to explicitly $clone()
the object beforehand if you want to keeps
the object in its previous state.
predict()
Uses the information stored during $train()
in $state
to create a new Prediction
for a set of observations of the provided task
.
Learner$predict(task, row_ids = NULL)
task
(Task).
row_ids
(integer()
)
Vector of test indices as subset of task$row_ids
.
For a simple split into training and test set, see partition()
.
Prediction.
predict_newdata()
Uses the model fitted during $train()
to create a new Prediction based on the new data in newdata
.
Object task
is the task used during $train()
and required for conversion of newdata
.
If the learner's $train()
method has been called, there is a (size reduced) version
of the training task stored in the learner.
If the learner has been fitted via resample()
or benchmark()
, you need to pass the corresponding task stored
in the ResampleResult or BenchmarkResult, respectively.
Learner$predict_newdata(newdata, task = NULL)
newdata
(any object supported by as_data_backend()
)
New data to predict on.
All data formats convertible by as_data_backend()
are supported, e.g.
data.frame()
or DataBackend.
If a DataBackend is provided as newdata
, the row ids are preserved,
otherwise they are set to to the sequence 1:nrow(newdata)
.
task
(Task).
Prediction.
reset()
Reset the learner, i.e. un-train by resetting the state
.
Learner$reset()
Returns the object itself, but modified by reference.
You need to explicitly $clone()
the object beforehand if you want to keeps
the object in its previous state.
base_learner()
Extracts the base learner from nested learner objects like
GraphLearner
in mlr3pipelines or AutoTuner
in
mlr3tuning.
Returns the Learner itself for regular learners.
Learner$base_learner(recursive = Inf)
recursive
(integer(1)
)
Depth of recursion for multiple nested objects.
Learner.
clone()
The objects of this class are cloneable with this method.
Learner$clone(deep = FALSE)
deep
Whether to make a deep clone.
Chapter in the mlr3book: https://mlr3book.mlr-org.com/chapters/chapter2/data_and_basic_modeling.html#sec-learners
Package mlr3learners for a solid collection of essential learners.
Package mlr3extralearners for more learners.
Dictionary of Learners: mlr_learners
as.data.table(mlr_learners)
for a table of available Learners in the running session (depending on the loaded packages).
mlr3pipelines to combine learners with pre- and postprocessing steps.
Package mlr3viz for some generic visualizations.
Extension packages for additional task types:
mlr3proba for probabilistic supervised regression and survival analysis.
mlr3cluster for unsupervised clustering.
mlr3tuning for tuning of hyperparameters, mlr3tuningspaces for established default tuning spaces.
Other Learner:
LearnerClassif
,
LearnerRegr
,
mlr_learners
,
mlr_learners_classif.debug
,
mlr_learners_classif.featureless
,
mlr_learners_classif.rpart
,
mlr_learners_regr.debug
,
mlr_learners_regr.featureless
,
mlr_learners_regr.rpart