Specification of the number of boosting iterations, step size and other parameters for boosting algorithms.
bst_control(mstop = 50, nu = 0.1, twinboost = FALSE, twintype=1, threshold=c("standard",
"adaptive"), f.init = NULL, coefir = NULL, xselect.init = NULL, center = FALSE,
trace = FALSE, numsample = 50, df = 4, s = NULL, sh = NULL, q = NULL, qh = NULL,
fk = NULL, start=FALSE, iter = 10, intercept = FALSE, trun=FALSE)
An object of class bst_control
, a list. Note fk
may be updated for robust boosting.
an integer giving the number of boosting iterations.
a small number (between 0 and 1) defining the step size or shrinkage parameter.
a logical value: TRUE
for twin boosting.
for twinboost=TRUE
only. For learner="ls"
, if twintype=1
, twin boosting with weights from magnitude of coefficients in the first round of boosting. If twintype=2
, weights are correlations between predicted values in the first round of boosting and current predicted values. For learners not componentwise least squares, twintype=2
.
if threshold="adaptive"
, the estimated function ctrl$fk
is updated in every boosting step. Otherwise, no update for ctrl$fk
in boosting steps. Only used in robust nonconvex loss function.
the estimate from the first round of twin boosting. Only useful when twinboost=TRUE
and learner="sm" or "tree"
.
the estimated coefficients from the first round of twin boosting. Only useful when twinboost=TRUE
and learner="ls"
.
the variable selected from the first round of twin boosting. Only useful when twinboost=TRUE
.
a logical value: TRUE
to center covariates with mean.
a logical value for printout of more details of information during the fitting process.
number of random sample variable selected in the first round of twin boosting. This is potentially useful in the future implementation.
degree of freedom used in smoothing splines.
nonconvex loss tuning parameter s
or frequency q
of outliers for robust regression and classification. If s
is missing but q
is available, s
may be computed as the 1-q
quantile of robust loss values using conventional software.
threshold value or frequency qh
of outliers for Huber regression family="huber"
or family="rhuberDC"
.
For family="huber"
, if sh
is not provided, sh
is then updated adaptively with the median of y-yhat
where yhat
is the estimated y
in the last boosting iteration. For family="rhuberDC"
, if sh
is missing but qh
is available, sh
may be computed as the 1-qh
quantile of robust loss values using conventional software.
predicted values at an iteration in the MM algorithm
a logical value, if start=TRUE
and fk
is a vector of values, then bst
iterations begin with fk
. Otherwise, bst
iterations begin with the default values. This can be useful, for instance, in rbst
for the MM boosting algorithm.
number of iteration in the MM algorithm
logical value, if TRUE, estimation of intercept with linear predictor model
logical value, if TRUE, predicted value in each boosting iteration is truncated at -1, 1, for family="closs"
in bst
and rfamily="closs"
in rbst
Objects to specify parameters of the boosting algorithms implemented in bst
, via the ctrl
argument.
The s
value is for robust nonconvex loss where smaller s
value is more robust to outliers with family="closs", "tbinom", "thinge", "tbinomd"
, and larger s
value more robust with family="clossR", "gloss", "qloss"
.
For family="closs"
, if s=2
, the loss is similar to the square loss; if s=1
, the loss function is an approximation of the hinge loss; for smaller values, the loss function approaches the 0-1 loss function if s<1
, the loss function is a nonconvex function of the margin.
The default value of s
is -1 if family="thinge"
, -log(3) if family="tbinom"
, and 4 if family="binomd"
. If trun=TRUE
, boosting classifiers can produce real values in [-1, 1] indicating their confidence in [-1, 1]-valued classification. cf. R. E. Schapire and Y. Singer. Improved boosting algorithms using confidence-rated predictions. In Proceedings of the Eleventh Annual Conference on Computational Learning Theory, pages 80-91, 1998.
bst