This function wraps MADlib's SVM for classification, regresssion and novelty detection.
madlib.svm (formula, data,
na.action = NULL, na.as.level = FALSE,
type = c("classification", "regression", "one-class"),
kernel = c("gaussian", "linear", "polynomial"),
degree = 3, gamma = NULL, coef0 = 1.0, class.weight = NULL,
tolerance = 1e-10, epsilon = NULL, cross = 0, lambda = 0.01,
control = list(), verbose = FALSE, ...)
an object of class "formula
" (or one that can be coerced to
that class): a symbolic description of the model to be
fitted. The details of model specification are given under
'Details'.
An object of db.obj
class. Currently, this parameter is
mandatory. If it is an object of class db.Rquery
or
db.view
, a temporary table will be created, and further
computation will be done on the temporary table. After the
computation, the temporary will be dropped from the corresponding
database.
A string which indicates what should happen when the data
contain NA
s. Possible values include na.omit
, "na.exclude"
, "na.fail"
and NULL
. Right now, na.omit
has been implemented. When the value is NULL
, nothing is done on the R side and NA
values are filtered on the MADlib side. User defined na.action
function is allowed.
A logical value, default is FALSE
. Whether to treat NA
value as a level in a categorical variable or just ignore it.
A string, default: "classification". Indicate type of analysis to perform: "classification", "regression" or "one-class".
A string, default: "gaussian". Type of kernel. Currently three kernel types are supported: 'linear', 'gaussian', and 'polynomial'.
Default: 3. The parameter needed for polynomial kernel
Default: 1/num_features. The parameter needed for gaussian kernel
Default: 1.0. The independent term in polynomial kernel
Default: 1.0. Set the weight for the positive and negative classes. If not given, all classes are set to have weight one. If class_weight = balanced, values of y are automatically adjusted as inversely proportional to class frequencies in the input data i.e. the weights are set as n_samples / (n_classes * bincount(y)).
Alternatively, class_weight can be a mapping, giving the weight for each class. Eg. For dependent variable values 'a' and 'b', the class_weight can be a: 2, b: 3. This would lead to each 'a' tuple's y value multiplied by 2 and each 'b' y value will be multiplied by 3.
For regression, the class weights are always one.
Default: 1e-10. The criterion to end iterations. The training stops whenever <the difference between the training models of two consecutive iterations is <smaller than tolerance or the iteration number is larger than max_iter.
Default: [0.01]. Determines the epsilon for epsilon-SVR. Ignored during classification. When training the model, differences of less than epsilon between estimated labels and actual labels are ignored. A larger epsilon will yield a model with fewer support vectors, but will not generalize as well to future data. Generally, it has been suggested that epsilon should increase with noisier data, and decrease with the number of samples. See [5].
Default: 0. Number of folds (k). Must be at least 2 to activate cross validation. If a value of k > 2 is specified, each fold is then used as a validation set once, while the other k - 1 folds form the training set.
Default: [0.01]. Regularization parameter. Must be non-negative.
A list, which contains the more control parameters for the optimizer.
- init.stepsize
: Default: [0.01]. Also known as the initial learning rate. A small value is usually desirable to ensure convergence, while a large value provides more room for progress during training. Since the best value depends on the condition number of the data, in practice one often searches in an exponential grid using built-in cross validation; e.g., "init_stepsize = [1, 0.1, 0.001]". To reduce training time, it is common to run cross validation on a subsampled dataset, since this usually provides a good estimate of the condition number of the whole dataset. Then the resulting init_stepsize can be run on the whole dataset.
- decay.factor
: Default: [0.9]. Control the learning rate schedule: 0 means constant rate; <-1 means inverse scaling, i.e., stepsize = init_stepsize / iteration; > 0 means <exponential decay, i.e., stepsize = init_stepsize * decay_factor^iteration.
- max.iter
: Default: [100]. The maximum number of iterations allowed.
- norm
: Default: 'L2'. Name of the regularization, either 'L2' or 'L1'.
- eps.table
: Default: NULL. Name of the input table that contains values of epsilon for different groups. Ignored when grouping_col is NULL. Define this input table if you want different epsilon values for different groups. The table consists of a column named epsilon which specifies the epsilon values, and one or more columns for grouping_col. Extra groups are ignored, and groups not present in this table will use the epsilon value specified in parameter epsilon.
- validation.result
: Default: NULL. Name of the table to store the cross validation results including the values of parameters and their averaged error values. For now, metric like 0-1 loss is used for classification and mean square error is used for regression. The table is only created if the name is not NULL.
A logical value, default: FALSE. Verbose output of the results of training.
More parameters can be passed into this function. Currently, it is just a place holder and any parameter here is not used.
If there is no grouping (i.e. no |
in the formula), the result
is a svm.madlib
object. Otherwise, it is a svm.madlib.grps
object, which is just a list of svm.madlib
objects.
A svm.madlib
object is a list which contains the following items:
A vector, the fitting coefficients.
An integer, the number of groups that the data is divided into according to the grouping columns in the formula.
An array of strings. The column names of the grouping columns.
A logical, whether the intercept is included in the fitting.
An array of strings, all the different terms used as independent variables in the fitting.
A string. The independent variables in an array format string.
A language object. The function call that generates this result.
An array of strings. The column names used in the fitting.
An array of strings, the same length as the number of independent
variables. The strings are used to print a clean result, especially when
we are dealing with the factor variables, where the dummy variable
names can be very long due to the inserting of a random string to
avoid naming conflicts, see as.factor,db.obj-method
for details. The list also contains dummy
and dummy.expr
, which are also used for processing the categorical variables, but do not contain any important information.
A '>db.data.frame
object, which wraps the model
table of this function.
A '>db.data.frame
object, which wraps the summary
table of this function.
A '>db.data.frame
object, which wraps the kernel
table of this function. Only created when non-linear kernel is used.
A terms
object, describing the terms in
the model formula.
The number of observations used to fit the model.
A db.obj
object, which wraps all the data used in the database.
If there are fittings for multiple groups, then this is only the wrapper for the data in one group.
The original db.obj
object. When there is no grouping, it is equal to data
above, otherwise it is the "sum" of data
from all groups.
Note that if there is grouping done, and there are multiple svm.madlib objects in the final result, each one of them contains the same copy model.
For details about how to write a formula, see formula
for details. "|" can be used at the end of the formula to denote that
the fitting is done conditioned on the values of one or more
variables. For example, y ~ x + sin(z) | v + w
will do the
fitting each distinct combination of the values of v
and
w
.
madlib.lm
,
madlib.summary
, madlib.arima
are MADlib
wrapper functions.
as.factor
creates categorical variables for fitting.
delete
safely deletes the result of this function.
# NOT RUN {
<!-- %% @test .port Database port number -->
<!-- %% @test .dbname Database name -->
## set up the database connection
## Assume that .port is port number and .dbname is the database name
cid <- db.connect(port = .port, dbname = .dbname, verbose = FALSE)
data <- as.db.data.frame(abalone, conn.id = cid, verbose = FALSE)
lk(data, 10)
## svm regression
## i.e. grouping on multiple columns
fit <- madlib.svm(length ~ height + shell | sex + (rings > 7), data = data, type = "regression")
fit
## use I(.) for expressions
fit <- madlib.svm(rings > 7 ~ height + shell + diameter + I(diameter^2),
data = data, type = "classification")
fit # display the result
## Adding new column for training
dat <- data
dat$arr <- db.array(data[,-c(1,2)])
array.data <- as.db.data.frame(dat)
fit <- madlib.svm(rings > 7 ~ arr, data = array.data)
db.disconnect(cid)
# }
Run the code above in your browser using DataLab