Calculate a coefficient of determination based on the likelihood-ratio test (\(R_{LR}^{2}\)).
r.squaredLR(object, null = NULL, null.RE = FALSE, ...)null.fit(object, evaluate = FALSE, RE.keep = FALSE, envir = NULL, ...)
r.squaredLR
returns a value of \(R_{LR}^{2}\), and the
attribute "adj.r.squared"
gives the Nagelkerke's modified statistic.
Note that this is not the same as nor equivalent to the classical
‘adjusted R squared’.
null.fit
returns the fitted null model object (if
evaluate = TRUE
) or an unevaluated call to fit a null model.
a fitted model object.
a fitted null model. If not provided, null.fit
will
be used to construct it. null.fit
's capabilities are limited to only
a few model classes, for others the null model has to be specified
manually.
logical, should the null model contain random factors? Only used if no null model is given, otherwise omitted, with a warning.
if TRUE
evaluate the fitted model object else return
the call.
if TRUE
, the random effects of the original model are
included.
the environment in which the null model is to be evaluated, defaults to the environment of the original model's formula.
further arguments, of which only x
would be used, to
maintain compatibility with older versions (x
has been replaced
with object
).
This statistic is is one of the several proposed pseudo-\(R^{2}\)'s for nonlinear regression models. It is based on an improvement from null (intercept only) model to the fitted model, and calculated as
$$ R_{LR}^{2}=1-\exp(-\frac{2}{n}(\log\mathcal{L}(x)-\log\mathcal{L}(0))) $$
where \(\log\mathcal{L}(x)\) and \(\log\mathcal{L}(0)\) are the log-likelihoods of the
fitted and the null model respectively.
ML estimates are used if models have been
fitted by REstricted ML (by calling logLik
with argument
REML = FALSE
). Note that the null model can include the random
factors of the original model, in which case the statistic represents the
‘variance explained’ by fixed effects.
For OLS models the value is consistent with classical \(R^{2}\). In some cases (e.g. in logistic regression), the maximum \(R_{LR}^{2}\) is less than one. The modification proposed by Nagelkerke (1991) adjusts the \(R_{LR}^{2}\) to achieve 1 at its maximum: \(\bar{R}^{2} = R_{LR}^{2} / \max(R_{LR}^{2}) \) where \(\max(R_{LR}^{2}) = 1 - \exp(\frac{2}{n}\log\mathcal{L}(\textrm{0})) \).
null.fit
tries to guess the null model call, given the provided
fitted model object. This would be usually a glm
. The function will give
an error for an unrecognised class.
Cox, D. R. and Snell, E. J. 1989 The analysis of binary data, 2nd ed. London, Chapman and Hall.
Magee, L. 1990 \(R^{2}\) measures based on Wald and likelihood ratio joint significance tests. Amer. Stat. 44, 250--253.
Nagelkerke, N. J. D. 1991 A note on a general definition of the coefficient of determination. Biometrika 78, 691--692.
summary.lm
, r.squaredGLMM
r2
from package performance calculates
many different types of \(R^{2}\).