
absdPsiMC(t, family, theta, degree = 1, n.MC, method = c("log", "direct", "pois.direct", "pois"), log = FALSE, is.log.t = FALSE)
numeric
vector of evaluation points.
t
contains the logarithm of the
“mathematical” $t$, i.e., conceptually,
psi(t, *) == psi(log(t), *, is.log.t=TRUE)
, where the latter
may potentially be numerically accurate, e.g., for $t =
10^{500}$, where as the former would just return $psi(Inf, *) = 0$.numeric
vector of the same length as t
containing
the absolute values of the generator derivatives.
degree
and
$N=$n.MC
. Note that this is comparably fast even if
t
contains many evaluation points, since the random variates
$V_k ~ F, k in {1,...,N}$ only have
to be generated once, not depending on t
.
acopula-families
.
t <- c(0:100,Inf)
set.seed(1)
(ps <- absdPsiMC(t, family="Gumbel", theta=2, degree=10, n.MC=10000, log=TRUE))
## Note: The absolute value of the derivative at 0 should be Inf for
## Gumbel, however, it is always finite for the Monte Carlo approximation
set.seed(1)
ps2 <- absdPsiMC(log(t), family="Gumbel", theta=2, degree=10,
n.MC=10000, log=TRUE, is.log.t = TRUE)
stopifnot(all.equal(ps[-1], ps2[-1], tolerance=1e-14))
## Now is there an advantage of using "is.log.t" ?
sapply(eval(formals(absdPsiMC)$method), function(MM)
absdPsiMC(780, family="Gumbel", method = MM,
theta=2, degree=10, n.MC=10000, log=TRUE, is.log.t = TRUE))
## not really better, yet...
Run the code above in your browser using DataLab