Learn R Programming

alabama (version 2022.4-1)

auglag: Nonlinear optimization with constraints

Description

Augmented Lagrangian Minimization Algorithm for optimizing smooth nonlinear objective functions with constraints. Linear or nonlinear equality and inequality constraints are allowed.

Usage

auglag(par, fn, gr, hin, hin.jac, heq, heq.jac, 
control.outer=list(), control.optim = list(), ...)

Value

A list with the following components:

par

Parameters that optimize the nonlinear objective function, satisfying constraints, if convergence is successful.

value

The value of the objective function at termination.

counts

A vector of length 2 denoting the number of times the objective fn and the gr were evaluated, respectively.

convergence

An integer code indicating type of convergence. 0 indicates successful convergence. Positive integer codes indicate failure to converge.

outer.iterations

Number of outer iterations

lambda

Values of the Lagrangian parameter. This is a vector of same length as the total number of inequalities and equalities. It must be zero for inactive inequalities; non-negative for active inequalities; and can have any sign for equalities.

sigma

Value of augmented penalty parameter for the quadratic term

gradient

Gradient of the augmented Lagrangian function at convergence. It should be small.

hessian

Hessian of the augmented Lagrangian function at convergence. It should be positive (negative) definite for minimization (maximization)

ineq

Values of inequlaity constraints at convergence. All of them must be non-negative

equal

Values of equlaity constraints at convergence. All of them must be close to zero.

kkt1

A logical variable indicating whether or not the first-order KKT conditions were satisfied.

kkt2

A logical variable indicating whether or not the second-order KKT conditions were satisfied.

Arguments

par

Starting vector of parameter values. Any initial vector, even those violating inequality constraints, may be specified. This is in contrast to constrOptim.nl which requires "feasible" initial values with respect to inequality constraints

fn

Nonlinear objective function that is to be optimized. A scalar function that takes a real vector as argument and returns a scalar that is the value of the function at that point (see details).

gr

The gradient of the objective function fn evaluated at the argument. This is a vector-function that takes a real vector as argument and returns a real vector of the same length. It defaults to "NULL", which means that gradient is evaluated numerically. Computations are dramatically faster in high-dimensional problems when the exact gradient is provided. See *Example*.

hin

a vector function specifying inequality constraints such that hin[j] > 0 for all j

hin.jac

Jacobian of hin. If unspecified, it will be computed using finite-difference, but computations will be faster if specified.

heq

a vector function specifying equality constraints such that heq[j] = 0 for all j

heq.jac

Jacobian of heq. If unspecified, it will be computed using finite-difference, but computations will be faster if specified.

control.outer

A list of control parameters to be used by the outer loop in constrOptim.nl. See *Details* for more information.

control.optim

A list of control parameters to be used by the unconstrained optimization algorithm in the inner loop. Identical to that used in optim or in nlminb.

...

Additional arguments passed to fn, gr, hin, heq. All of them must accept any specified arguments, either explicitly or by having a ... argument, but they do not need to use them all.

Author

Ravi Varadhan, Center on Aging and Health, Johns Hopkins University.

Details

Argument control.outer is a list specifing any changes to default values of algorithm control parameters for the outer loop. Note that the names of these must be specified completely. Partial matching will not work. The list items are as follows:

lam0: Initial value for the Lagrangian parameter.

sig0: A scaling parameter for penalty term that is augmented to the Lagrangian.

eps: Tolerance for convergence of outer iterations of the barrier and/or augmented lagrangian algorithm

itmax: Maximum number of outer iterations.

ilack.max: Maximum number of outer iterations where no change in parameters is tolerated.

trace: A logical variable indicating whether information on outer iterations should be printed out. If TRUE, at each outer iteration information is displayed on: (i) how well the inequality and equalities are satisfied, (ii) current parameter values, and (iii) current objective function value.

method: Unconstrained optimization algorithm for inner loop optimization. User can specify any algorithm in optim(). The default is the "BFGS" variable metric method. However, the user can also invoke the nlminb() algorithm by specifying method="nlminb", which can often perform better than "BFGS."

NMinit: A logical variable indicating whether "Nelder-Mead" algorithm should be used in optim() for the first outer iteration.

i.scale: A vector of length equal to number of inequalities that may be used to scale the inequalities or it can be a scalar in which case all the inequalities are scaled by the same value.

e.scale: A vector of length equal to number of equalities that may be used to scale the equalities or it can be a scalar in which case all the equalities are scaled by the same value.

kkt2.check: A logical variable (TRUE/FALSE) indicating whether the second-order KKT condition should be checked. Deafult is TRUE. It may be set to FALSE in problems where the Hessian computation can b etime consuming.

References

Lange K, Optimization, 2004, Springer.

Madsen K, Nielsen HB, Tingleff O, Optimization With Constraints, 2004, IMM, Technical University of Denmark.

See Also

See Also constrOptim.nl, nlminb, optim.

Examples

Run this code

fn <- function(x) (x[1] + 3*x[2] + x[3])^2 + 4 * (x[1] - x[2])^2

gr <- function(x) {
g <- rep(NA, 3)
g[1] <- 2*(x[1] + 3*x[2] + x[3]) + 8*(x[1] - x[2]) 
g[2] <- 6*(x[1] + 3*x[2] + x[3]) - 8*(x[1] - x[2]) 
g[3] <- 2*(x[1] + 3*x[2] + x[3])
g
}

heq <- function(x) {
h <- rep(NA, 1)
h[1] <- x[1] + x[2] + x[3] - 1
h
}


heq.jac <- function(x) {
j <- matrix(NA, 1, length(x))
j[1, ] <- c(1, 1, 1)
j
}

hin <- function(x) {
h <- rep(NA, 1)
h[1] <- 6*x[2] + 4*x[3] - x[1]^3 - 3
h[2] <- x[1]
h[3] <- x[2]
h[4] <- x[3]
h
}


hin.jac <- function(x) {
j <- matrix(NA, 4, length(x))
j[1, ] <- c(-3*x[1]^2, 6, 4)
j[2, ] <- c(1, 0, 0)
j[3, ] <- c(0, 1, 0)
j[4, ] <- c(0, 0, 1)
j
}

# Note: `auglag' accepts infeasible starting values
#
p0 <- runif(3)
ans <- auglag(par=p0, fn=fn, gr=gr, heq=heq, heq.jac=heq.jac, hin=hin, hin.jac=hin.jac) 
ans

# Not specifying the gradient and the Jacobians
set.seed(12)
p0 <- runif(3)
ans2 <- auglag(par=p0, fn=fn, heq=heq, hin=hin) 
ans2

# Using "nlminb" algorithm
ans3 <- auglag(par=p0, fn=fn, heq=heq, hin=hin, control.outer=list(method="nlminb")) 
ans3

# Turning off the second-order KKT condition check
ans4 <- auglag(par=p0, fn=fn, heq=heq, hin=hin, control.outer=list(kkt2.check=FALSE)) 
ans4

Run the code above in your browser using DataLab