Univariate and multivariate optimization in Julia.
Optim.jl is part of the JuliaNLSolvers family.
|Documentation||PackageEvaluator||Build Status||Social||References to cite|
Optim.jl is a package for univariate and multivariate optimization of functions. A typical example of the usage of Optim.jl is
using Optim rosenbrock(x) = (1.0 - x)^2 + 100.0 * (x - x^2)^2 result = optimize(rosenbrock, zeros(2), BFGS())
This minimizes the Rosenbrock function
with a = 1, b = 100 and the initial values x=0, y=0. The minimum is at (a,a^2).
The above code gives the output
Results of Optimization Algorithm * Algorithm: BFGS * Starting Point: [0.0,0.0] * Minimizer: [0.9999999926033423,0.9999999852005353] * Minimum: 5.471433e-17 * Iterations: 16 * Convergence: true * |x - x'| ≤ 1.0e-32: false |x - x'| = 3.47e-07 * |f(x) - f(x')| ≤ 1.0e-32 |f(x)|: false |f(x) - f(x')| = 1.20e+03 |f(x)| * |g(x)| ≤ 1.0e-08: true |g(x)| = 2.33e-09 * Stopped by an increasing objective: false * Reached Maximum Number of Iterations: false * Objective Calls: 53 * Gradient Calls: 53
To get information on the keywords used to construct method instances, use the Julia REPL help prompt (
help?> LBFGS search: LBFGS LBFGS ≡≡≡≡≡≡≡ Constructor ============= LBFGS(; m::Integer = 10, alphaguess = LineSearches.InitialStatic(), linesearch = LineSearches.HagerZhang(), P=nothing, precondprep = (P, x) -> nothing, manifold = Flat(), scaleinvH0::Bool = true && (typeof(P) <: Void)) LBFGS has two special keywords; the memory length m, and the scaleinvH0 flag. The memory length determines how many previous Hessian approximations to store. When scaleinvH0 == true, then the initial guess in the two-loop recursion to approximate the inverse Hessian is the scaled identity, as can be found in Nocedal and Wright (2nd edition) (sec. 7.2). In addition, LBFGS supports preconditioning via the P and precondprep keywords. Description ============= The LBFGS method implements the limited-memory BFGS algorithm as described in Nocedal and Wright (sec. 7.2, 2006) and original paper by Liu & Nocedal (1989). It is a quasi-Newton method that updates an approximation to the Hessian using past approximations as well as the gradient. References ============ • Wright, S. J. and J. Nocedal (2006), Numerical optimization, 2nd edition. Springer • Liu, D. C. and Nocedal, J. (1989). "On the Limited Memory Method for Large Scale Optimization". Mathematical Programming B. 45 (3): 503–528
The package is registered in
METADATA.jl and can be installed with
5 days ago