Nonlinear Optimization Methods

Overview

There are several optimization techniques available, as shown in Table 3. You can choose a particular optimizer with the TECH=name option in the PROC statement or NLOPTIONS statement.

Table 3: Optimization Techniques

Algorithm TECH=
Trust region method TRUREG
Newton-Raphson method with line search NEWRAP
Newton-Raphson method with ridging NRRIDG
Quasi-Newton methods (DBFGS, DDFP, BFGS, DFP) QUANEW
Double-dogleg method (DBFGS, DDFP) DBLDOG
Conjugate gradient methods (PB, FR, PR, CD) CONGRA
Nelder-Mead simplex method NMSIMP


No algorithm for optimizing general nonlinear functions exists that always finds the global optimum for a general nonlinear minimization problem in a reasonable amount of time. Since no single optimization technique is invariably superior to others, NLO provides a variety of optimization techniques that work well in various circumstances. However, you can devise problems for which none of the techniques in NLO can find the correct solution. Moreover, nonlinear optimization can be computationally expensive in terms of time and memory, so you must be careful when matching an algorithm to a problem.

All optimization techniques in NLO use upper O left-parenthesis n squared right-parenthesis memory except the conjugate gradient methods, which use only upper O left-parenthesis n right-parenthesis of memory and are designed to optimize problems with many parameters. These iterative techniques require repeated computation of the following:

  • the function value (optimization criterion)

  • the gradient vector (first-order partial derivatives)

  • for some techniques, the (approximate) Hessian matrix (second-order partial derivatives)

However, since each of the optimizers requires different derivatives, some computational efficiencies can be gained. Table 4 shows, for each optimization technique, which derivatives are required. (FOD means that first-order derivatives or the gradient is computed; SOD means that second-order derivatives or the Hessian is computed.)

Table 4: Optimization Computations

Algorithm FOD SOD
TRUREG x x
NEWRAP x x
NRRIDG x x
QUANEW x -
DBLDOG x -
CONGRA x -
NMSIMP - -


Each optimization method employs one or more convergence criteria that determine when it has converged. The various termination criteria are listed and described in the previous section. An algorithm is considered to have converged when any one of the convergence criterion is satisfied. For example, under the default settings, the QUANEW algorithm will converge if normal upper A normal upper B normal upper S normal upper G normal upper C normal upper O normal upper N normal upper V less-than 1 normal upper E minus 5, normal upper F normal upper C normal upper O normal upper N normal upper V less-than 10 Superscript minus normal upper F normal upper D normal upper I normal upper G normal upper I normal upper T normal upper S, or normal upper G normal upper C normal upper O normal upper N normal upper V less-than 1 normal upper E minus 8.

Last updated: June 19, 2025