language-icon Old Web
English
Sign In

Nonlinear conjugate gradient method

In numerical optimization, the nonlinear conjugate gradient method generalizes the conjugate gradient method to nonlinear optimization. For a quadratic function f ( x ) {displaystyle displaystyle f(x)} In numerical optimization, the nonlinear conjugate gradient method generalizes the conjugate gradient method to nonlinear optimization. For a quadratic function f ( x ) {displaystyle displaystyle f(x)} the minimum of f {displaystyle f} is obtained when the gradient is 0: Whereas linear conjugate gradient seeks a solution to the linear equation A T A x = A T b {displaystyle displaystyle A^{T}Ax=A^{T}b} , the nonlinear conjugate gradient method is generally used to find the local minimum of a nonlinear function using its gradient ∇ x f {displaystyle abla _{x}f} alone. It works when the function is approximately quadratic near the minimum, which is the case when the function is twice differentiable at the minimum and the second derivative is non-singular there. Given a function f ( x ) {displaystyle displaystyle f(x)} of N {displaystyle N} variables to minimize, its gradient ∇ x f {displaystyle abla _{x}f} indicates the direction of maximum increase.One simply starts in the opposite (steepest descent) direction: with an adjustable step length α {displaystyle displaystyle alpha } and performs a line search in this direction until it reaches the minimum of f {displaystyle displaystyle f} : After this first iteration in the steepest direction Δ x 0 {displaystyle displaystyle Delta x_{0}} , the following steps constitute one iteration of moving along a subsequent conjugate direction s n {displaystyle displaystyle s_{n}} , where s 0 = Δ x 0 {displaystyle displaystyle s_{0}=Delta x_{0}} : With a pure quadratic function the minimum is reached within N iterations (excepting roundoff error), but a non-quadratic function will make slower progress. Subsequent search directions lose conjugacy requiring the search direction to be reset to the steepest descent direction at least every N iterations, or sooner if progress stops. However, resetting every iteration turns the method into steepest descent. The algorithm stops when it finds the minimum, determined when no progress is made after a direction reset (i.e. in the steepest descent direction), or when some tolerance criterion is reached. Within a linear approximation, the parameters α {displaystyle displaystyle alpha } and β {displaystyle displaystyle eta } are the same as in thelinear conjugate gradient method but have been obtained with line searches.The conjugate gradient method can follow narrow (ill-conditioned) valleys, where the steepest descent method slows down and follows a criss-cross pattern. Four of the best known formulas for β n {displaystyle displaystyle eta _{n}} are named after their developers:

[ "Conjugate gradient method", "Gradient method", "Gradient descent", "Biconjugate gradient stabilized method", "Neighbourhood components analysis", "Descent direction", "Biconjugate gradient method", "wolfe line search" ]
Parent Topic
Child Topic
    No Parent Topic