Descent direction explained

In optimization, a descent direction is a vector

p\inRn

that points towards a local minimum

x*

of an objective function

f:Rn\toR

.

Computing

x*

by an iterative method, such as line search defines a descent direction
n
p
k\inR
at the

k

th iterate to be any

pk

such that

\langlepk,\nablaf(xk)\rangle<0

, where

\langle,\rangle

denotes the inner product. The motivation for such an approach is that small steps along

pk

guarantee that

\displaystylef

is reduced, by Taylor's theorem.

Using this definition, the negative of a non-zero gradient is always adescent direction, as

\langle-\nablaf(xk),\nablaf(xk)\rangle=-\langle\nablaf(xk),\nablaf(xk)\rangle<0

.

Numerous methods exist to compute descent directions, all with differing merits, such as gradient descent or the conjugate gradient method.

More generally, if

P

is a positive definite matrix, then

pk=-P\nablaf(xk)

is a descent direction at

xk

.[1] This generality is used in preconditioned gradient descent methods.

See also

Notes and References

  1. Book: J. M. Ortega and W. C. Rheinbold . Iterative Solution of Nonlinear Equations in Several Variables . 243 . 1970 . 10.1137/1.9780898719468 .