Descent direction explained
In optimization, a descent direction is a vector
that points towards a local minimum
of an objective function
.
Computing
by an iterative method, such as
line search defines a descent direction
at the
th iterate to be any
such that
\langlepk,\nablaf(xk)\rangle<0
, where
denotes the
inner product. The motivation for such an approach is that small steps along
guarantee that
is reduced, by
Taylor's theorem.
Using this definition, the negative of a non-zero gradient is always adescent direction, as
\langle-\nablaf(xk),\nablaf(xk)\rangle=-\langle\nablaf(xk),\nablaf(xk)\rangle<0
.
Numerous methods exist to compute descent directions, all with differing merits, such as gradient descent or the conjugate gradient method.
More generally, if
is a
positive definite matrix, then
is a descent direction at
.
[1] This generality is used in preconditioned gradient descent methods.
See also
Notes and References
- Book: J. M. Ortega and W. C. Rheinbold . Iterative Solution of Nonlinear Equations in Several Variables . 243 . 1970 . 10.1137/1.9780898719468 .