Descent direction

From testwiki
Jump to navigation Jump to search

In optimization, a descent direction is a vector π©βˆˆβ„n that points towards a local minimum π±βˆ— of an objective function f:ℝn→ℝ.

Computing π±βˆ— by an iterative method, such as line search defines a descent direction 𝐩kβˆˆβ„n at the kth iterate to be any 𝐩k such that ⟨𝐩k,βˆ‡f(𝐱k)⟩<0, where ⟨,⟩ denotes the inner product. The motivation for such an approach is that small steps along 𝐩k guarantee that f is reduced, by Taylor's theorem.

Using this definition, the negative of a non-zero gradient is always a descent direction, as βŸ¨βˆ’βˆ‡f(𝐱k),βˆ‡f(𝐱k)⟩=βˆ’βŸ¨βˆ‡f(𝐱k),βˆ‡f(𝐱k)⟩<0.

Numerous methods exist to compute descent directions, all with differing merits, such as gradient descent or the conjugate gradient method.

More generally, if P is a positive definite matrix, then pk=βˆ’Pβˆ‡f(xk) is a descent direction at xk.[1] This generality is used in preconditioned gradient descent methods.

See also

References

Template:Reflist