Powell's dog leg method

From testwiki
Jump to navigation Jump to search

Powell's dog leg method, also called Powell's hybrid method, is an iterative optimisation algorithm for the solution of non-linear least squares problems, introduced in 1970 by Michael J. D. Powell.[1] Similarly to the Levenberg–Marquardt algorithm, it combines the Gauss–Newton algorithm with gradient descent, but it uses an explicit trust region. At each iteration, if the step from the Gauss–Newton algorithm is within the trust region, it is used to update the current solution. If not, the algorithm searches for the minimum of the objective function along the steepest descent direction, known as Cauchy point. If the Cauchy point is outside of the trust region, it is truncated to the boundary of the latter and it is taken as the new solution. If the Cauchy point is inside the trust region, the new solution is taken at the intersection between the trust region boundary and the line joining the Cauchy point and the Gauss-Newton step (dog leg step).[2]

The name of the method derives from the resemblance between the construction of the dog leg step and the shape of a dogleg hole in golf.[2]

Formulation

Construction of the dog leg step

Given a least squares problem in the form

F(𝒙)=12𝒇(𝒙)2=12i=1m(fi(𝒙))2

with fi:n, Powell's dog leg method finds the optimal point 𝒙=argmin𝒙F(𝒙) by constructing a sequence 𝒙k=𝒙k1+δk that converges to 𝒙. At a given iteration, the Gauss–Newton step is given by

𝜹𝒈𝒏=(𝑱𝑱)1𝑱𝒇(𝒙)

where 𝑱=(fixj) is the Jacobian matrix, while the steepest descent direction is given by

𝜹𝒔𝒅=𝑱𝒇(𝒙).

The objective function is linearised along the steepest descent direction

F(𝒙+t𝜹𝒔𝒅)12𝒇(𝒙)+t𝑱(𝒙)𝜹𝒔𝒅2=F(𝒙)+t𝜹𝒔𝒅𝑱𝒇(𝒙)+12t2𝑱𝜹𝒔𝒅2.

To compute the value of the parameter t at the Cauchy point, the derivative of the last expression with respect to t is imposed to be equal to zero, giving

t=𝜹𝒔𝒅𝑱𝒇(𝒙)𝑱𝜹𝒔𝒅2=𝜹𝒔𝒅2𝑱𝜹𝒔𝒅2.


Given a trust region of radius Δ, Powell's dog leg method selects the update step 𝜹𝒌 as equal to:

  • 𝜹𝒈𝒏, if the Gauss–Newton step is within the trust region (𝜹𝒈𝒏Δ);
  • Δ𝜹𝒔𝒅𝜹𝒔𝒅 if both the Gauss–Newton and the steepest descent steps are outside the trust region (t𝜹𝒔𝒅>Δ);
  • t𝜹𝒔𝒅+s(𝜹𝒈𝒏t𝜹𝒔𝒅) with s such that 𝜹=Δ, if the Gauss–Newton step is outside the trust region but the steepest descent step is inside (dog leg step).[1]

References

  1. 1.0 1.1 Powell (1970)
  2. 2.0 2.1 Yuan (2000)

Sources

Template:Optimization algorithms