Frank–Wolfe algorithm

From testwiki
Revision as of 20:37, 11 July 2024 by 129.97.124.26 (talk) (The direction-finding subproblem and the update rule did not comply with each other. Either x_k +s in D in the subproblem and x_k+1 <-- x_k + \alpha s in the update or s in D in the subproblem and x_k+1 <-- x_k + \alpha (s - x_k) in the update are used. See Jaggi (2013))
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Template:Short description The Frank–Wolfe algorithm is an iterative first-order optimization algorithm for constrained convex optimization. Also known as the conditional gradient method,[1] reduced gradient algorithm and the convex combination algorithm, the method was originally proposed by Marguerite Frank and Philip Wolfe in 1956.[2] In each iteration, the Frank–Wolfe algorithm considers a linear approximation of the objective function, and moves towards a minimizer of this linear function (taken over the same domain).

Problem statement

Suppose 𝒟 is a compact convex set in a vector space and f:𝒟 is a convex, differentiable real-valued function. The Frank–Wolfe algorithm solves the optimization problem

Minimize f(𝐱)
subject to 𝐱𝒟.

Algorithm

A step of the Frank–Wolfe algorithm
Initialization: Let k0, and let 𝐱0 be any point in 𝒟.
Step 1. Direction-finding subproblem: Find 𝐬k solving
Minimize 𝐬Tf(𝐱k)
Subject to 𝐬𝒟
(Interpretation: Minimize the linear approximation of the problem given by the first-order Taylor approximation of f around 𝐱k constrained to stay within 𝒟.)
Step 2. Step size determination: Set α2k+2, or alternatively find α that minimizes f(𝐱k+α(𝐬k𝐱k)) subject to 0α1 .
Step 3. Update: Let 𝐱k+1𝐱k+α(𝐬k𝐱k), let kk+1 and go to Step 1.

Properties

While competing methods such as gradient descent for constrained optimization require a projection step back to the feasible set in each iteration, the Frank–Wolfe algorithm only needs the solution of a convex problem over the same set in each iteration, and automatically stays in the feasible set.

The convergence of the Frank–Wolfe algorithm is sublinear in general: the error in the objective function to the optimum is O(1/k) after k iterations, so long as the gradient is Lipschitz continuous with respect to some norm. The same convergence rate can also be shown if the sub-problems are only solved approximately.[3]

The iterations of the algorithm can always be represented as a sparse convex combination of the extreme points of the feasible set, which has helped to the popularity of the algorithm for sparse greedy optimization in machine learning and signal processing problems,[4] as well as for example the optimization of minimum–cost flows in transportation networks.[5]

If the feasible set is given by a set of linear constraints, then the subproblem to be solved in each iteration becomes a linear program.

While the worst-case convergence rate with O(1/k) can not be improved in general, faster convergence can be obtained for special problem classes, such as some strongly convex problems.[6]

Lower bounds on the solution value, and primal-dual analysis

Since f is convex, for any two points 𝐱,𝐲𝒟 we have:

f(𝐲)f(𝐱)+(𝐲𝐱)Tf(𝐱)

This also holds for the (unknown) optimal solution 𝐱*. That is, f(𝐱*)f(𝐱)+(𝐱*𝐱)Tf(𝐱). The best lower bound with respect to a given point 𝐱 is given by

f(𝐱*)f(𝐱)+(𝐱*𝐱)Tf(𝐱)min𝐲D{f(𝐱)+(𝐲𝐱)Tf(𝐱)}=f(𝐱)𝐱Tf(𝐱)+min𝐲D𝐲Tf(𝐱)

The latter optimization problem is solved in every iteration of the Frank–Wolfe algorithm, therefore the solution 𝐬k of the direction-finding subproblem of the k-th iteration can be used to determine increasing lower bounds lk during each iteration by setting l0= and

lk:=max(lk1,f(𝐱k)+(𝐬k𝐱k)Tf(𝐱k))

Such lower bounds on the unknown optimal value are important in practice because they can be used as a stopping criterion, and give an efficient certificate of the approximation quality in every iteration, since always lkf(𝐱*)f(𝐱k).

It has been shown that this corresponding duality gap, that is the difference between f(𝐱k) and the lower bound lk, decreases with the same convergence rate, i.e. f(𝐱k)lk=O(1/k).

Notes

Template:Reflist

Bibliography

See also

Template:Optimization algorithms