Wolfe duality

From HandWiki

In mathematical optimization, Wolfe duality, named after Philip Wolfe, is type of dual problem in which the objective function and constraints are all differentiable functions. Using this concept a lower bound for a minimization problem can be found because of the weak duality principle.[1]

Mathematical formulation

For a minimization problem with inequality constraints,

minimizexf(x)subjecttogi(x)0,i=1,,m

the Lagrangian dual problem is

maximizeuinfx(f(x)+j=1mujgj(x))subjecttoui0,i=1,,m

where the objective function is the Lagrange dual function. Provided that the functions f and g1,,gm are convex and continuously differentiable, the infimum occurs where the gradient is equal to zero. The problem

maximizex,uf(x)+j=1mujgj(x)subjecttof(x)+j=1mujgj(x)=0ui0,i=1,,m

is called the Wolfe dual problem.[2][clarification needed] This problem employs the KKT conditions as a constraint. Also, the equality constraint f(x)+j=1mujgj(x) is nonlinear in general, so the Wolfe dual problem may be a nonconvex optimization problem. In any case, weak duality holds.[3]

See also

  • Lagrangian duality
  • Fenchel duality

References

  1. Philip Wolfe (1961). "A duality theorem for non-linear programming". Quarterly of Applied Mathematics 19 (3): 239–244. doi:10.1090/qam/135625. 
  2. Eiselt, Horst A. (2019). Nonlinear Optimization: Methods and Applications. International Series in Operations Research and Management Science Ser. Carl-Louis Sandblom. Cham: Springer International Publishing AG. pp. 147. ISBN 978-3-030-19462-8. https://books.google.com/books?id=SlW9DwAAQBAJ. 
  3. Geoffrion, Arthur M. (1971). "Duality in Nonlinear Programming: A Simplified Applications-Oriented Development". SIAM Review 13 (1): 1–37. doi:10.1137/1013001.