|This article needs additional citations for verification. (May 2012) (Learn how and when to remove this template message)|
In mathematical optimization, Wolfe duality, named after Philip Wolfe, is type of dual problem in which the objective function and constraints are all differentiable functions. Using this concept a lower bound for a minimization problem can be found because of the weak duality principle.
For a minimization problem with inequality constraints,
the Lagrangian dual problem is
where the objective function is the Lagrange dual function. Provided that the functions and are continuously differentiable, the infimum occurs where the gradient is equal to zero. The problem
is called the Wolfe dual problem. This problem employs the KKT conditions as a constraint. This problem may be difficult to deal with computationally, because the objective function is not concave in the joint variables . Also, the equality constraint is nonlinear in general, so the Wolfe dual problem is typically a nonconvex optimization problem. In any case, weak duality holds.
- Philip Wolfe (1961). "A duality theorem for non-linear programming". Quarterly of Applied Mathematics. 19: 239–244.
- "Chapter 3. Duality in convex optimization" (pdf). October 30, 2011. Retrieved May 20, 2012.
- Geoffrion, Arthur M. (1971). "Duality in Nonlinear Programming: A Simplified Applications-Oriented Development". SIAM Review. 13 (1): 1–37. doi:10.1137/1013001. JSTOR 2028848.
|This applied mathematics-related article is a stub. You can help Wikipedia by expanding it.|