A Simple Convex Optimization AlgorithmIn this section, we aim to show that convex optimization is a computationally tractable problem, by presenting a simple method for general-purpose convex optimization, the subgradient method.
The subgradient methodThe subgradient method is a simple algorithm for minimizing a non-differentiable convex function, and more generally, solving convex optimization problems. Its complexity in terms of problem size is very good (each iteration is cheap), but in terms of accuracy, very poor (the algorithm typically requires thousands or millions of iterations). The convergence analysis of the algorithm has an important theoretical consequence: it shows that convex optimization is essentially ‘‘tractable’’, provided we can compute subgradients of the objective and constraints functions efficiently. The method is part of a large class of methods known as first-order methods, which are based on subgradients or closely related objects. In constrast, second-order methods typically use Hessian information; each iteration is expensive (both in flops and memory) but convergence is obtained very quickly. SetupThe setup is as follows. Let be a convex function. We seek to solve the uconstrained problem
Now consider the convex, inequality constrained problem:
AlgorithmThe algorithm proceeds as follows:
Convergence analysis and complexityTo analyze the convergence, we make the following assumptions:
Now let be any minimizer of . We have
This shows that
ComplexityThe convergence analysis result (ref{eq:conv-anal-L19}) depends on the choice of the step size rule. Which rule is optimal for this bound? The problem of minimizing the upper bound above is convex and symmetric (the function does not change when we exchange variables). Hence, the optimal ’s are all equal at optimum, to a number . With this choice of step length rule, the number of iterations needed to achieve -sub-optimality, as predicted by the analysis, is . The dependence on is now , which is very slow. Interior-point methods alluded to earlier typically behave in ), but they do not apply to general convex problems, only to a special class. The above discusses the dependence of complexity on the accuracy . The complexity depends on problem size as well, and more generally, on the structure of the function involved, via the need to compute subgradients at each step, and also the parameters and . Example: We seek to find a point in a given intersection of closed convex sets,
Projected subgradient methodOne extension of the subgradient method for solving constrained
optimization problems, is the projected subgradient method. We consider the problem
Example:
Let , , with , full rank (hence, ). We consider the problem
|