Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Subgradient method

From Wikipedia, the free encyclopedia
Concept in convex optimization mathematics

Subgradient methods areconvex optimization methods which usesubderivatives. Originally developed byNaum Z. Shor and others in the 1960s and 1970s, subgradient methods are convergent when applied even to a non-differentiable objective function. When the objective function is differentiable, subgradient methods for unconstrained problems use the same search direction as the method ofgradient descent.

Subgradient methods are slower than Newton's method when applied to minimize twice continuously differentiable convex functions. However, Newton's method fails to converge on problems that have non-differentiable kinks.

In recent years, someinterior-point methods have been suggested for convex minimization problems, but subgradient projection methods and related bundle methods of descent remain competitive. For convex minimization problems with very large number of dimensions, subgradient-projection methods are suitable, because they require little storage.

Subgradient projection methods are often applied to large-scale problems with decomposition techniques. Such decomposition methods often allow a simple distributed method for a problem.

Classical subgradient rules

[edit]

Letf:RnR{\displaystyle f:\mathbb {R} ^{n}\to \mathbb {R} } be aconvex function with domainRn.{\displaystyle \mathbb {R} ^{n}.}A classical subgradient method iteratesx(k+1)=x(k)αkg(k) {\displaystyle x^{(k+1)}=x^{(k)}-\alpha _{k}g^{(k)}\ }whereg(k){\displaystyle g^{(k)}} denotesanysubgradient off {\displaystyle f\ } atx(k), {\displaystyle x^{(k)},\ } andx(k){\displaystyle x^{(k)}} is thekth{\displaystyle k^{th}} iterate ofx.{\displaystyle x.} Iff {\displaystyle f\ } is differentiable, then its only subgradient is the gradient vectorf{\displaystyle \nabla f} itself.It may happen thatg(k){\displaystyle -g^{(k)}} is not a descent direction forf {\displaystyle f\ } atx(k).{\displaystyle x^{(k)}.} We therefore maintain a listfbest {\displaystyle f_{\rm {best}}\ } that keeps track of the lowest objective function value found so far, i.e.fbest(k)=min{fbest(k1),f(x(k))}.{\displaystyle f_{\rm {best}}^{(k)}=\min\{f_{\rm {best}}^{(k-1)},f(x^{(k)})\}.}

Step size rules

[edit]

Many different types of step-size rules are used by subgradient methods. This article notes five classical step-size rules for which convergenceproofs are known:

For all five rules, the step-sizes are determined "off-line", before the method is iterated; the step-sizes do not depend on preceding iterations. This "off-line" property of subgradient methods differs from the "on-line" step-size rules used for descent methods for differentiable functions: Many methods for minimizing differentiable functions satisfy Wolfe's sufficient conditions for convergence, where step-sizes typically depend on the current point and the current search-direction. An extensive discussion of stepsize rules for subgradient methods, including incremental versions, is given in the books by Bertsekas[1] and by Bertsekas, Nedic, and Ozdaglar.[2]

Convergence results

[edit]

For constant step-length and scaled subgradients havingEuclidean norm equal to one, the subgradient method converges to an arbitrarily close approximation to the minimum value, that is

limkfbest(k)f<ϵ{\displaystyle \lim _{k\to \infty }f_{\rm {best}}^{(k)}-f^{*}<\epsilon } by a result ofShor.[3]

These classical subgradient methods have poor performance and are no longer recommended for general use.[4][5] However, they are still used widely in specialized applications because they are simple and they can be easily adapted to take advantage of the special structure of the problem at hand.

Subgradient-projection and bundle methods

[edit]

During the 1970s,Claude Lemaréchal and Phil Wolfe proposed "bundle methods" of descent for problems of convex minimization.[6] The meaning of the term "bundle methods" has changed significantly since that time. Modern versions and full convergence analysis were provided by Kiwiel.[7] Contemporary bundle-methods often use "level control" rules for choosing step-sizes, developing techniques from the "subgradient-projection" method of Boris T. Polyak (1969). However, there are problems on which bundle methods offer little advantage over subgradient-projection methods.[4][5]

Constrained optimization

[edit]

Projected subgradient

[edit]

One extension of the subgradient method is theprojected subgradient method, which solves the constrainedoptimization problem

minimizef(x) {\displaystyle f(x)\ } subject toxC{\displaystyle x\in {\mathcal {C}}}

whereC{\displaystyle {\mathcal {C}}} is aconvex set. The projected subgradient method uses the iterationx(k+1)=P(x(k)αkg(k)){\displaystyle x^{(k+1)}=P\left(x^{(k)}-\alpha _{k}g^{(k)}\right)}whereP{\displaystyle P} is projection onC{\displaystyle {\mathcal {C}}} andg(k){\displaystyle g^{(k)}} is any subgradient off {\displaystyle f\ } atx(k).{\displaystyle x^{(k)}.}

General constraints

[edit]

The subgradient method can be extended to solve the inequality constrained problem

minimizef0(x) {\displaystyle f_{0}(x)\ } subject tofi(x)0,i=1,,m{\displaystyle f_{i}(x)\leq 0,\quad i=1,\ldots ,m}

wherefi{\displaystyle f_{i}} are convex. The algorithm takes the same form as the unconstrained casex(k+1)=x(k)αkg(k) {\displaystyle x^{(k+1)}=x^{(k)}-\alpha _{k}g^{(k)}\ }whereαk>0{\displaystyle \alpha _{k}>0} is a step size, andg(k){\displaystyle g^{(k)}} is a subgradient of the objective or one of the constraint functions atx. {\displaystyle x.\ } Takeg(k)={f0(x) if fi(x)0i=1mfj(x) for some j such that fj(x)>0{\displaystyle g^{(k)}={\begin{cases}\partial f_{0}(x)&{\text{ if }}f_{i}(x)\leq 0\;\forall i=1\dots m\\\partial f_{j}(x)&{\text{ for some }}j{\text{ such that }}f_{j}(x)>0\end{cases}}}wheref{\displaystyle \partial f} denotes thesubdifferential off. {\displaystyle f.\ } If the current point is feasible, the algorithm uses an objective subgradient; if the current point is infeasible, the algorithm chooses a subgradient of any violated constraint.

See also

[edit]

References

[edit]
  1. ^Bertsekas, Dimitri P. (2015).Convex Optimization Algorithms (Second ed.). Belmont, MA.: Athena Scientific.ISBN 978-1-886529-28-1.
  2. ^Bertsekas, Dimitri P.; Nedic, Angelia; Ozdaglar, Asuman (2003).Convex Analysis and Optimization (Second ed.). Belmont, MA.: Athena Scientific.ISBN 1-886529-45-0.
  3. ^The approximate convergence of the constant step-size (scaled) subgradient method is stated as Exercise 6.3.14(a) inBertsekas (page 636):Bertsekas, Dimitri P. (1999).Nonlinear Programming (Second ed.). Cambridge, MA.: Athena Scientific.ISBN 1-886529-00-0. On page 636, Bertsekas attributes this result to Shor:Shor, Naum Z. (1985).Minimization Methods for Non-differentiable Functions.Springer-Verlag.ISBN 0-387-12763-1.
  4. ^abLemaréchal, Claude (2001). "Lagrangian relaxation". In Michael Jünger and Denis Naddef (ed.).Computational combinatorial optimization: Papers from the Spring School held in Schloß Dagstuhl, May 15–19, 2000. Lecture Notes in Computer Science. Vol. 2241. Berlin: Springer-Verlag. pp. 112–156.doi:10.1007/3-540-45586-8_4.ISBN 3-540-42877-1.MR 1900016.S2CID 9048698.
  5. ^abKiwiel, Krzysztof C.; Larsson, Torbjörn; Lindberg, P. O. (August 2007)."Lagrangian relaxation via ballstep subgradient methods"(PDF).Mathematics of Operations Research.32 (3):669–686.doi:10.1287/moor.1070.0261.MR 2348241.
  6. ^Bertsekas, Dimitri P. (1999).Nonlinear Programming (Second ed.). Cambridge, MA.: Athena Scientific.ISBN 1-886529-00-0.
  7. ^Kiwiel, Krzysztof (1985).Methods of Descent for Nondifferentiable Optimization. Berlin:Springer Verlag. p. 362.ISBN 978-3540156420.MR 0797754.

Further reading

[edit]

External links

[edit]
  • EE364A andEE364B, Stanford's convex optimization course sequence.
Basic concepts
Topics (list)
Maps
Mainresults (list)
Sets
Series
Duality
Applications and related
Functions
Gradients
Convergence
Quasi–Newton
Other methods
Hessians
Graph of a strictly concave quadratic function with unique maximum.
Optimization computes maxima and minima.
General
Differentiable
Convex
minimization
Linear and
quadratic
Interior point
Basis-exchange
Paradigms
Graph
algorithms
Minimum
spanning tree
Shortest path
Network flows
Retrieved from "https://en.wikipedia.org/w/index.php?title=Subgradient_method&oldid=1334089787"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp