Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Karush–Kuhn–Tucker conditions

From Wikipedia, the free encyclopedia
Concept in mathematical optimization

Inmathematical optimization, theKarush–Kuhn–Tucker (KKT)conditions, also known as theKuhn–Tucker conditions, arefirst derivative tests (sometimes called first-ordernecessary conditions) for a solution innonlinear programming to beoptimal, provided that someregularity conditions are satisfied.

Allowing inequality constraints, the KKT approach to nonlinear programming generalizes the method ofLagrange multipliers, which allows only equality constraints. Similar to the Lagrange approach, the constrained maximization (minimization) problem is rewritten as a Lagrange function whose optimal point is a global maximum or minimum over the domain of the choice variables and a global minimum (maximum) over the multipliers. The Karush–Kuhn–Tucker theorem is sometimes referred to as the saddle-point theorem.[1]

The KKT conditions were originally named afterHarold W. Kuhn andAlbert W. Tucker, who first published the conditions in 1951.[2] Later scholars discovered that the necessary conditions for this problem had been stated byWilliam Karush in his master's thesis in 1939.[3][4]

Nonlinear optimization problem

[edit]

Consider the following nonlinear optimization problem instandard form:

minimizef(x){\displaystyle f(\mathbf {x} )}
subject to
gi(x)0,{\displaystyle g_{i}(\mathbf {x} )\leq 0,}
hj(x)=0.{\displaystyle h_{j}(\mathbf {x} )=0.}

wherexX{\displaystyle \mathbf {x} \in \mathbf {X} } is the optimization variable chosen from aconvex subset ofRn{\displaystyle \mathbb {R} ^{n}},f{\displaystyle f} is theobjective orutility function,gi (i=1,,m){\displaystyle g_{i}\ (i=1,\ldots ,m)} are the inequalityconstraint functions andhj (j=1,,){\displaystyle h_{j}\ (j=1,\ldots ,\ell )} are the equalityconstraint functions. The numbers of inequalities and equalities are denoted bym{\displaystyle m} and{\displaystyle \ell } respectively. Corresponding to the constrained optimization problem one can form the Lagrangian function

L(x,μ,λ)=f(x)+μg(x)+λh(x)=L(x,α)=f(x)+α(g(x)h(x)){\displaystyle {\mathcal {L}}(\mathbf {x} ,\mathbf {\mu } ,\mathbf {\lambda } )=f(\mathbf {x} )+\mathbf {\mu } ^{\top }\mathbf {g} (\mathbf {x} )+\mathbf {\lambda } ^{\top }\mathbf {h} (\mathbf {x} )=L(\mathbf {x} ,\mathbf {\alpha } )=f(\mathbf {x} )+\mathbf {\alpha } ^{\top }{\begin{pmatrix}\mathbf {g} (\mathbf {x} )\\\mathbf {h} (\mathbf {x} )\end{pmatrix}}}

where

g(x)=[g1(x)gi(x)gm(x)],h(x)=[h1(x)hj(x)h(x)],μ=[μ1μiμm],λ=[λ1λjλ]andα=[μλ].{\displaystyle \mathbf {g} \left(\mathbf {x} \right)={\begin{bmatrix}g_{1}\left(\mathbf {x} \right)\\\vdots \\g_{i}\left(\mathbf {x} \right)\\\vdots \\g_{m}\left(\mathbf {x} \right)\end{bmatrix}},\quad \mathbf {h} \left(\mathbf {x} \right)={\begin{bmatrix}h_{1}\left(\mathbf {x} \right)\\\vdots \\h_{j}\left(\mathbf {x} \right)\\\vdots \\h_{\ell }\left(\mathbf {x} \right)\end{bmatrix}},\quad \mathbf {\mu } ={\begin{bmatrix}\mu _{1}\\\vdots \\\mu _{i}\\\vdots \\\mu _{m}\\\end{bmatrix}},\quad \mathbf {\lambda } ={\begin{bmatrix}\lambda _{1}\\\vdots \\\lambda _{j}\\\vdots \\\lambda _{\ell }\end{bmatrix}}\quad {\text{and}}\quad \mathbf {\alpha } ={\begin{bmatrix}\mu \\\lambda \end{bmatrix}}.} TheKarush–Kuhn–Tucker theorem then states the following.

Theorem(sufficiency) If(x,α){\displaystyle (\mathbf {x} ^{\ast },\mathbf {\alpha } ^{\ast })} is asaddle point ofL(x,α){\displaystyle L(\mathbf {x} ,\mathbf {\alpha } )} inxX{\displaystyle \mathbf {x} \in \mathbf {X} },μ0{\displaystyle \mathbf {\mu } \geq \mathbf {0} }, thenx{\displaystyle \mathbf {x} ^{\ast }} is an optimal vector for the above optimization problem.

(necessity) Suppose thatf(x){\displaystyle f(\mathbf {x} )} andgi(x){\displaystyle g_{i}(\mathbf {x} )},i=1,,m{\displaystyle i=1,\ldots ,m}, areconvex inX{\displaystyle \mathbf {X} } and that there existsx0relint(X){\displaystyle \mathbf {x} _{0}\in \operatorname {relint} (\mathbf {X} )} such thatg(x0)<0{\displaystyle \mathbf {g} (\mathbf {x} _{0})<\mathbf {0} } (i.e.,Slater's condition holds). Then with an optimal vectorx{\displaystyle \mathbf {x} ^{\ast }} for the above optimization problem there is associated a vectorα=[μλ]{\displaystyle \mathbf {\alpha } ^{\ast }={\begin{bmatrix}\mu ^{*}\\\lambda ^{*}\end{bmatrix}}} satisfyingμ0{\displaystyle \mathbf {\mu } ^{*}\geq \mathbf {0} } such that(x,α){\displaystyle (\mathbf {x} ^{\ast },\mathbf {\alpha } ^{\ast })} is a saddle point ofL(x,α){\displaystyle L(\mathbf {x} ,\mathbf {\alpha } )}.[5]

Since the idea of this approach is to find asupporting hyperplane on the feasible setΓ={xX:gi(x)0,i=1,,m}{\displaystyle \mathbf {\Gamma } =\left\{\mathbf {x} \in \mathbf {X} :g_{i}(\mathbf {x} )\leq 0,i=1,\ldots ,m\right\}}, the proof of the Karush–Kuhn–Tucker theorem makes use of thehyperplane separation theorem.[6]

The system of equations and inequalities corresponding to the KKT conditions is usually not solved directly, except in the few special cases where aclosed-form solution can be derived analytically. In general, many optimization algorithms can be interpreted as methods for numerically solving the KKT system of equations and inequalities.[7]

Necessary conditions

[edit]

Suppose that theobjective functionf:RnR{\displaystyle f\colon \mathbb {R} ^{n}\rightarrow \mathbb {R} } and the constraint functionsgi:RnR{\displaystyle g_{i}\colon \mathbb {R} ^{n}\rightarrow \mathbb {R} } andhj:RnR{\displaystyle h_{j}\colon \mathbb {R} ^{n}\rightarrow \mathbb {R} } havesubderivatives at a pointxRn{\displaystyle x^{*}\in \mathbb {R} ^{n}}. Ifx{\displaystyle x^{*}} is alocal optimum and the optimization problem satisfies some regularity conditions (see below), then there exist constantsμi (i=1,,m){\displaystyle \mu _{i}\ (i=1,\ldots ,m)} andλj (j=1,,){\displaystyle \lambda _{j}\ (j=1,\ldots ,\ell )}, called KKT multipliers, such that the following four groups of conditions hold:[8]

Inequality constraint diagram for optimization problems
Stationarity
For minimizingf(x){\displaystyle f(x)}:f(x)+j=1λjhj(x)+i=1mμigi(x)0{\displaystyle \partial f(x^{*})+\sum _{j=1}^{\ell }\lambda _{j}\partial h_{j}(x^{*})+\sum _{i=1}^{m}\mu _{i}\partial g_{i}(x^{*})\ni \mathbf {0} }
For maximizingf(x){\displaystyle f(x)}:f(x)+j=1λjhj(x)+i=1mμigi(x)0{\displaystyle -\partial f(x^{*})+\sum _{j=1}^{\ell }\lambda _{j}\partial h_{j}(x^{*})+\sum _{i=1}^{m}\mu _{i}\partial g_{i}(x^{*})\ni \mathbf {0} }
Primal feasibility
hj(x)=0, for j=1,,{\displaystyle h_{j}(x^{*})=0,{\text{ for }}j=1,\ldots ,\ell \,\!}
gi(x)0, for i=1,,m{\displaystyle g_{i}(x^{*})\leq 0,{\text{ for }}i=1,\ldots ,m}
Dual feasibility
μi0, for i=1,,m{\displaystyle \mu _{i}\geq 0,{\text{ for }}i=1,\ldots ,m}
Complementary slackness
i=1mμigi(x)=0.{\displaystyle \sum _{i=1}^{m}\mu _{i}g_{i}(x^{*})=0.}

The last condition is sometimes written in the equivalent form:μigi(x)=0, for i=1,,m.{\displaystyle \mu _{i}g_{i}(x^{*})=0,{\text{ for }}i=1,\ldots ,m.}

In the particular casem=0{\displaystyle m=0}, i.e., when there are no inequality constraints, the KKT conditions turn into the Lagrange conditions, and the KKT multipliers are calledLagrange multipliers.

Interpretation: KKT conditions as balancing constraint-forces in state space

[edit]

The primal problem can be interpreted as moving a particle in the space ofx{\displaystyle x}, and subjecting it to three kinds of force fields:

Primal stationarity states that the "force" off(x){\displaystyle \partial f(x^{*})} is exactly balanced by a linear sum of forceshj(x){\displaystyle \partial h_{j}(x^{*})} andgi(x){\displaystyle \partial g_{i}(x^{*})}.

Dual feasibility additionally states that all thegi(x){\displaystyle \partial g_{i}(x^{*})} forces must be one-sided, pointing inwards into the feasible set forx{\displaystyle x}.

Complementary slackness states that ifgi(x)<0{\displaystyle g_{i}(x^{*})<0}, then the force coming fromgi(x){\displaystyle \partial g_{i}(x^{*})} must be zero i.e.,μi(x)=0{\displaystyle \mu _{i}(x^{*})=0}, since the particle is not on the boundary, the one-sided constraint force cannot activate.

Matrix representation

[edit]

The necessary conditions can be written withJacobian matrices of the constraint functions. Letg(x):RnRm{\displaystyle \mathbf {g} (x):\,\!\mathbb {R} ^{n}\rightarrow \mathbb {R} ^{m}} be defined asg(x)=(g1(x),,gm(x)){\displaystyle \mathbf {g} (x)=\left(g_{1}(x),\ldots ,g_{m}(x)\right)^{\top }} and leth(x):RnR{\displaystyle \mathbf {h} (x):\,\!\mathbb {R} ^{n}\rightarrow \mathbb {R} ^{\ell }} be defined ash(x)=(h1(x),,h(x)){\displaystyle \mathbf {h} (x)=\left(h_{1}(x),\ldots ,h_{\ell }(x)\right)^{\top }}. Letμ=(μ1,,μm){\displaystyle {\boldsymbol {\mu }}=\left(\mu _{1},\ldots ,\mu _{m}\right)^{\top }} andλ=(λ1,,λ){\displaystyle {\boldsymbol {\lambda }}=\left(\lambda _{1},\ldots ,\lambda _{\ell }\right)^{\top }}. Then the necessary conditions can be written as:

Stationarity
For maximizingf(x){\displaystyle f(x)}:f(x)Dg(x)μDh(x)λ=0{\displaystyle \partial f(x^{*})-D\mathbf {g} (x^{*})^{\top }{\boldsymbol {\mu }}-D\mathbf {h} (x^{*})^{\top }{\boldsymbol {\lambda }}=\mathbf {0} }
For minimizingf(x){\displaystyle f(x)}:f(x)+Dg(x)μ+Dh(x)λ=0{\displaystyle \partial f(x^{*})+D\mathbf {g} (x^{*})^{\top }{\boldsymbol {\mu }}+D\mathbf {h} (x^{*})^{\top }{\boldsymbol {\lambda }}=\mathbf {0} }
Primal feasibility
g(x)0{\displaystyle \mathbf {g} (x^{*})\leq \mathbf {0} }
h(x)=0{\displaystyle \mathbf {h} (x^{*})=\mathbf {0} }
Dual feasibility
μ0{\displaystyle {\boldsymbol {\mu }}\geq \mathbf {0} }
Complementary slackness
μg(x)=0.{\displaystyle {\boldsymbol {\mu }}^{\top }\mathbf {g} (x^{*})=0.}

Regularity conditions (or constraint qualifications)

[edit]

One can ask whether a minimizer pointx{\displaystyle x^{*}} of the original, constrained optimization problem (assuming one exists) has to satisfy the above KKT conditions. This is similar to asking under what conditions the minimizerx{\displaystyle x^{*}} of a functionf(x){\displaystyle f(x)} in an unconstrained problem has to satisfy the conditionf(x)=0{\displaystyle \nabla f(x^{*})=0}. For the constrained case, the situation is more complicated, and one can state a variety of (increasingly complicated) "regularity" conditions under which a constrained minimizer also satisfies the KKT conditions. Some common examples for conditions that guarantee this are tabulated in the following, with the LICQ the most frequently used one:

ConstraintAcronymStatement
Linearity constraint qualificationLCQIfgi{\displaystyle g_{i}} andhj{\displaystyle h_{j}} areaffine functions, then no other condition is needed.
Linear independence constraint qualificationLICQThe gradients of the active inequality constraints and the gradients of the equality constraints arelinearly independent atx{\displaystyle x^{*}}.
Mangasarian-Fromovitz constraint qualificationMFCQThe gradients of the equality constraints are linearly independent atx{\displaystyle x^{*}} and there exists a vectordRn{\displaystyle d\in \mathbb {R} ^{n}} such thatgi(x)d<0{\displaystyle \nabla g_{i}(x^{*})^{\top }d<0} for all active inequality constraints andhj(x)d=0{\displaystyle \nabla h_{j}(x^{*})^{\top }d=0} for all equality constraints.[9]
Constant rank constraint qualificationCRCQFor each subset of the gradients of the active inequality constraints and the gradients of the equality constraints the rank at a vicinity ofx{\displaystyle x^{*}} is constant.
Constant positive linear dependence constraint qualificationCPLDFor each subset of gradients of active inequality constraints and gradients of equality constraints, if the subset of vectors is linearly dependent atx{\displaystyle x^{*}} with non-negative scalars associated with the inequality constraints, then it remains linearly dependent in a neighborhood ofx{\displaystyle x^{*}}.
Quasi-normality constraint qualificationQNCQIf the gradients of the active inequality constraints and the gradients of the equality constraints are linearly dependent atx{\displaystyle x^{*}} with associated multipliersλj{\displaystyle \lambda _{j}} for equalities andμi0{\displaystyle \mu _{i}\geq 0} for inequalities, then there is no sequencexkx{\displaystyle x_{k}\to x^{*}} such thatλj0λjhj(xk)>0{\displaystyle \lambda _{j}\neq 0\Rightarrow \lambda _{j}h_{j}(x_{k})>0} andμi0μigi(xk)>0.{\displaystyle \mu _{i}\neq 0\Rightarrow \mu _{i}g_{i}(x_{k})>0.}
Slater's conditionSCFor aconvex problem (i.e., assuming minimization,f,gi{\displaystyle f,g_{i}} are convex andhj{\displaystyle h_{j}} is affine), there exists a pointx{\displaystyle x} such thathj(x)=0{\displaystyle h_{j}(x)=0} andgi(x)<0.{\displaystyle g_{i}(x)<0.}

The strict implications can be shown

LICQ ⇒ MFCQ ⇒ CPLD ⇒ QNCQ

and

LICQ ⇒ CRCQ ⇒ CPLD ⇒ QNCQ

In practice weaker constraint qualifications are preferred since they apply to a broader selection of problems.

Sufficient conditions

[edit]

In some cases, the necessary conditions are also sufficient for optimality. In general, the necessary conditions are not sufficient for optimality and additional information is required, such as the Second Order Sufficient Conditions (SOSC). For smooth functions, SOSC involve the second derivatives, which explains its name.

The necessary conditions are sufficient for optimality if the objective functionf{\displaystyle f} of a maximization problem is a differentiableconcave function, the inequality constraintsgj{\displaystyle g_{j}} are differentiableconvex functions, the equality constraintshi{\displaystyle h_{i}} areaffine functions, andSlater's condition holds.[10] Similarly, if the objective functionf{\displaystyle f} of a minimization problem is a differentiableconvex function, the necessary conditions are also sufficient for optimality.

It was shown by Martin in 1985 that the broader class of functions in which KKT conditions guarantees global optimality are the so-called Type 1invex functions.[11][12]

Second-order sufficient conditions

[edit]

For smooth,non-linear optimization problems, a second order sufficient condition is given as follows.

The solutionx,λ,μ{\displaystyle x^{*},\lambda ^{*},\mu ^{*}} found in the above section is a constrained local minimum if for the Lagrangian,

L(x,λ,μ)=f(x)+i=1mμigi(x)+j=1λjhj(x){\displaystyle L(x,\lambda ,\mu )=f(x)+\sum _{i=1}^{m}\mu _{i}g_{i}(x)+\sum _{j=1}^{\ell }\lambda _{j}h_{j}(x)}

then,

sTxx2L(x,λ,μ)s0{\displaystyle s^{T}\nabla _{xx}^{2}L(x^{*},\lambda ^{*},\mu ^{*})s\geq 0}

wheres0{\displaystyle s\neq 0} is a vector satisfying the following,

[xgi(x),xhj(x)]Ts=0R2{\displaystyle \left[\nabla _{x}g_{i}(x^{*}),\nabla _{x}h_{j}(x^{*})\right]^{T}s=0_{\mathbb {R} ^{2}}}

where only those active inequality constraintsgi(x){\displaystyle g_{i}(x)} corresponding to strict complementarity (i.e. whereμi>0{\displaystyle \mu _{i}>0}) are applied. The solution is a strict constrained local minimum in the case the inequality is also strict.

IfsTxx2L(x,λ,μ)s=0{\displaystyle s^{T}\nabla _{xx}^{2}L(x^{*},\lambda ^{*},\mu ^{*})s=0}, the third order Taylor expansion of the Lagrangian should be used to verify ifx{\displaystyle x^{*}} is a local minimum. The minimization off(x1,x2)=(x2x12)(x23x12){\displaystyle f(x_{1},x_{2})=(x_{2}-x_{1}^{2})(x_{2}-3x_{1}^{2})} is a good counter-example, see alsoPeano surface.

Economics

[edit]
See also:Profit maximization

Often inmathematical economics the KKT approach is used in theoretical models in order to obtain qualitative results. For example,[13] consider a firm that maximizes its sales revenue subject to a minimum profit constraint. LettingQ{\displaystyle Q} be the quantity of output produced (to be chosen),R(Q){\displaystyle R(Q)} be sales revenue with a positive first derivative and with a zero value at zero output,C(Q){\displaystyle C(Q)} be production costs with a positive first derivative and with a non-negative value at zero output, andGmin{\displaystyle G_{\min }} be the positive minimal acceptable level ofprofit, then the problem is a meaningful one if the revenue function levels off so it eventually is less steep than the cost function. The problem expressed in the previously given minimization form is

MinimizeR(Q){\displaystyle -R(Q)}
subject to
GminR(Q)C(Q){\displaystyle G_{\min }\leq R(Q)-C(Q)}
Q0,{\displaystyle Q\geq 0,}

and the KKT conditions are

(dRdQ)(1+μ)μ(dCdQ)0,Q0,Q[(dRdQ)(1+μ)μ(dCdQ)]=0,R(Q)C(Q)Gmin0,μ0,μ[R(Q)C(Q)Gmin]=0.{\displaystyle {\begin{aligned}&\left({\frac {{\text{d}}R}{{\text{d}}Q}}\right)(1+\mu )-\mu \left({\frac {{\text{d}}C}{{\text{d}}Q}}\right)\leq 0,\\[5pt]&Q\geq 0,\\[5pt]&Q\left[\left({\frac {{\text{d}}R}{{\text{d}}Q}}\right)(1+\mu )-\mu \left({\frac {{\text{d}}C}{{\text{d}}Q}}\right)\right]=0,\\[5pt]&R(Q)-C(Q)-G_{\min }\geq 0,\\[5pt]&\mu \geq 0,\\[5pt]&\mu [R(Q)-C(Q)-G_{\min }]=0.\end{aligned}}}

SinceQ=0{\displaystyle Q=0} would violate the minimum profit constraint, we haveQ>0{\displaystyle Q>0} and hence the third condition implies that the first condition holds with equality. Solving that equality gives

dRdQ=μ1+μ(dCdQ).{\displaystyle {\frac {{\text{d}}R}{{\text{d}}Q}}={\frac {\mu }{1+\mu }}\left({\frac {{\text{d}}C}{{\text{d}}Q}}\right).}

Because it was given thatdR/dQ{\displaystyle {\text{d}}R/{\text{d}}Q} anddC/dQ{\displaystyle {\text{d}}C/{\text{d}}Q} are strictly positive, this inequality along with the non-negativity condition onμ{\displaystyle \mu } guarantees thatμ{\displaystyle \mu } is positive and so the revenue-maximizing firm operates at a level of output at whichmarginal revenuedR/dQ{\displaystyle {\text{d}}R/{\text{d}}Q} is less thanmarginal costdC/dQ{\displaystyle {\text{d}}C/{\text{d}}Q} — a result that is of interest because it contrasts with the behavior of aprofit maximizing firm, which operates at a level at which they are equal.

Value function

[edit]

If we reconsider the optimization problem as a maximization problem with constant inequality constraints:

Maximize f(x){\displaystyle {\text{Maximize }}\;f(x)}
subject to  {\displaystyle {\text{subject to }}\ }
gi(x)ai,hj(x)=0.{\displaystyle g_{i}(x)\leq a_{i},h_{j}(x)=0.}

The value function is defined as

V(a1,,an)=supxf(x){\displaystyle V(a_{1},\ldots ,a_{n})=\sup \limits _{x}f(x)}
subject to  {\displaystyle {\text{subject to }}\ }
gi(x)ai,hj(x)=0{\displaystyle g_{i}(x)\leq a_{i},h_{j}(x)=0}
j{1,,},i{1,,m},{\displaystyle j\in \{1,\ldots ,\ell \},i\in \{1,\ldots ,m\},}

so the domain ofV{\displaystyle V} is{aRmfor some xX,gi(x)ai,i{1,,m}}.{\displaystyle \{a\in \mathbb {R} ^{m}\mid {\text{for some }}x\in X,g_{i}(x)\leq a_{i},i\in \{1,\ldots ,m\}\}.}

Given this definition, each coefficientμi{\displaystyle \mu _{i}} is the rate at which the value function increases asai{\displaystyle a_{i}} increases. Thus if eachai{\displaystyle a_{i}} is interpreted as a resource constraint, the coefficients tell you how much increasing a resource will increase the optimum value of our functionf{\displaystyle f}. This interpretation is especially important in economics and is used, for instance, inutility maximization problems.

Generalizations

[edit]

With an extra multiplierμ00{\displaystyle \mu _{0}\geq 0}, which may be zero (as long as(μ0,μ,λ)0{\displaystyle (\mu _{0},\mu ,\lambda )\neq 0}), in front off(x){\displaystyle \nabla f(x^{*})} the KKT stationarity conditions turn into

μ0f(x)+i=1mμigi(x)+j=1λjhj(x)=0,μjgi(x)=0,i=1,,m,{\displaystyle {\begin{aligned}&\mu _{0}\,\nabla f(x^{*})+\sum _{i=1}^{m}\mu _{i}\,\nabla g_{i}(x^{*})+\sum _{j=1}^{\ell }\lambda _{j}\,\nabla h_{j}(x^{*})=0,\\[4pt]&\mu _{j}g_{i}(x^{*})=0,\quad i=1,\dots ,m,\end{aligned}}}

which are called theFritz John conditions. This optimality conditions holds without constraint qualifications and it is equivalent to the optimality conditionKKT or (not-MFCQ).

The KKT conditions belong to a wider class of the first-order necessary conditions (FONC), which allow for non-smooth functions usingsubderivatives.

See also

[edit]

References

[edit]
  1. ^Tabak, Daniel; Kuo, Benjamin C. (1971).Optimal Control by Mathematical Programming. Englewood Cliffs, NJ: Prentice-Hall. pp. 19–20.ISBN 0-13-638106-5.
  2. ^Kuhn, H. W.;Tucker, A. W. (1951)."Nonlinear programming".Proceedings of 2nd Berkeley Symposium. Berkeley: University of California Press. pp. 481–492.MR 0047303.
  3. ^W. Karush (1939).Minima of Functions of Several Variables with Inequalities as Side Constraints (M.Sc. thesis). Dept. of Mathematics, Univ. of Chicago, Chicago, Illinois.
  4. ^Kjeldsen, Tinne Hoff (2000)."A contextualized historical analysis of the Kuhn-Tucker theorem in nonlinear programming: the impact of World War II".Historia Math.27 (4):331–361.doi:10.1006/hmat.2000.2289.MR 1800317.
  5. ^Walsh, G. R. (1975)."Saddle-point Property of Lagrangian Function".Methods of Optimization. New York: John Wiley & Sons. pp. 39–44.ISBN 0-471-91922-5.
  6. ^Kemp, Murray C.; Kimura, Yoshio (1978).Introduction to Mathematical Economics. New York: Springer. pp. 38–44.ISBN 0-387-90304-6.
  7. ^Boyd, Stephen; Vandenberghe, Lieven (2004).Convex Optimization. Cambridge:Cambridge University Press. p. 244.ISBN 0-521-83378-7.MR 2061575.
  8. ^Ruszczyński, Andrzej (2006).Nonlinear Optimization. Princeton, NJ:Princeton University Press.ISBN 978-0691119151.MR 2199043.
  9. ^Dimitri Bertsekas (1999).Nonlinear Programming (2 ed.). Athena Scientific. pp. 329–330.ISBN 9781886529007.
  10. ^Boyd, Stephen; Vandenberghe, Lieven (2004).Convex Optimization. Cambridge:Cambridge University Press. p. 244.ISBN 0-521-83378-7.MR 2061575.
  11. ^Martin, D. H. (1985). "The Essence of Invexity".J. Optim. Theory Appl.47 (1):65–76.doi:10.1007/BF00941316.S2CID 122906371.
  12. ^Hanson, M. A. (1999)."Invexity and the Kuhn-Tucker Theorem".J. Math. Anal. Appl.236 (2):594–604.doi:10.1006/jmaa.1999.6484.
  13. ^Chiang, Alpha C.Fundamental Methods of Mathematical Economics, 3rd edition, 1984, pp. 750–752.

Further reading

[edit]

External links

[edit]
Retrieved from "https://en.wikipedia.org/w/index.php?title=Karush–Kuhn–Tucker_conditions&oldid=1316796544"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp