Karush–Kuhn–Tucker conditions#Regularity conditions .28or constraint qualifications.29

{{Short description|Concept in mathematical optimization}}

In mathematical optimization, the Karush–Kuhn–Tucker (KKT) conditions, also known as the Kuhn–Tucker conditions, are first derivative tests (sometimes called first-order necessary conditions) for a solution in nonlinear programming to be optimal, provided that some regularity conditions are satisfied.

Allowing inequality constraints, the KKT approach to nonlinear programming generalizes the method of Lagrange multipliers, which allows only equality constraints. Similar to the Lagrange approach, the constrained maximization (minimization) problem is rewritten as a Lagrange function whose optimal point is a global maximum or minimum over the domain of the choice variables and a global minimum (maximum) over the multipliers. The Karush–Kuhn–Tucker theorem is sometimes referred to as the saddle-point theorem.{{cite book |first1=Daniel |last1=Tabak |first2=Benjamin C. |last2=Kuo |title=Optimal Control by Mathematical Programming |location=Englewood Cliffs, NJ |publisher=Prentice-Hall |year=1971 |isbn=0-13-638106-5 |pages=19–20 }}

The KKT conditions were originally named after Harold W. Kuhn and Albert W. Tucker, who first published the conditions in 1951.{{cite conference

| first1 = H. W.

| last1 = Kuhn

|author-link1=Harold W. Kuhn

| first2 = A. W.

| last2 = Tucker

|author-link2=Albert W. Tucker

| book-title = Proceedings of 2nd Berkeley Symposium

| pages = 481–492

| title=Nonlinear programming

| publisher = University of California Press

| year = 1951

| location = Berkeley

| url = http://projecteuclid.org/euclid.bsmsp/1200500249

|mr=47303}} Later scholars discovered that the necessary conditions for this problem had been stated by William Karush in his master's thesis in 1939.{{cite thesis

|author=W. Karush

|title=Minima of Functions of Several Variables with Inequalities as Side Constraints

|type=M.Sc. thesis

|publisher=Dept. of Mathematics, Univ. of Chicago, Chicago, Illinois

|year=1939

|url=http://pi.lib.uchicago.edu/1001/cat/bib/4111654

}}

{{cite journal |last=Kjeldsen |first=Tinne Hoff|author-link= Tinne Hoff Kjeldsen |title=A contextualized historical analysis of the Kuhn-Tucker theorem in nonlinear programming: the impact of World War II |journal=Historia Math. |volume=27 |year=2000 |issue=4 |pages=331–361 |mr=1800317 |doi=10.1006/hmat.2000.2289|doi-access=free }}

Nonlinear optimization problem

Consider the following nonlinear optimization problem in standard form:

:minimize f(\mathbf{x})

:subject to

:: g_i(\mathbf{x}) \leq 0,

:: h_j(\mathbf{x}) =0.

where \mathbf{x} \in \mathbf{X} is the optimization variable chosen from a convex subset of \mathbb{R}^{n}, f is the objective or utility function, g_i \ (i = 1, \ldots, m) are the inequality constraint functions and h_j \ (j = 1, \ldots, \ell) are the equality constraint functions. The numbers of inequalities and equalities are denoted by m and \ell respectively. Corresponding to the constrained optimization problem one can form the Lagrangian function

\mathcal{L}(\mathbf{x},\mathbf{\mu},\mathbf{\lambda}) = f(\mathbf{x}) + \mathbf{\mu}^\top \mathbf{g}(\mathbf{x}) + \mathbf{\lambda}^\top \mathbf{h}(\mathbf{x})=L(\mathbf{x}, \mathbf{\alpha})=f(\mathbf{x})+\mathbf{\alpha}^\top \begin{pmatrix}

\mathbf{g}(\mathbf{x}) \\

\mathbf{h}(\mathbf{x})

\end{pmatrix}

where

\mathbf{g}\left(\mathbf{x}\right) = \begin{bmatrix}

g_{1}\left(\mathbf{x}\right) \\

\vdots \\

g_{i}\left(\mathbf{x}\right) \\

\vdots \\

g_{m}\left(\mathbf{x}\right)

\end{bmatrix},

\quad

\mathbf{h}\left(\mathbf{x}\right) = \begin{bmatrix}

h_{1}\left(\mathbf{x}\right) \\

\vdots \\

h_{j}\left(\mathbf{x}\right) \\

\vdots \\

h_{\ell}\left(\mathbf{x}\right)

\end{bmatrix},

\quad

\mathbf{\mu} = \begin{bmatrix}

\mu_{1} \\

\vdots \\

\mu_{i} \\

\vdots \\

\mu_{m} \\

\end{bmatrix},

\quad

\mathbf{\lambda} = \begin{bmatrix}

\lambda_{1} \\

\vdots \\

\lambda_{j} \\

\vdots \\

\lambda_{\ell}

\end{bmatrix}

\quad \text{and} \quad

\mathbf{\alpha} = \begin{bmatrix}

\mu \\

\lambda

\end{bmatrix}. The Karush–Kuhn–Tucker theorem then states the following.

{{Math theorem

| math_statement =

(sufficiency) If (\mathbf{x}^{\ast},\mathbf{\alpha}^\ast) is a saddle point of L(\mathbf{x},\mathbf{\alpha}) in \mathbf{x} \in \mathbf{X}, \mathbf{\mu} \geq \mathbf{0}, then \mathbf{x}^{\ast} is an optimal vector for the above optimization problem.

(necessity) Suppose that f(\mathbf{x}) and g_i(\mathbf{x}), i = 1, \ldots, m, are convex in \mathbf{X} and that there exists \mathbf{x}_{0} \in \operatorname{relint}(\mathbf{X}) such that \mathbf{g}(\mathbf{x}_{0}) < \mathbf 0 (i.e., Slater's condition holds). Then with an optimal vector \mathbf{x}^{\ast} for the above optimization problem there is associated a vector \mathbf{\alpha}^\ast=\begin{bmatrix}\mu^* \\ \lambda^*\end{bmatrix} satisfying \mathbf \mu^*\ge \mathbf 0 such that (\mathbf{x}^{\ast},\mathbf{\alpha}^\ast) is a saddle point of L(\mathbf{x},\mathbf{\alpha}).{{cite book |first=G. R. |last=Walsh |title=Methods of Optimization |location=New York |publisher=John Wiley & Sons |year=1975 |isbn=0-471-91922-5 |chapter=Saddle-point Property of Lagrangian Function |pages=39–44 |chapter-url=https://books.google.com/books?id=K0EZAQAAIAAJ&pg=PA39 }}

}}

Since the idea of this approach is to find a supporting hyperplane on the feasible set \mathbf{\Gamma} = \left\{ \mathbf{x} \in \mathbf{X} : g_i(\mathbf{x}) \leq 0, i = 1, \ldots, m \right\}, the proof of the Karush–Kuhn–Tucker theorem makes use of the hyperplane separation theorem.{{cite book |first1=Murray C. |last1=Kemp |first2=Yoshio |last2=Kimura |title=Introduction to Mathematical Economics |location=New York |publisher=Springer |year=1978 |isbn=0-387-90304-6 |pages=[https://archive.org/details/introductiontoma0000kemp/page/38 38–44] |url=https://archive.org/details/introductiontoma0000kemp/page/38 }}

The system of equations and inequalities corresponding to the KKT conditions is usually not solved directly, except in the few special cases where a closed-form solution can be derived analytically. In general, many optimization algorithms can be interpreted as methods for numerically solving the KKT system of equations and inequalities.{{cite book|last1=Boyd|first1=Stephen|last2=Vandenberghe|first2=Lieven|title=Convex Optimization|publisher=Cambridge University Press|location=Cambridge |year=2004|pages=244|isbn=0-521-83378-7|mr=2061575}}

Necessary conditions

Suppose that the objective function f\colon\mathbb{R}^n \rightarrow \mathbb{R} and the constraint functions g_i\colon\mathbb{R}^n \rightarrow \mathbb{R} and h_j\colon\mathbb{R}^n \rightarrow \mathbb{R} have subderivatives at a point x^* \in \mathbb{R}^n. If x^* is a local optimum and the optimization problem satisfies some regularity conditions (see below), then there exist constants \mu_i\ (i = 1,\ldots,m) and \lambda_j\ (j = 1,\ldots,\ell), called KKT multipliers, such that the following four groups of conditions hold:

{{cite book |last=Ruszczyński |first=Andrzej |title=Nonlinear Optimization |publisher=Princeton University Press |year=2006 |isbn=978-0691119151 |location=Princeton, NJ |mr=2199043 |author-link=Andrzej Piotr Ruszczyński}}

File:Inequality constraint diagram.svg

;Stationarity

:For minimizing f(x): \partial f(x^*) + \sum_{j=1}^\ell \lambda_j \partial h_j(x^*) + \sum_{i=1}^m \mu_i \partial g_i(x^*) \ni \mathbf 0

:For maximizing f(x): -\partial f(x^*) + \sum_{j=1}^\ell \lambda_j \partial h_j(x^*) + \sum_{i=1}^m \mu_i \partial g_i(x^*) \ni \mathbf 0

;Primal feasibility

:h_j(x^*) = 0, \text{ for } j = 1, \ldots, \ell \,\!

:g_i(x^*) \le 0, \text{ for } i = 1, \ldots, m

;Dual feasibility

:\mu_i \ge 0, \text{ for } i = 1, \ldots, m

;Complementary slackness

:\sum_{i=1}^m \mu_i g_i (x^*) = 0.

The last condition is sometimes written in the equivalent form: \mu_i g_i (x^*) = 0, \text{ for } i = 1, \ldots, m.

In the particular case m=0, i.e., when there are no inequality constraints, the KKT conditions turn into the Lagrange conditions, and the KKT multipliers are called Lagrange multipliers.

= Proof =

{{Math theorem

| math_statement =

(sufficiency) If there exists a solution x^* to the primal problem, a solution (\mu^*, \lambda^*) to the dual problem, such that together they satisfy the KKT conditions, then the problem pair has strong duality, and x^*, (\mu^*, \lambda^*) is a solution pair to the primal and dual problems.

(necessity) If the problem pair has strong duality, then for any solution x^* to the primal problem and any solution (\mu^*, \lambda^*) to the dual problem, the pair x^*, (\mu^*, \lambda^*) must satisfy the KKT conditions.{{Cite web |last=Geoff Gordon & Ryan Tibshirani |title=Karush-Kuhn-Tucker conditions, Optimization 10-725 / 36-725 |url=http://www.cs.cmu.edu/~ggordon/10725-F12/slides/16-kkt.pdf |archive-url=https://web.archive.org/web/20220617163414/http://www.cs.cmu.edu/~ggordon/10725-F12/slides/16-kkt.pdf |archive-date=2022-06-17}}

}}

{{Math proof|proof=

First, for the x^*, (\mu^*, \lambda^*) to satisfy the KKT conditions is equivalent to them being a Nash equilibrium.

Fix (\mu^*, \lambda^*), and vary x: equilibrium is equivalent to primal stationarity.

Fix x^*, and vary (\mu, \lambda): equilibrium is equivalent to primal feasibility and complementary slackness.

Sufficiency: the solution pair x^*, (\mu^*, \lambda^*) satisfies the KKT conditions, thus is a Nash equilibrium, and therefore closes the duality gap.

Necessity: any solution pair x^*, (\mu^*, \lambda^*) must close the duality gap, thus they must constitute a Nash equilibrium (since neither side could do any better), thus they satisfy the KKT conditions.

}}

= Interpretation: KKT conditions as balancing constraint-forces in state space =

The primal problem can be interpreted as moving a particle in the space of x, and subjecting it to three kinds of force fields:

  • f is a potential field that the particle is minimizing. The force generated by f is -\partial f.
  • g_i are one-sided constraint surfaces. The particle is allowed to move inside g_i \leq 0, but whenever it touches g_i=0, it is pushed inwards.
  • h_j are two-sided constraint surfaces. The particle is allowed to move only on the surface h_j.

Primal stationarity states that the "force" of \partial f(x^*) is exactly balanced by a linear sum of forces \partial h_j(x^*) and \partial g_i(x^*).

Dual feasibility additionally states that all the \partial g_i(x^*) forces must be one-sided, pointing inwards into the feasible set for x.

Complementary slackness states that if g_i(x^*) < 0, then the force coming from \partial g_i(x^*) must be zero i.e., \mu_i(x^*) = 0, since the particle is not on the boundary, the one-sided constraint force cannot activate.

= Matrix representation =

The necessary conditions can be written with Jacobian matrices of the constraint functions. Let \mathbf g(x) : \,\!\mathbb{R}^n \rightarrow \mathbb{R}^m be defined as \mathbf{g}(x) = \left( g_{1}(x), \ldots, g_{m}(x) \right)^\top and let \mathbf h(x) : \,\!\mathbb{R}^n \rightarrow \mathbb{R}^{\ell} be defined as \mathbf{h}(x) = \left( h_{1}(x), \ldots, h_{\ell}(x) \right)^\top. Let \boldsymbol{\mu} = \left( \mu_1, \ldots, \mu_m \right)^\top and \boldsymbol{\lambda} = \left( \lambda_1, \ldots, \lambda_{\ell} \right)^\top. Then the necessary conditions can be written as:

;Stationarity

:For maximizing f(x): \partial f(x^*) - D \mathbf g(x^*)^{\top} \boldsymbol{\mu} - D \mathbf h(x^*)^{\top} \boldsymbol{\lambda} = \mathbf 0

:For minimizing f(x): \partial f(x^*) + D \mathbf g(x^*)^{\top} \boldsymbol{\mu} + D \mathbf h(x^*)^{\top} \boldsymbol{\lambda} = \mathbf 0

;Primal feasibility

:\mathbf g(x^*) \le \mathbf 0

:\mathbf h(x^*) = \mathbf 0

;Dual feasibility

:\boldsymbol \mu \ge \mathbf 0

;Complementary slackness

:\boldsymbol \mu^{\top} \mathbf g (x^*) = 0.

Regularity conditions (or constraint qualifications)

One can ask whether a minimizer point x^* of the original, constrained optimization problem (assuming one exists) has to satisfy the above KKT conditions. This is similar to asking under what conditions the minimizer x^* of a function f(x) in an unconstrained problem has to satisfy the condition \nabla f(x^*)=0. For the constrained case, the situation is more complicated, and one can state a variety of (increasingly complicated) "regularity" conditions under which a constrained minimizer also satisfies the KKT conditions. Some common examples for conditions that guarantee this are tabulated in the following, with the LICQ the most frequently used one:

class="wikitable"

!Constraint

!Acronym

!Statement

Linearity constraint qualification

|LCQ

|If g_i and h_j are affine functions, then no other condition is needed.

Linear independence constraint qualification

|LICQ

|The gradients of the active inequality constraints and the gradients of the equality constraints are linearly independent at x^*.

Mangasarian-Fromovitz constraint qualification

|MFCQ

|The gradients of the equality constraints are linearly independent at x^* and there exists a vector d \in \mathbb{R}^n such that \nabla g_i(x^*)^\top d < 0 for all active inequality constraints and \nabla h_j(x^*)^\top d = 0 for all equality constraints.{{cite book|author=Dimitri Bertsekas|author-link=Dimitri Bertsekas|title=Nonlinear Programming|year=1999|publisher=Athena Scientific|edition=2|pages=329–330|isbn=9781886529007}}

Constant rank constraint qualification

|CRCQ

|For each subset of the gradients of the active inequality constraints and the gradients of the equality constraints the rank at a vicinity of x^* is constant.

Constant positive linear dependence constraint qualification

|CPLD

|For each subset of gradients of active inequality constraints and gradients of equality constraints, if the subset of vectors is linearly dependent at x^* with non-negative scalars associated with the inequality constraints, then it remains linearly dependent in a neighborhood of x^*.

Quasi-normality constraint qualification

|QNCQ

|If the gradients of the active inequality constraints and the gradients of the equality constraints are linearly dependent at x^* with associated multipliers \lambda_j for equalities and \mu_i\geq0 for inequalities, then there is no sequence x_k\to x^* such that \lambda_j \neq 0 \Rightarrow \lambda_j h_j(x_k)>0 and \mu_i \neq 0 \Rightarrow \mu_i g_i(x_k)>0.

Slater's condition

|SC

|For a convex problem (i.e., assuming minimization, f,g_i are convex and h_j is affine), there exists a point x such that h_j(x)=0 and g_i(x) < 0.

The strict implications can be shown

: LICQ ⇒ MFCQ ⇒ CPLD ⇒ QNCQ

and

: LICQ ⇒ CRCQ ⇒ CPLD ⇒ QNCQ

In practice weaker constraint qualifications are preferred since they apply to a broader selection of problems.

Sufficient conditions

In some cases, the necessary conditions are also sufficient for optimality. In general, the necessary conditions are not sufficient for optimality and additional information is required, such as the Second Order Sufficient Conditions (SOSC). For smooth functions, SOSC involve the second derivatives, which explains its name.

The necessary conditions are sufficient for optimality if the objective function f of a maximization problem is a differentiable concave function, the inequality constraints g_j are differentiable convex functions, the equality constraints h_i are affine functions, and Slater's condition holds.{{cite book|last1=Boyd|first1=Stephen|last2=Vandenberghe|first2=Lieven|title=Convex Optimization|publisher=Cambridge University Press|location=Cambridge |year=2004|pages=244|isbn=0-521-83378-7|mr=2061575}} Similarly, if the objective function f of a minimization problem is a differentiable convex function, the necessary conditions are also sufficient for optimality.

It was shown by Martin in 1985 that the broader class of functions in which KKT conditions guarantees global optimality are the so-called Type 1 invex functions.{{cite journal |first=D. H. |last=Martin |journal=J. Optim. Theory Appl. |volume=47 |issue=1 |pages=65–76 |title=The Essence of Invexity |year= 1985 |doi=10.1007/BF00941316 |s2cid=122906371 }}{{cite journal |first=M. A. |last=Hanson |title=Invexity and the Kuhn-Tucker Theorem |journal=J. Math. Anal. Appl. |volume=236 |issue=2 |pages=594–604 |year=1999 |doi=10.1006/jmaa.1999.6484 |doi-access=free }}

= Second-order sufficient conditions =

For smooth, non-linear optimization problems, a second order sufficient condition is given as follows.

The solution x^*, \lambda^*, \mu^* found in the above section is a constrained local minimum if for the Lagrangian,

: L(x,\lambda,\mu) = f(x) + \sum_{i=1}^m \mu_i g_i(x) + \sum_{j=1}^\ell \lambda_j h_j(x)

then,

:s^T\nabla ^2_{xx}L(x^*,\lambda^*,\mu^*)s \ge 0

where s \ne 0 is a vector satisfying the following,

:\left[ \nabla_x g_i(x^*), \nabla_x h_j(x^*) \right]^T s = 0_{\mathbb{R}^{2}}

where only those active inequality constraints g_i(x) corresponding to strict complementarity (i.e. where \mu_i > 0) are applied. The solution is a strict constrained local minimum in the case the inequality is also strict.

If s^T\nabla ^2_{xx}L(x^*,\lambda^*,\mu^*)s = 0, the third order Taylor expansion of the Lagrangian should be used to verify if x^* is a local minimum. The minimization of f(x_1,x_2)=(x_2-x_1^2)(x_2-3x_1^2) is a good counter-example, see also Peano surface.

Economics

{{see also|Profit maximization}}

Often in mathematical economics the KKT approach is used in theoretical models in order to obtain qualitative results. For example,Chiang, Alpha C. Fundamental Methods of Mathematical Economics, 3rd edition, 1984, pp. 750–752. consider a firm that maximizes its sales revenue subject to a minimum profit constraint. Letting Q be the quantity of output produced (to be chosen), R(Q) be sales revenue with a positive first derivative and with a zero value at zero output, C(Q) be production costs with a positive first derivative and with a non-negative value at zero output, and G_{\min} be the positive minimal acceptable level of profit, then the problem is a meaningful one if the revenue function levels off so it eventually is less steep than the cost function. The problem expressed in the previously given minimization form is

:Minimize -R(Q)

:subject to

: G_{\min} \le R(Q) - C(Q)

: Q \ge 0,

and the KKT conditions are

:

\begin{align}

& \left(\frac{\text{d} R}{\text{d} Q} \right) (1+\mu ) - \mu \left( \frac{\text{d} C}{\text{d} Q} \right) \le 0, \\[5pt]

& Q \ge 0, \\[5pt]

& Q \left[ \left( \frac{\text{d} R}{\text{d} Q} \right) (1+\mu ) - \mu \left( \frac{\text{d} C}{\text{d} Q}\right) \right] = 0, \\[5pt]

& R(Q) - C(Q) - G_{\min} \ge 0, \\[5pt]

& \mu \ge 0, \\[5pt]

& \mu [R(Q) - C(Q) - G_{\min}] = 0.

\end{align}

Since Q = 0 would violate the minimum profit constraint, we have Q > 0 and hence the third condition implies that the first condition holds with equality. Solving that equality gives

: \frac{\text{d} R}{\text{d} Q} = \frac{\mu}{1+ \mu} \left( \frac{\text{d} C}{\text{d} Q} \right).

Because it was given that \text{d} R / \text{d} Q and \text{d} C / \text{d} Q are strictly positive, this inequality along with the non-negativity condition on \mu guarantees that \mu is positive and so the revenue-maximizing firm operates at a level of output at which marginal revenue \text{d} R / \text{d} Q is less than marginal cost \text{d} C / \text{d} Q — a result that is of interest because it contrasts with the behavior of a profit maximizing firm, which operates at a level at which they are equal.

Value function

If we reconsider the optimization problem as a maximization problem with constant inequality constraints:

: \text{Maximize }\; f(x)

: \text{subject to }\

:: g_i(x) \le a_i , h_j(x) = 0.

The value function is defined as

:V(a_1, \ldots, a_n) = \sup\limits_x f(x)

: \text{subject to }\

:: g_i(x) \le a_i , h_j(x) = 0

:: j \in \{1,\ldots, \ell\}, i\in\{1,\ldots,m\},

so the domain of V is \{a \in \mathbb{R}^m \mid \text{for some }x\in X, g_i(x) \leq a_i, i \in \{1,\ldots,m\}\}.

Given this definition, each coefficient \mu_i is the rate at which the value function increases as a_i increases. Thus if each a_i is interpreted as a resource constraint, the coefficients tell you how much increasing a resource will increase the optimum value of our function f. This interpretation is especially important in economics and is used, for instance, in utility maximization problems.

Generalizations

With an extra multiplier \mu_0\geq0, which may be zero (as long as (\mu_0,\mu,\lambda)\neq0), in front of \nabla f(x^*) the KKT stationarity conditions turn into

:

\begin{align}

& \mu_0 \,\nabla f(x^*) + \sum_{i=1}^m \mu_i \,\nabla g_i(x^*) + \sum_{j=1}^\ell \lambda_j \,\nabla h_j(x^*) = 0, \\[4pt]

& \mu_jg_i(x^*)=0, \quad i=1,\dots,m,

\end{align}

which are called the Fritz John conditions. This optimality conditions holds without constraint qualifications and it is equivalent to the optimality condition KKT or (not-MFCQ).

The KKT conditions belong to a wider class of the first-order necessary conditions (FONC), which allow for non-smooth functions using subderivatives.

See also

References

{{Reflist}}

Further reading

  • {{cite journal |first1=R. |last1=Andreani |first2=J. M. |last2=Martínez |first3=M. L. |last3=Schuverdt |title=On the relation between constant positive linear dependence condition and quasinormality constraint qualification |journal=Journal of Optimization Theory and Applications |volume=125 |issue=2 |pages=473–485 |year=2005 |doi=10.1007/s10957-004-1861-9 |s2cid=122212394 }}
  • {{cite book |last=Avriel |first=Mordecai |year=2003 |title=Nonlinear Programming: Analysis and Methods |publisher=Dover |isbn=0-486-43227-0 }}
  • {{cite book |first1=V. |last1=Boltyanski |first2=H. |last2=Martini |first3=V. |last3=Soltan |title=Geometric Methods and Optimization Problems |location=New York |publisher=Springer |year=1998 |isbn=0-7923-5454-0 |pages=78–92 |chapter=The Kuhn–Tucker Theorem |chapter-url=https://books.google.com/books?id=YD7UBwAAQBAJ&pg=PA78 }}
  • {{cite book |first1=S. |last1=Boyd |first2=L. |last2=Vandenberghe |year=2004 |chapter=Optimality Conditions |chapter-url=https://web.stanford.edu/~boyd/cvxbook/bv_cvxbook.pdf#page=255 |title=Convex Optimization |publisher=Cambridge University Press |isbn=0-521-83378-7 |pages=241–249 }}
  • {{cite book |first1=Murray C. |last1=Kemp |first2=Yoshio |last2=Kimura |title=Introduction to Mathematical Economics |location=New York |publisher=Springer |year=1978 |isbn=0-387-90304-6 |pages=[https://archive.org/details/introductiontoma0000kemp/page/38 38–73] |url=https://archive.org/details/introductiontoma0000kemp/page/38 }}
  • {{cite book |first=Nicholas |last=Rau |chapter=Lagrange Multipliers |title=Matrices and Mathematical Programming |location=London |publisher=Macmillan |year=1981 |isbn=0-333-27768-6 |pages=156–174 }}
  • {{cite book |first1=J. |last1=Nocedal |first2=S. J. |last2=Wright |year=2006 |title=Numerical Optimization |publisher=Springer |location=New York |isbn=978-0-387-30303-1 }}
  • {{cite book |first=Rangarajan K. |last=Sundaram |chapter=Inequality Constraints and the Theorem of Kuhn and Tucker |title=A First Course in Optimization Theory |location=New York |publisher=Cambridge University Press |year=1996 |isbn=0-521-49770-1 |pages=145–171 |chapter-url=https://books.google.com/books?id=yAfug81P-8YC&pg=PA145 }}