Processing math: 100%

Constrained Optimization

Introduction

Constrained Optimization Methods

Problem

Minimize f(x), xRn, subject to gj(x)=0, j=1,,m (equality constraints) or gj(x)0 (inequality constraints).

  • Direct Substitution: Solve gj(x)=0 for some variables, substitute into f(x). Simple but limited to explicit solutions.
  • Constrained Variation: Optimize along admissible directions satisfying constraints. Complex for nonlinear constraints.
  • Lagrange Multipliers: Form L(x,λ)=f(x)+λjgj(x), solve xL=0, gj(x)=0. General, extends to inequalities via KKT.

Direct Substitution Method

Approach

  • Solve: Constraint g(x)=0 for one variable.
  • Substitute: Into f(x) to eliminate the constraint.
  • Optimize: The reduced unconstrained problem.

Example

Minimize f(x,y)=x2+y2 subject to x+y=1:

  • Substitute: y=1x into f.
  • Solve: ddx(2x22x+1)=0x=0.5.
  • Optimal point: (0.5,0.5).

Limitations

  • Fails: For complex constraints (e.g., x+ey=1) or multiple constraints.

Constrained Variation Method

Approach

  • Parameterize: Variables to satisfy g(x)=0.
  • Compute: Admissible variations δx (tangent to constraint).
  • Ensure: f(x) has no component along admissible directions.

Example

Minimize f(x,y)=x2+y2 subject to x+y=1:

  • Parameterize: x=t, y=1t.
  • Optimize: f(t)=2t22t+1.
  • Solution: Same as Direct Substitution, x=0.5,y=0.5.

Limitations

  • Challenges: Manual parameterization; harder for nonlinear/multiple constraints.

Summary: Direct vs. Constrained Methods

Direct Substitution

  • Solves: Constraints explicitly.
  • Reduces: Problem dimensionality.
  • Limited: To simple constraints.

Constrained Variation

  • Works: Along constraint surfaces.
  • Geometric: Interpretation.
  • Harder: For multiple constraints.

General Rule

Use Direct Substitution when possible; resort to Constrained Variation for complex constraints.

Lagrange Multiplier Method

Problem Formulation

Minimize f(x), xRn, subject to gj(x)=0, j=1,,m.

  • Assumption: f and gj are differentiable, constraints satisfy regularity (e.g., gj linearly independent).

Lagrangian Function

L(x,λ)=f(x)+mj=1λjgj(x)

where λ=(λ1,,λm) are Lagrange multipliers.

Theorem: Necessary Conditions

At a critical point:

xL=Lxi=0(i=1,,n)Lλj=gj(x)=0(j=1,,m)

Geometric Intuition

At the optimum, f=mj=1λjgj, meaning gradient alignment of f with constraint gradients.

Interpretation of Lagrange Multipliers

  • Sensitivity: λj measures how f changes with constraint perturbations.
  • Shadow price: For gj(x)=bj, λj=fbj.
  • Physical: Represents constraint forces.
  • Sign:
    • λj>0: Tightening increases f;
    • λj<0: Tightening decreases f;
    • λj=0: Inactive constraint.

Detailed Example: Lagrange Multipliers

Example

Minimize f(x,y)=x2+y2 subject to g(x,y)=x+y1=0.

Solution

  • Lagrangian: L(x,y,λ)=x2+y2+λ(x+y1).
  • Conditions:

    Lx=2x+λ=0Ly=2y+λ=0Lλ=x+y1=0

  • Solution: x=y=12, λ=1.
  • Verification: f(12,12)=14. Hessian positive definite, confirming a minimum.
  • Interpretation: Minimum distance from origin to line x+y=1 is 22 at (12,12).

Inequality Constraints

Problem Formulation

Minimize f(x), xRn, subject to gj(x)0, j=1,,m.

Karush-Kuhn-Tucker (KKT) Conditions

Form Lagrangian: L(x,λ)=f(x)+mj=1λjgj(x).

Necessary conditions:

  • Stationarity: xL=0.
  • Primal feasibility: gj(x)0.
  • Complementary slackness: λjgj(x)=0.
  • Dual feasibility: λj0.

Example

Minimize f(x,y)=x2+y2 subject to x+y1.

  • Solution: Check KKT conditions.
  • Inactive: If x+y<1, λ=0, minimum at (0,0), f=0.
  • Active: If x+y=1, x=y=12, f=14.
  • Conclusion: Since f(0,0)<f(12,12), minimum is at (0,0).

KKT Conditions Interpretation

  • Active constraints: gj(x)=0, λj>0 (binding).
  • Inactive constraints: gj(x)<0, λj=0 (non-binding).
  • Constraint qualification: E.g., LICQ: Gradients of active constraints are linearly independent.
  • Notes: KKT necessary under qualification; sufficient for convex f and gj.

KKT Example Problem

Example

Minimize f(x,y)=(x1)2+(y2)2 subject to:

x+y2yx

Solution

  • Rewrite: g1=x+y20, g2=xy0.
  • KKT conditions:

    2(x1)+λ1+λ2=02(y2)+λ1λ2=0λ1(x+y2)=0λ2(xy)=0λ1,λ20

  • Solve: Assume λ1>0, λ2=0. Then x+y=2, yielding x=0.5, y=1.5, λ1=1. Verify: yx, λ10.
  • Solution: x=0.5, y=1.5, λ1=1, λ2=0.

Summary