Skip to content

PHR Conic Augmented Lagrangian Method

Slide: Starting from the penalty method, we extend to the augmented Lagrangian method for improved stability. By introducing s, symmetric cone constraints are integrated, forming a unified framework for solving constrained optimization problems iteratively. Inspired by Dr. Zhepei Wang's Lecture "Numerical Optimization for Robotics".

Introduction

Penalty Method

Consider the constrained optimization problem:

minxf(x)s.t.h(x)=0

Penalty method solves a series of unconstrained problems:

Qρ(x)=f(x)+ρ2||h(x)||2

Challenge:

  • Requires ρ for exact solution, causing ill-conditioned Hessian.
  • Finite ρ leads to constraint violation h(x)0.

Lagrangian Relaxation

The Lagrangian is defined as:

L(x,λ)=f(x)+λh(x)

At the optimal solution x, there exists λ such that xL(x,λ)=0.

Uzawa's method iteratively updates x and λ:

{xk+1=argminxL(x,λk)λk+1=λk+αkh(xk+1)

where αk>0 is a step size.

However, when λ is fixed and one attempts to minimize L(x,λ):

  • minxL(x,λ) can be non-smooth even for smooth f and h.
  • minxL(x,λ) may be unbounded or have no finite solution.

Equality Constraint

PHR Augmented Lagrangian Method

Consider the optimization problem with a penalty on the deviation from a prior λ¯:

minxmaxλf(x)+λh(x)12ρ||λλ¯||2

The inner problem:

λ=h(x)1ρ(λλ¯)λ(λ¯)=λ¯+ρh(x)

The outer problem:

minxmaxλf(x)+λh(x)12ρ||λλ¯||2=minxf(x)+[λ(λ¯)]h(x)12ρ||λ(λ¯)λ¯||2=minxf(x)+[λ¯+ρh(x)]h(x)ρ2||h(x)||2=minxf(x)+λ¯h(x)+ρ2||h(x)||2

PHR Augmented Lagrangian Method (cont.)

To increase precision:

  • Reduce the penalty weight 1/ρ
  • Update the prior multiplier λ¯λ(λ¯)

Uzawa's method for the augmented Lagrangian function is:

  1. xargminxf(x)+λ¯h(x)+ρ2||h(x)||2
  2. λ¯λ¯+ρh(x)

Penalty Method Perspective

The corresponding primal problem of the augumented Lagrangian Function is obviously:

minx f(x)+ρ2||h(x)||2s.t. h(x)=0

Advantages:

  • Even without ρ, the constraints can be exactly satisfied in the limit through multiplier updates.
  • For large ρ, the penalty term ρ2||h(x)||2 dominates, ensuring minxLρ(x,λ) has a local solution.
  • The augmented dual function qρ(λ) is smooth in proper conditions, with qρ(λ)h(x(λ)).

Practical PHR-ALM

In practical, we use its equivalent form:

Lρ(x,λ)=f(x)+ρ2h(x)+λρ212ρ||λ||2x-independent

The KKT solution can be solved via:

{xk+1=argminxLρk(x,λk)λk+1=λk+ρkh(xk+1)ρk+1=min[(1+γ)ρk,ρmax]

where ρk can be any nondecreasing positive sequence.

Inequality Constraint

Slack Variables Relaxation

Consider the optimization problem with inequality constraints:

minxf(x)s.t.g(x)0

We use the equivalent formulation using slack variables:

minx,sf(x)s.t.g(x)+[s]2=0

where []2 means element-wise squaring.

We can directly form Lagrangian like equality-constrained case

minx,s{f(x)+ρ2g(x)+[s]2+λρ2}=minxf(x)+minxminsρ2g(x)+[s]2+λρ2=minxf(x)+ρ2max[g(x)+λρ,0]2

Simplified Form

Summing over all components gives the final form:

Lρ(x,μ)=f(x)+ρ2max[g(x)+μρ,0]212ρ||μ||2x-independent

For the dual update, from the optimality condition:

μk+1=μk+ρ(g(xk+1)+[sk+1]2)=max[μk+ρg(xk+1),0]

Summary

PHR Augmented Lagrangian Method for General Nonconvex cases:

minxf(x)s.t. h(x)=0g(x)0

Its PHR Augmented Lagrangian is defined as

Lρ=f(x)+ρ2h(x)+λρ2+ρ2max[g(x)+μρ,0]212ρ{||λ||2+||μ||2}

The PHR-ALM is simply repeating the primal descent and dual ascent iterations:

{xk+1=argminxLρk(x,λk,μk)λk+1=λk+ρkh(xk+1)μk+1=max[μk+ρkg(xk+1),0]ρk+1=min[(1+γ)ρk,ρmax]

Courtesy: Z. Wang

Symmetric Cone Constraint

Extension to Symmetric Cone Constraints

Consider the symmetric cone constrained optimization problem:

minxf(x)s.t. h(x)=0g(x)K
Second Order ConePositive Definite Cone

Generalized Inequality Constraint

For the symmetric cone constraint xK, we can equivalently express it as:

g(x)=xK0

The standard inequality constraint g(x)0 corresponds to the nonnegative orthant cone:

K=R+n={xRn:xi0,i=1,,n}

t's projection operator is exactly element-wise max function:

ΠR+n(v)=max[v,0],(R+n)=R+n

Slack Variables Relaxation

Consider the optimization problem with inequality constraints:

minxf(x)s.t.g(x)K

By Euclidean Jordan algebra, the conit program is equivalent to

minx,sf(x)s.t.g(x)=ss

We can directly form Lagrangian like equality-constrained case

minx,s{f(x)+ρ2g(x)ss+λρ2}=minxf(x)+minxminsρ2g(x)ss+λρ2=minxf(x)+ρ2ΠK(g(x)λρ)2

Simplified Form

Let μ=λ, we get the final form:

Lρ(x,μ)=f(x)+ρ2ΠK(g(x)+μρ)212ρ||μ||2x-independent

For the dual update, from the optimality condition:

μk+1=μk+ρ[g(xk+1)sk+1sk+1]=ΠK[μkρg(xk+1)]

where K is the dual cone of K.

Summary

PHR Augmented Lagrangian Method for General Nonconvex cases:

minxf(x)s.t. h(x)=0g(x)K

Its PHR Augmented Lagrangian is defined as

Lρ=f(x)+ρ2h(x)+λρ2+ρ2ΠK(g(x)+μρ)212ρ{||λ||2+||μ||2}

The PHR-ALM is simply repeating the primal descent and dual ascent iterations:

{xk+1=argminxLρk(x,λk,μk)λk+1=λk+ρkh(xk+1)μk+1=ΠK(μkρkxk+1)ρk+1=min[(1+γ)ρk,ρmax]