The Augmented Lagrangian Method (ALM), also known as the method of multipliers, combines the advantages of penalty methods and Lagrangian methods to solve constrained optimization problems. It addresses the ill-conditioning issues of penalty methods while maintaining their computational advantages.
Mathematical Formulation
Consider a constrained optimization problem:
\begin{align} \min_{\theta} \quad & L(\theta) \ \text{subject to} \quad & h_j(\theta) = 0, \quad j = 1,2,\ldots,m \ & I_k(\theta) \geq 0, \quad k = 1,2,\ldots,n \end{align}
where:
- is the design vector
- is the objective function to be minimized
- are the equality constraint functions
- are the inequality constraint functions
The Augmented Lagrangian function for this problem is:
where:
- are the Lagrange multipliers for equality constraints
- are the Lagrange multipliers for inequality constraints
- is the penalty parameter
Solution Approach
-
Initialization:
- Select an initial design point
- Choose initial Lagrange multiplier estimates and
- Set an initial penalty parameter
- Set iteration counter
-
Iteration:
- Solve the unconstrained subproblem:
- Update the Lagrange multipliers:
- Check convergence based on constraint violations and multiplier changes
- If not converged, update penalty parameter: (if needed)
- Increment and repeat
-
Convergence:
- The method converges to a KKT point of the original problem
Key Properties
Advantages
- Improved Conditioning: Avoids the severe ill-conditioning of pure penalty methods
- Faster Convergence: Typically requires fewer iterations and lower penalty parameters than penalty methods
- Exact Constraint Satisfaction: Can achieve precise constraint satisfaction with finite penalty parameters
- Dual Information: Provides estimates of Lagrange multipliers throughout the optimization process
Disadvantages
- Complexity: More complex than pure penalty or Lagrangian methods
- Parameter Tuning: Requires careful selection of penalty update parameters
- Subproblem Difficulty: Each unconstrained subproblem may be more challenging to solve
- Non-Convexity: For non-convex problems, convergence to global optima is not guaranteed
Method Variations
Method of Multipliers (for Equality Constraints)
For problems with only equality constraints, the Augmented Lagrangian simplifies to:
The multiplier update becomes:
ADMM (Alternating Direction Method of Multipliers)
A specialized version of ALM for problems with separable objective functions and linear constraints:
\begin{align} \min_{\theta_1, \theta_2} \quad & L_1(\theta_1) + L_2(\theta_2) \ \text{subject to} \quad & A\theta_1 + B\theta_2 = c \end{align}
ADMM alternates between optimizing for and before updating multipliers.
Separable Augmented Lagrangian
For problems with block structure, the method can be adapted to update different blocks of variables sequentially:
Implementation Considerations
Unconstrained Subproblem Solver
The choice of algorithm for solving the unconstrained subproblems affects performance:
- Quasi-Newton Methods: BFGS or L-BFGS for smooth problems
- Conjugate Gradient: For large-scale problems with good conditioning
- Hessian-Free Newton: When second derivatives are available
Multiplier Updates
Multiplier updates can be damped to improve stability:
where is a damping factor.
Penalty Parameter Updates
The penalty parameter is typically increased only when constraint violations don’t decrease sufficiently:
- If , where , then
- Otherwise, keep
Convergence Criteria
Common stopping conditions include:
- Constraint violations below tolerance
- KKT conditions satisfied to desired accuracy
- Change in Lagrange multipliers below tolerance
- Maximum iterations reached
Theoretical Foundations
Connection to KKT Conditions
The fixed points of the ALM iteration satisfy the KKT conditions for the original constrained problem:
- Stationarity:
- Primal Feasibility: and
- Dual Feasibility:
- Complementary Slackness:
Local vs. Global Convergence
- For convex problems, ALM converges to global optima
- For non-convex problems, ALM converges to local optima or stationary points
- The quality of the final solution depends on the initial point
Example: Augmented Lagrangian Application
Consider the inequality-constrained problem:
\begin{align} \min_{\theta_1, \theta_2} \quad & \theta_1^2 + \theta_2^2 \ \text{subject to} \quad & \theta_1 + \theta_2 \geq 1 \end{align}
We rewrite the constraint as . The Augmented Lagrangian is:
For a given and , we solve for , then update:
As the method progresses, converges to and to .
Practical Applications
Structural Optimization
In structural design problems, ALM effectively handles:
- Stress constraints:
- Displacement constraints:
- Natural frequency constraints:
Optimal Control
For trajectory optimization with path constraints:
- Dynamic constraints:
- Control limits:
- State constraints:
Machine Learning
In constrained learning problems:
- Fairness constraints:
- Sparsity constraints:
- Budget constraints:
Numerical Considerations
Scaling
Proper scaling of constraints improves numerical behavior:
- Normalize constraints to have similar magnitudes
- Scale objective function to be comparable to constraint penalty terms
Starting Point
The initial point affects convergence:
- Starting from a feasible point may accelerate convergence
- Multiple starting points can be used to find better local optima
Precision
Balancing precision requirements:
- Tighten subproblem tolerances as outer iterations progress
- Use adaptive precision based on constraint violation magnitudes
Software Implementations
Several optimization libraries implement Augmented Lagrangian methods:
- LANCELOT and ALGENCAN for large-scale constrained optimization
- IPOPT with augmented Lagrangian preconditioners
- SciPy’s
scipy.optimize.minimizewith method=‘trust-constr’ - MATLAB’s Optimization Toolbox with
fminconalgorithm=‘sqp’
Related Pages
- Penalty Method
- Lagrange Multipliers
- KKT Conditions
- Constrained Optimization
- Equality Constraints
- Inequality Constraints
Augmented Lagrangian Method
The Augmented Lagrangian Method (ALM), also known as the method of multipliers, combines the Lagrange multipliers approach with penalty terms.
Mathematical Formulation
For the constrained optimization problem:
The Augmented Lagrangian function is:
Method of Operation
Instead of solving for optimal directly, it is updated in an “outer” iterative loop:
Advantages
- Combines benefits of both Lagrange multipliers and penalty methods
- Improves convergence properties compared to simple penalty methods
Disadvantages
- More complex to implement due to two nested loops