The penalty method is a technique for solving constrained optimization problems by converting them into unconstrained problems. It works by adding penalty terms to the objective function that impose a high cost for violating the constraints.
Mathematical Formulation
Consider a constrained optimization problem:
\begin{align} \min_{\theta} \quad & L(\theta) \ \text{subject to} \quad & h_j(\theta) = 0, \quad j = 1,2,\ldots,m \ & I_k(\theta) \geq 0, \quad k = 1,2,\ldots,n \end{align}
where:
- is the design vector
- is the objective function to be minimized
- are the equality constraint functions
- are the inequality constraint functions
The penalty method transforms this into an unconstrained problem:
where:
- is the penalty parameter
- The term penalizes violations of equality constraints
- The term penalizes violations of inequality constraints
Solution Approach
-
Initialization:
- Select an initial design point (not necessarily feasible)
- Choose an initial penalty parameter
- Set a parameter for increasing the penalty parameter
- Set iteration counter
-
Iteration:
- Solve the unconstrained optimization problem:
- Evaluate constraint violations
- If constraints are satisfied to desired tolerance, terminate
- Otherwise, update penalty parameter:
- Increment and repeat
-
Convergence:
- As , the solution approaches the constrained optimum
Key Properties
Advantages
- Simplicity: Transforms constrained problems into unconstrained problems
- Generality: Handles both equality and inequality constraints
- Initial Point: Can start from infeasible points
- Implementation: Reuses unconstrained optimization algorithms
Disadvantages
- Ill-Conditioning: As increases, the problem becomes increasingly ill-conditioned
- Inexact Constraint Satisfaction: Constraints are only approximately satisfied
- Slow Convergence: Requires solving multiple unconstrained problems with increasing penalty parameters
- Parameter Selection: Choosing good values for and can be challenging
Variants of Penalty Methods
Exterior Penalty Method
The standard formulation described above is known as the exterior penalty method because it allows iterates to be outside the feasible region. The solution approaches the feasible region from the exterior as increases.
Interior Penalty Method (Barrier Method)
The interior penalty method adds a barrier term that prevents solutions from leaving the feasible region:
This method requires a strictly feasible initial point. As , the solution approaches the constrained optimum.
Exact Penalty Method
The exact penalty method uses non-differentiable penalty functions such as L1-norm:
For sufficiently large but finite , this method can yield the exact solution to the constrained problem.
Implementation Considerations
Penalty Parameter Update
The rate at which is increased affects both convergence speed and numerical stability:
- Small (e.g., ): Slower convergence but better numerical stability
- Large (e.g., ): Faster convergence but potential numerical issues
Unconstrained Optimization Algorithm
The choice of algorithm for solving the unconstrained subproblems depends on problem characteristics:
- Gradient-Based Methods: BFGS, conjugate gradient, or steepest descent
- Derivative-Free Methods: Nelder-Mead simplex or pattern search for non-smooth penalties
Stopping Criteria
Common termination conditions include:
- Maximum constraint violation below tolerance
- Change in objective function or solution vector below tolerance
- Maximum number of iterations or function evaluations reached
Example: Penalty Method Application
Consider the problem:
\begin{align} \min_{\theta_1, \theta_2} \quad & \theta_1^2 + \theta_2^2 \ \text{subject to} \quad & \theta_1 + \theta_2 = 1 \end{align}
Using the penalty method, we form:
For any finite , the optimum of will be close to but not exactly on the constraint . As increases, the solution approaches the constrained optimum .
Numerical Issues and Remedies
Scaling
Poor scaling of constraints can lead to numerical difficulties. Normalizing constraints to similar magnitudes helps balance their contributions to the penalty term.
Gradual Penalty Increase
Starting with a moderate penalty parameter and gradually increasing it allows the algorithm to adapt to the constrained landscape.
Constraint Prioritization
For problems with multiple constraints, applying different penalty weights to different constraints can improve solution quality.
Theoretical Connections
The penalty method is related to several other optimization concepts:
- Lagrange Multipliers: As , the ratio of penalty term gradients to approaches the optimal Lagrange multipliers
- Duality Theory: The penalty function provides an upper bound on the optimal dual function value
- Merit Functions: Penalty functions serve as merit functions in line search methods for constrained optimization
Engineering Applications
The penalty method is widely used in engineering for:
- Structural Optimization: Minimizing weight subject to stress and displacement constraints
- Parameter Estimation: Fitting models to data with physical constraints
- Path Planning: Finding optimal trajectories subject to kinematic constraints
- Machine Learning: Regularizing models with equality and inequality constraints