Penalty methods are a certain class of algorithms for solving constrained optimization problems.
A penalty method replaces a constrained optimization problem by a series of unconstrained problems whose solutions ideally converge to the solution of the original constrained problem. The unconstrained problems are formed by adding a term, called a penalty function, to the objective function that consists of a penalty parameter multiplied by a measure of violation of the constraints. The measure of violation is nonzero when the constraints are violated and is zero in the region where constraints are not violated.
Let us say we are solving the following constrained problem:
minxf(x)
ci(x)\le0~\foralli\inI.
This problem can be solved as a series of unconstrained minimization problems
minfp(x):=f(x)+p~\sumi\in~g(ci(x))
g(ci(x))=max(0,c
2. | |
i(x)) |
In the above equations,
g(ci(x))
p
p
Common penalty functions in constrained optimization are the quadratic penalty function and the deadzone-linear penalty function[1] .
We first consider the set of global optimizers of the original problem, X*.[2] Assume that the objective f has bounded level sets, and that the original problem is feasible. Then:
This theorem is helpful mostly when fp is convex, since in this case, we can find the global optimizers of fp.
A second theorem considers local optimizers. Let x* be a non-degenerate local optimizer of the original problem ("nondegenerate" means that the gradients of the active constraints are linearly independent and the second-order sufficient optimality condition is satisfied). Then, there exists a neighborhood V* of x*, and some p0>0, such that for all p>p0, the penalized objective fp has exactly one critical point in V* (denoted by x*(p)), and x*(p) approaches x* as p→∞. Also, the objective value f(x*(p)) is weakly-increasing with p.
Image compression optimization algorithms can make use of penalty functions for selecting how best to compress zones of colour to single representative values.[3] [4] The penalty method is often used in computational mechanics, especially in the Finite element method, to enforce conditions such as e.g. contact.
The advantage of the penalty method is that, once we have a penalized objective with no constraints, we can use any unconstrained optimization method to solve it. The disadvantage is that, as the penalty coefficient p grows, the unconstrained problem becomes ill-conditioned - the coefficients are very large, and this may cause numeric errors and slow convergence of the unconstrained minimization.
Barrier methods constitute an alternative class of algorithms for constrained optimization. These methods also add a penalty-like term to the objective function, but in this case the iterates are forced to remain interior to the feasible domain and the barrier is in place to bias the iterates to remain away from the boundary of the feasible region. They are practically more efficient than penalty methods.
Augmented Lagrangian methods are alternative penalty methods, which allow to get high-accuracy solutions without pushing the penalty coefficient to infinity. This makes the unconstrained penalized problems easier to solve.
Other nonlinear programming algorithms:
Coello, A.C.https://www.researchgate.net/publication/222857595_Coello_AC_Theoretical_and_Numerical_Constraint-Handling_Techniques_Used_with_Evolutionary_Algorithms_A_Survey_of_the_State_of_the_Art_Comput_Methods_Appl_Mech_Engrg_19111-12_1245-1287: Theoretical and Numerical Constraint-Handling Techniques Used with Evolutionary Algorithms: A Survey of the State of the Art. Comput. Methods Appl. Mech. Engrg. 191(11-12), 1245-1287
Courant, R. Variational methods for the solution of problems of equilibrium and vibrations. Bull. Amer. Math. Soc., 49, 1 - 23, 1943.
Wotao, Y. Optimization Algorithms for constrained optimization. Department of Mathematics, UCLA, 2015.