Augmented Lagrangian amd Method of Multipliers (ALM)
The dual ascent requires conditions to ensure convergence. The limitation of dual ascent is solved by augmented Lagrangian method, also called method of multipliers.
This method is to robust dual ascent.
- Convergence without assumptions like strict convexity or finiteness of $f$.
Consider the augmented Lagrangian function,
where $\rho > 0$ is the penalty parameter.
The augmented Lagrangian can be regarded as the unaugmented Lagrangian associated with the constraints,
Clearly equivalent to original problem. Hence the both equation are same, because for any feasible $x$, the term $\frac{\rho}{2} \left || Ax-b \right ||^{2}_{2}$ becomes zero.
The associated dual function is as follows,
Then, adding term $\frac{\rho}{2} \left || Ax-b \right ||^{2}_{2}$
to $f\left ( x \right )$ makes $g_{\rho}(y)$ differentiable under rather mild conditions than on the original problem.
Applying dual ascent to the modified algorithm is as follows,
Now here’s a question. Why is the convergence of ALM better than duan ascent?
Since $x^{k}$ minimizes
over $x$, we have,
This is the stationarity condition for original primal problem under mild condition
as
, so Karush–Kuhn–Tucker (KKT) conditions are satisfied in the limit and $x^{k}$, $u^{k}$ converge to solutions.
The advantage of ALM is that much better convergence properties. However, the disadvantage of ALM is the loss of decomposability.
There’s always a trade-off