Smoothness of the Augmented Lagrangian Dual in Convex Optimization
Abstract: This paper investigates the general linearly constrained optimization problem: $\min_{x \in \Rd} f(x) \ \st \ A x = b$, where $f: \Rn \rightarrow \exs$ is a closed proper convex function, $A \in \R{p \times d}$, and $b \in \Rp$. We establish the following results without requiring additional regularity conditions: (1) the augmented Lagrangian dual function $\phi_{\rho}(\lambda) = \inf_x \cL_{\rho}(x, \lambda)$ is $\frac{1}{\rho}$-smooth everywhere; and (2) the solution to $\min_{x \in \Rd} \cL_{\rho}(x, \lambda)$ exists for any dual variable $\lambda \in \Rp$, where $\rho > 0$ is the augmented parameter and $\cL_{\rho}(x, \lambda) = f(x) + \dotprod{\lambda, A x - b} + \frac{\rho}{2}\norm{A x - b}2$ is the augmented Lagrangian. These findings significantly relax the strong assumptions commonly imposed in existing literature to guarantee similar properties.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.