A comprehensive guide to the theory of Lagrangean Relaxation and Decomposition, with deep-dive applications in process systems engineering.
Minimize: \(f_1(x_1) + f_2(x_2)\)
Subject to: \(x_1 \in C_1, \quad x_2 \in C_2\)
Actially cvx is very good at finding and exploiting the separable structure.
Consider the unconstrained problem:
Minimize: \(f(x) = f_1(x_1, y) + f_2(x_2, y)\)
where \(x = (x_1, x_2, y)\)
Fix $y$ and define:
Subproblem 1: \(\text{Minimize}_{x_1} \quad f_1(x_1, y)\)
Subproblem 2: \(\text{Minimize}_{x_2} \quad f_2(x_2, y)\)
Minimize: \(\phi_1(y) + \phi_2(y)\)
with variable $y$.
(Using subgradient algorithm for master)
Repeat:
Note: The step length $\alpha_k$ can be chosen in any of the standard ways.
in the linear algebra level, sturctuer can be exploted. how the hessian look like for complicating variable case?
block elimination shur complement block diagonal submatrix
Step 1: introduce new variables $y_1$, $y_2$
\[\begin{aligned} \text{minimize} \quad & f(x) = f_1(x_1, y_1) + f_2(x_2, y_2) \\ \text{subject to} \quad & y_1 = y_2 \end{aligned}\]Step 2: form dual problem
\[L(x_1, y_1, x_2, y_2) = f_1(x_1, y_1) + f_2(x_2, y_2) + \nu^T(y_1 - y_2)\]separable; can minimize over $(x_1, y_1)$ and $(x_2, y_2)$ separately
\[\begin{aligned} g_1(\nu) &= \inf_{x_1, y_1} (f_1(x_1, y_1) + \nu^T y_1) = -f_1^*(0, -\nu) \\ g_2(\nu) &= \inf_{x_2, y_2} (f_2(x_2, y_2) - \nu^T y_2) = -f_2^*(0, \nu) \end{aligned}\]dual problem is: $\text{maximize} \quad g(\nu) = g_1(\nu) + g_2(\nu)$
In large-scale optimization, the challenge often lies in complicating constraints—constraints that, if removed, would allow the problem to decompose into smaller, independent subproblems. Lagrangean Decomposition is a powerful technique designed to exploit such structures, providing tighter bounds than standard LP relaxations and enabling the solution of massive-scale Mixed-Integer Linear Programs (MILP) and Nonlinear Programs (MINLP).
This tutorial follows the theoretical foundations and industrial applications developed by Ignacio Grossmann and Bora Tarhan at Carnegie Mellon University.
Consider a Mixed-Integer Programming (MIP) problem with complicating constraints:
\[\begin{aligned} Z = \min \quad & c^T x \\ \text{s.t.} \quad & A x \ge b \quad \color{red}{\text{(Complicating Constraints)}} \\ & D x \ge e \quad \color{blue}{\text{(Easy Constraints)}} \\ & x \in \mathbb{Z}^n_+ \end{aligned}\]The Lagrangean Relaxation absorbs the complicating constraints into the objective function with a penalty term (Lagrangean multiplier $u \ge 0$):
\[Z_{LR}(u) = \min_{x \in X} \{ c^T x + u^T (b - Ax) \}\]where $X = { x \in \mathbb{Z}^n_+ : Dx \ge e }$.
For any $u \ge 0$, $Z_{LR}(u)$ provides a lower bound to the original minimization problem. This is because:
To find the tightest bound, we solve the Lagrangean Dual: \(Z_D = \max_{u \ge 0} Z_{LR}(u)\)
Optimizing the Lagrangean multipliers can be interpreted as optimizing the primal objective function on the intersection of the convex hull of the easy constraints and the LP relaxation of the complicating constraints.
\[Z_D = \min \{ c^T x : x \in \text{Conv}(Dx \ge e, x \in \mathbb{Z}^n_+) \cap \{ x : Ax \ge b \} \}\]As established by (missing reference), Lagrangean relaxation yields a bound at least as tight as the LP relaxation: \(Z_{LP} \le Z_D \le Z(P)\)
Lagrangean Decomposition (or Variable Splitting) is a special case of relaxation. Instead of dualizing existing constraints, we define separate variables for different sets of constraints and add a “linking constraint” equating them.
Consider: \(Z = \min \{ c^T x : Ax \ge b, Dx \ge e, x \in \mathbb{Z}^n_+ \}\)
We split $x$ into $x$ and $y$: \(\begin{aligned} Z = \min \quad & c^T x \\ \text{s.t.} \quad & Ax \ge b, x \in \mathbb{Z}^n_+ \\ & Dy \ge e, y \in \mathbb{Z}^n_+ \\ & x = y \quad \color{red}{\text{(Linking Constraint)}} \end{aligned}\)
Dualizing $x = y$ with multipliers $v$ (unrestricted in sign) yields: \(Z_{LD}(v) = \min_{x, y} \{ c^T x + v^T (y - x) : Ax \ge b, Dy \ge e, x, y \in \mathbb{Z}^n_+ \}\)
The problem decomposes into two independent subproblems:
The dual problem is solved iteratively to update multipliers $u$ or $v$:
The overall algorithm for Lagrangean decomposition can be visualized as an iterative feedback loop between the master problem (multiplier update) and the subproblems.
graph TD
Start((Start)) --> Init[Set Initial Multipliers v0]
Init --> Subprobs[Solve Subproblem 1 & 2 independently]
Subprobs --> Bound[Compute Z_LD = Z_LD1 + Z_LD2]
Bound --> Check{Convergence?}
Check -- No --> Update[Update Multipliers v using Subgradient]
Update --> Subprobs
Check -- Yes --> End((Optimal Bound Found))
Lagrangean decomposition is particularly effective in high-stakes engineering decisions.
Managing the design and planning of offshore hydrocarbon fields involves complex economic objectives (Minimize Total Cost) across multiple time periods.
Motivation: Complex royalties and taxes are often local to specific platforms. By dualizing the linking flows between platforms, the model decomposes into independent models for each Well Platform (WP).
The simultaneous optimization of testing tasks (clinical trials) and batch plant manufacturing facilities is a massive MILP.
The Trick: Decouple the Scheduling of tests from the Design/Planning of the manufacturing site.
Global multisite distribution involves determining product manufacturing sites and supply chains for multiple markets.
Spatial vs. Temporal Decomposition:
Gasfield planning under uncertainty leads to massive scenario trees.
Non-Anticipativity: Scenarios are linked by non-anticipativity constraints (decisions must be the same if scenarios have not yet diverged).
For a deeper dive, see the foundational work by (missing reference) and the applications by (missing reference).