Biros, George, Volkan Akçelik, Omar Ghattas, Judith Hill, David Keyes, Bart Van, and Bloemen Waanders. n.d. “Parallel Algorithms for PDE-Constrained Optimization.”
Hicken, Jason, and Juan Alonso. 2013. “Comparison of Reduced- and Full-Space Algorithms for PDE-Constrained Optimization.” In 51st AIAA Aerospace Sciences Meeting Including the New Horizons Forum and Aerospace Exposition. Aerospace Sciences Meetings. American Institute of Aeronautics and Astronautics. https://doi.org/10.2514/6.2013-1043.
Adavani, Santi S., and George Biros. 2008. “Multigrid Algorithms for Inverse Problems with Linear Parabolic PDE Constraints.” SIAM Journal on Scientific Computing 31 (1): 369–97. https://doi.org/10.1137/070687426.
Within PDEConstrained Optimization there are broadly two classes of methods: full and reduced space methods. For a state vector , design/optimization variables , an objective function , and a PDE residual , the general problem is
The PDE problem is well posed; given , we can uniquely determine from . Although multiple may produce the same value of , meaning the optimization problem can be ill-posed. In the discretized case, we have , , , and . With adjoint variables or Lagrange multipliers as , the Lagrangian is
and the first-order optimality conditions (KKT Conditions) require that its gradient vanishes, implying
Now to focus on a few terms:
is the standard Jacobian of the PDE, which is often used in its solution. This first equation is also known as the Adjoint Equation.
The condition (Primal Feasibility) simply states that must satisfy the PDE corresponding to .
In many cases, is , as does not explicitly appear in the loss, e.g. in (unregularized) Inverse Problems.
There are resulting equations.
Full Space methods solve the KKT system directly, optimizing over variables, while Reduced Space methods solve the first and third sets of equations before tackling the second set. In other words, they remove and from the optimization problem by solving the original PDE and the adjoint equation, respectively, for a given value of .
Full Space
Applying a Newton step to the KKT system gives a matrix system for the update to and . This is generally too large for a direct solve, but Iterative Linear Solvers also suffer from ill-conditioning. These methods generally require specialized Preconditioners and may be slow to converge.
Reduced Space
We can rewrite the optimization such that depends on , which happens implicitly through . Then our problem is
where , and by construction. This option is more clearly dependent on the well-posedness of the forward problem, as we must invert to get . With this simplification we can either reduce the original KKT system or apply first order optimality again to get the same result of
You can't use 'macro parameter character #' in math mode\partial_{\mathbf{d}}\mathcal{J} + [\partial_{\mathbf{d}} \mathbf{c}] { #T} \mathbf{\lambda} = 0.
Note also that , implicitly through the adjoint equation. These methods require full forward (and adjoint) solves far from the optimum. Through the inversion of (or strict enforcement of ), the problem may be more nonlinear in .