For instance, we may have and (for the PDEs describing the same domain ).
The goal of the lifting approach is to find a lifting operator
Then,
First, let's look at some properties for operators.
Linearity
The operator is linear if for
Such a linear operator is bounded (or continuous) if there exists such that
This is similar to Lipschitz Continuity. The smallest (infimum) such that the above holds is the operator norm, which for and a symmetric operator is the largest absolute eigenvalue, as , and the norm is the same between both sides (other than taking the absolute value of ). In this case, the operator norm is the same as the spectral norm. For non-square and non-symmetric matrices, this is the square root of the largest eigenvalue of .
For an operator between function spaces, an integral Kernel and a (linear) differential operator are linear:
Further, the sum and composition of linear operators is still a linear operator. Thus, we can combine and repeat each of these separate forms while still being a linear operator.
Example: Operator from Diffusion to Advection Diffusion
Let and be solution spaces for these equations. Without sufficient initial conditions / boundary conditions, multiple functions are in these function spaces (which are themselves in ). We seek an operator / a map of the form given above. Further, by leaving unspecified, we can update too: . However, we take to simplify things. In this, we assume a particular form of :
By applying and to the right most term, we can derive an expression involving only and . This eventually gives
for any . Here, we found not just a single operator , but a family of these operators. Thus the operator is not necessarily unique. Intuitively, we can have scaled by an arbitrary constant, and the -system will hold, as it is a linear PDE. By imposing an initial condition, we should have fewer solutions. If the -system has initial condition , then
For the case of , we see that . Thus, we still have the same scaling ambiguity. We can think of this as having multiple -systems. There are infinitely many diffusion systems (scaled by the constant) that can be transformed into a single -system. I doubt that adding boundary conditions would fix this issue, as it still persists after investigating the initial conditions. Let us suppose that we know the initial condition, and choose satisfying the initial condition. In other words, we align and at :
This requires
which only holds for , which means . Thus, for the nontrivial case (), there is no such transformation of the form assumed in . Due to the generality of the derivation, I also think that there is no such transformation of the more general form .
Example: Operator from Viscous Burgers' to Diffusion
We use the first transformation
Plugging this in and simplifying gives
We integrate with respect to and get an arbitrary constant, , and
Next, we introduce another transformation . Solving for and plugging into the above gives the diffusion equation,
Yet, for the initial condition, we have that
This indefinite integral introduces an arbitrary constant. Taking as the antiderivative of the initial condition , this gives
Note that is an arbitrary multiplication constant. Again, only in very special conditions will we have that the initial conditions of the two systems are the same. For this case, also note that the first transformation would be nonlinear. The second transformation seems to be linear, but as a whole the mapping from to () is nonlinear.
Initial Condition
We require that the -system and the -system begin with the same state, . That is,
In other words, the initial condition requires that at . For instance, consider the form of . Then,
For a kernel operator, this requires that . The derivative operator may not work without specific requirements on the IC.
Continuity
There must be some sort of regularity requirement based on the continuity of both and . If these are both continuous, then it is reasonable to expect that is in some way continuous too. For example, we may expect to be continuous. would have similar requirements, but maybe slightly looser, due to the smoothness introduced by the integral operator.
Operator as PDE Discovery with Solution Operator
One option to consider for is by modifying the original PDE. We wish to construct First, constructs the residual according to , and stacks the original state as the input to . Then, modifies the residual, potentially with a new term resembling . Finally, solves this modified PDE, returning the state, which is ideally . The modified PDE aims to approximate the term (keeping the time derivative).
Using a simpler , we can think of mapping directly between the state spaces, not relying on the PDE structure directly in the mapping of . Conversely, we can think of this simpler mapping as "inducing" some PDE for .
Parameterization
We take some inspiration from the exact initial conditions for PINNs and parameterize as the following general form
taking for instance
The consideration of the form of is discussed in PINNs, notably with , , and . For our application, describes how close is to , at a given time.
We should consider more closely the error term
as this informs the functional form. We may expect, due to the accumulation of errors in the time integration, . This is further discussed later
As mentioned above, this function maps from function spaces to function spaces, e.g. , or more generally Let us choose subspaces and , which have finite bases:
In other words,
Now, the operator can be fully defined by , or .
First implementation
In the spirit motivated above, a first implementation may be as follows:
Construct to satisfy the initial and boundary conditions. This may use the -system PDE too. In other words, this can be a variant of Physics Informed Neural Network, without using the available system data (which comes from the -system).
We will evaluate at a collection of points , for a fixed time , giving .
Our transformation will apply at multiple times, rather than mapping from to , it will map from to . By repeatedly applying this, we can get the original operator that maps over both and . This will also map the discretized form , i.e. . Notably, for the discretized initial condition, , this operator should behave as the identity. It may be worthwhile to design this operator in such a manner.
Train both of these objects simultaneously. They serve different purposes, so hopefully they do not conflict with one another. It might make sense to train first.
For a time invariant correction, we require that . One such form is
However, is discontinuous. Thus, we suggest using a new measure . If we consider as an input to this function, then it maps . This resembles normalized inner products. For instance, we may use
This is not differentiable at , but that might be okay.