Partial Differential Equation Discovery

Resources

Main Idea

Within Inverse Problems, PDE Discovery aims to recover a Partial Differential Equation for some system, given data from the system. This is an extension of Parameter Estimation and Neural ODEs. Often, the problem is pitched as discovering N such that ut=N(x,t,u,ux,uxx,). Note that this places stricter requirements on the function class for u than may be needed with a Weak Form . For instance ut=(κux)x is more general than ut=κxux+κuxx. This field can also be seen as a specific case of Operator Learning.

SINDy is one popular example of these methods, although there are other approaches based on numeric-symbolic hybrid deep networks (PDE-Net 2.0), and on Neural Networks (Deep Hidden Physics Models). The neural-network based methods may also represent the state as a neural network uθ(x,t). While SINDy-type methods often use Convex optimization techniques, the neural-network based methods rely on standard machine learning Unconstrained Optimization methods (like Gradient Descent).

Derivations from Constrained Formulation

Consider the following Constrained Optimization problem

minϕi=1Nu||uiu(xi,ti)||2,s.t.ut=Nϕ(u,ux,uxx),(x,t)Ω×[0,T].

Neural ODE Method

Consider a mesh grid of points in space, xRnx. Let us loosen the constraints to hold at just these points in space, rather than for the whole domain. Denote v(t)=u(x,t), and consider finite difference approximations of ux and uxx for x as D1 and D2 (which also include boundary conditions). Then the optimization problem is

minϕi=1Nu||uiu(xi,ti)||2,s.t.vt=Nϕ(v,D1v,D2v)f(v),t[0,T].

Let us consider the constraint separately. Under certain smoothness assumptions and through Fundamental Theorem of Calculus, the constraint is equivalent to

0tvtdt=v(t)v(0)=0tf(v)dt,t[0,T].

Rearranging and trading t and t, the rewritten constraint reads

v(t)=v(0)+0tfϕ(v)dt,t[0,T].

Here, we assume that this must hold for a finite set of times, and numerically approximate this integration. This is equivalent to taking a numerical solution to the method of lines ODE. Denote this vϕ(t):

vϕ(t)=ODESolve(t,v(0),fϕ).

While vϕ(t)u(x,t), let us consider an interpolation of vϕ(t) over x, uϕ(x,t). Said differently, let uϕ(x,t) be the numerical solution (with an interpolant) to the PDE.

Now, by construction the constraints are (numerically / approximately) satisfied and we have the following Unconstrained Optimization problem:

minϕi=1Nu||uiuϕ(xi,ti)||2.

The state uϕ(x,t) is an (implicit) function of the optimization variable ϕ (i.e. we solve for uϕ(x,t) in an inner-loop for each update of ϕ). Thus, this a reduced-space method.

PINNs Method

Consider the main optimization problem, but now loosen the constraints to hold only over a set of collocation points {(xj,tj)}j=1Nr, and to within some tolerance ε. Then we have

minϕi=1Nu||uiu(xi,ti)||2,s.t.|[utNϕ(u,ux,uxx)](xj,tj)|ε,j=1,,Nr.

Let us introduce a neural network representation of the state uθ(x,t), which will we also find during the optimization problem. Through automatic differentiation, we do not need to approximate ux numerically, and we can use the strong form with ut, rather than integrating in time. Thus, the final constrained optimization problem is as follows:

minθ,ϕi=1Nu||uiuθ(xi,ti)||2,s.t.|[utθNϕ(uθ,uxθ,uxxθ)](xj,tj)|ε,j=1,,Nr.

Here, the parameters θ of the state representation uθ(x,t) are updated on each iteration, placing them on equal footing with ϕ. Therefore, this approach is a full-space method.

SINDy

SINDy methods do not fit nicely into this formulation. We approximate u(x,t) in a weird way, and we also try to enforce the PDE constraints. Compared to these two methods, our approximation of u(x,t) (and ux,uxx,ut) is entirely independent of the PDE we're discovering. We construct u and derivatives through denoising and Numerical Differentiation applied to the data. We add in some Sparsity restriction to the form of Nϕ, but this should be fine.

Todo

Is there a way to show that for any approximation of u(x,t) independent of the PDE, there's an adversarial PDE such that SINDy fails? It seems like by choosing a (biased) denoising and numerical differentiation method, some PDE could take advantage of this.