论文标题
神经操作员的基于残余的误差校正加速无限贝叶斯逆问题
Residual-based error correction for neural operator accelerated infinite-dimensional Bayesian inverse problems
论文作者
论文摘要
我们使用功能空间之间非线性图的神经操作员或神经网络表示,以加速无限二维贝叶斯逆问题(BIPS),其模型受非线性参数偏微分方程(PDES)的模型。近年来,神经操作员因其在有限数量的参数样本下使用AS训练数据解决方案近似于PDE定义的参数到解决图的能力。如果将大量的PDE解决方案替换为训练有素的神经操作员的评估,则可以大大降低BIP的计算成本。但是,通过减少神经操作员在训练中的近似误差来减少所得BIP解决方案的误差可能是具有挑战性的,而且不可靠。我们提供了一个先验错误的结合结果,这意味着某些BIP可能会不适合神经操作员的近似错误,从而导致训练中无法访问的准确性要求。为了可靠地将神经操作员部署在BIP中,我们考虑了一种增强神经操作员性能的策略,即通过解决基于PDE残差的线性变异问题来纠正训练有素的神经操作员的预测。我们表明,具有误差校正的训练有素的神经操作员可以实现其近似误差的二次减少,同时,当模型受高度非线性PDES的控制时,同时保留了大量的后验计算加速。该策略应用于基于非线性反应问题的两个数值示例 - 扩散问题和超弹性材料的变形。我们证明,使用训练有素的神经操作员生产的两个BIP的后表示,通过误差校正大大增强。
We explore using neural operators, or neural network representations of nonlinear maps between function spaces, to accelerate infinite-dimensional Bayesian inverse problems (BIPs) with models governed by nonlinear parametric partial differential equations (PDEs). Neural operators have gained significant attention in recent years for their ability to approximate the parameter-to-solution maps defined by PDEs using as training data solutions of PDEs at a limited number of parameter samples. The computational cost of BIPs can be drastically reduced if the large number of PDE solves required for posterior characterization are replaced with evaluations of trained neural operators. However, reducing error in the resulting BIP solutions via reducing the approximation error of the neural operators in training can be challenging and unreliable. We provide an a priori error bound result that implies certain BIPs can be ill-conditioned to the approximation error of neural operators, thus leading to inaccessible accuracy requirements in training. To reliably deploy neural operators in BIPs, we consider a strategy for enhancing the performance of neural operators, which is to correct the prediction of a trained neural operator by solving a linear variational problem based on the PDE residual. We show that a trained neural operator with error correction can achieve a quadratic reduction of its approximation error, all while retaining substantial computational speedups of posterior sampling when models are governed by highly nonlinear PDEs. The strategy is applied to two numerical examples of BIPs based on a nonlinear reaction--diffusion problem and deformation of hyperelastic materials. We demonstrate that posterior representations of the two BIPs produced using trained neural operators are greatly and consistently enhanced by error correction.