论文标题
用于操作员学习的多变量错误建模
Error-in-variables modelling for operator learning
论文作者
论文摘要
深度操作员学习已成为减少订购建模和PDE模型发现的有前途的工具。利用深层神经网络的表达能力,尤其是在高维度中,这种方法了解功能状态变量之间的映射。虽然所提出的方法仅在因变量中假设噪声,但用于操作员学习的实验和数值数据也通常在自变量中显示出噪声,因为这两个变量都代表了符合测量误差的信号。在标量数据的回归中,未能说明嘈杂的自变量会导致参数估计。使用嘈杂的自变量,通过普通最小二乘(OLS)拟合的线性模型将显示衰减偏置,其中斜率将被低估。在这项工作中,我们得出了在自变量和因变量中都具有白噪声的线性操作器回归的衰减偏差的类似物。在非线性环境中,我们在计算上证明了在自变量中存在噪声的情况下,汉堡操作员的作用不足。我们提出了两种操作员回归方法,mor-physics和deponet的多变量错误模型(EIV)模型,并证明这些新模型在存在各种操作员学习问题的嘈杂自变量的情况下减少了偏见。考虑到1d和2d的汉堡操作员,我们证明了EIV操作员的学习能力稳健地在击败OLS操作员学习的高噪声政权中恢复运营商。我们还引入了一个EIV模型,以进行时间不断发展的PDE发现,并表明OLS和EIV从损坏的数据中学习Kuramoto-Sivashinsky Evolution Operator的表现相似,这表明偏见在OLS操作员学习中的影响取决于目标操作员的规律性。
Deep operator learning has emerged as a promising tool for reduced-order modelling and PDE model discovery. Leveraging the expressive power of deep neural networks, especially in high dimensions, such methods learn the mapping between functional state variables. While proposed methods have assumed noise only in the dependent variables, experimental and numerical data for operator learning typically exhibit noise in the independent variables as well, since both variables represent signals that are subject to measurement error. In regression on scalar data, failure to account for noisy independent variables can lead to biased parameter estimates. With noisy independent variables, linear models fitted via ordinary least squares (OLS) will show attenuation bias, wherein the slope will be underestimated. In this work, we derive an analogue of attenuation bias for linear operator regression with white noise in both the independent and dependent variables. In the nonlinear setting, we computationally demonstrate underprediction of the action of the Burgers operator in the presence of noise in the independent variable. We propose error-in-variables (EiV) models for two operator regression methods, MOR-Physics and DeepONet, and demonstrate that these new models reduce bias in the presence of noisy independent variables for a variety of operator learning problems. Considering the Burgers operator in 1D and 2D, we demonstrate that EiV operator learning robustly recovers operators in high-noise regimes that defeat OLS operator learning. We also introduce an EiV model for time-evolving PDE discovery and show that OLS and EiV perform similarly in learning the Kuramoto-Sivashinsky evolution operator from corrupted data, suggesting that the effect of bias in OLS operator learning depends on the regularity of the target operator.