论文标题
SVD的观点,用于增强deponet灵活性和可解释性
SVD Perspectives for Augmenting DeepONet Flexibility and Interpretability
论文作者
论文摘要
深度运算符网络(DeepOnets)是快速,准确仿真复杂动力学的强大架构。由于其非凡的概括能力主要是通过基于投影的属性来实现的,因此我们研究了与奇异值分解(SVD)得出的低级别技术的连接。我们证明,正交分解(POD) - 神经网络背后的某些概念可以改善Deponet的设计和训练阶段。这些想法使我们获得了我们命名SVD-Deeponet的方法扩展。此外,通过多个SVD分析,我们发现deponet从其基于投影的属性强效率上继承在表示以对称为特征的动力学时。受到移动pod的工作的启发,我们开发了FlexDeeponet,这是一种依赖于转换网络来生成移动参考框架并隔离动力学的刚性组件的体系结构增强。这样,物理可以在没有旋转,翻译和拉伸的潜在空间上表示,并且可以以低维度进行准确的投影。除了灵活性和解释性外,提出的观点还提高了DeWonet的概括能力和计算效率。例如,我们显示FlexDeeponet可以通过依靠比香草体系结构的训练参数少95%的训练参数来准确地替代19个变量的动力学。我们认为,基于DEAPONET和SVD的方法可以互相受益。特别是,前者以非结构化数据和物理知识的约束形式利用多个数据源和多重级知识的灵活性有可能极大地扩展诸如POD和PCA之类的方法的适用性。
Deep operator networks (DeepONets) are powerful architectures for fast and accurate emulation of complex dynamics. As their remarkable generalization capabilities are primarily enabled by their projection-based attribute, we investigate connections with low-rank techniques derived from the singular value decomposition (SVD). We demonstrate that some of the concepts behind proper orthogonal decomposition (POD)-neural networks can improve DeepONet's design and training phases. These ideas lead us to a methodology extension that we name SVD-DeepONet. Moreover, through multiple SVD analyses, we find that DeepONet inherits from its projection-based attribute strong inefficiencies in representing dynamics characterized by symmetries. Inspired by the work on shifted-POD, we develop flexDeepONet, an architecture enhancement that relies on a pre-transformation network for generating a moving reference frame and isolating the rigid components of the dynamics. In this way, the physics can be represented on a latent space free from rotations, translations, and stretches, and an accurate projection can be performed to a low-dimensional basis. In addition to flexibility and interpretability, the proposed perspectives increase DeepONet's generalization capabilities and computational efficiencies. For instance, we show flexDeepONet can accurately surrogate the dynamics of 19 variables in a combustion chemistry application by relying on 95% less trainable parameters than the ones of the vanilla architecture. We argue that DeepONet and SVD-based methods can reciprocally benefit from each other. In particular, the flexibility of the former in leveraging multiple data sources and multifidelity knowledge in the form of both unstructured data and physics-informed constraints has the potential to greatly extend the applicability of methodologies such as POD and PCA.