论文标题
通过物理引导的神经网络识别状态功能,具有身体上的内部层
Identification of state functions by physically-guided neural networks with physically-meaningful internal layers
论文作者
论文摘要
用数据驱动的预测替换良好的理论模型在工程和科学中并不像社会和经济领域那样简单。科学问题大多数时候都遭受了数据的匮乏,而它们可能涉及大量以复杂和非平稳方式相互作用的变量和参数,从而遵守某些物理定律。此外,一个基于物理的模型不仅有用,还可以通过解释其结构,参数和数学属性来获得知识。解决这些缺点的方法似乎是数据驱动方法的巨大预测能力与基于物理模型的科学一致性和解释性的无缝融合。我们在这里使用物理约束的神经网络(PCNN)的概念来预测物理系统中的输入输出关系,同时履行物理约束。有了这个目标,系统的内部隐藏状态变量与一组内部神经元层相关联,其值受已知的物理关系以及对系统的任何其他知识的约束。此外,当拥有足够的数据时,可以推断有关系统内部结构的知识,如果对系统的内部结构进行了了解,以预测特定输入输出关系的状态参数。我们表明,这种方法除了获得基于物理的预测,加速训练过程,还减少了获得相似精度所需的数据量,部分过滤了实验数据中的内在噪声并提供了提高的外推能力。
Substitution of well-grounded theoretical models by data-driven predictions is not as simple in engineering and sciences as it is in social and economic fields. Scientific problems suffer most times from paucity of data, while they may involve a large number of variables and parameters that interact in complex and non-stationary ways, obeying certain physical laws. Moreover, a physically-based model is not only useful for making predictions, but to gain knowledge by the interpretation of its structure, parameters, and mathematical properties. The solution to these shortcomings seems to be the seamless blending of the tremendous predictive power of the data-driven approach with the scientific consistency and interpretability of physically-based models. We use here the concept of physically-constrained neural networks (PCNN) to predict the input-output relation in a physical system, while, at the same time fulfilling the physical constraints. With this goal, the internal hidden state variables of the system are associated with a set of internal neuron layers, whose values are constrained by known physical relations, as well as any additional knowledge on the system. Furthermore, when having enough data, it is possible to infer knowledge about the internal structure of the system and, if parameterized, to predict the state parameters for a particular input-output relation. We show that this approach, besides getting physically-based predictions, accelerates the training process, reduces the amount of data required to get similar accuracy, filters partly the intrinsic noise in the experimental data and provides improved extrapolation capacity.