论文标题

使图形神经网络值得低数据

Making Graph Neural Networks Worth It for Low-Data Molecular Machine Learning

论文作者

Pappu, Aneesh, Paige, Brooks

论文摘要

由于学到的表示的表现力,图形神经网络在分子上的机器学习中变得非常流行。但是,分子机器学习是一种经典的低DATA制度,尚不清楚图形神经网络可以避免在低资源设置中过度拟合。相反,由于参数数量减少和手动设计的功能,指纹方法是低数据表环境的传统标准。在这项工作中,我们研究了图形神经网络在小数据设置中是否具有与指纹方法的“便宜”替代方案相比,在小型数据设置中是否具有竞争力。当我们发现它们不是时,我们将探索预训练和元学习方法MAML(和变体FO-MAML和ANIL),以通过从相关任务中转移学习来改善图形神经网络性能。我们发现MAML和FO-MAML确实使图神经网络能够超过指纹的模型,即使在具有严格限制的数据可用性的设置中,也为使用图神经网络提供了一条路径。与以前的工作相反,我们发现在这种分子环境中其他元学习方法的表现更糟糕。我们的结果提出了两个原因:分子机器学习任务可能需要特定于任务的大量适应,并且相对于火车任务的测试任务的分配变化可能会导致较差的Anil性能。

Graph neural networks have become very popular for machine learning on molecules due to the expressive power of their learnt representations. However, molecular machine learning is a classically low-data regime and it isn't clear that graph neural networks can avoid overfitting in low-resource settings. In contrast, fingerprint methods are the traditional standard for low-data environments due to their reduced number of parameters and manually engineered features. In this work, we investigate whether graph neural networks are competitive in small data settings compared to the parametrically 'cheaper' alternative of fingerprint methods. When we find that they are not, we explore pretraining and the meta-learning method MAML (and variants FO-MAML and ANIL) for improving graph neural network performance by transfer learning from related tasks. We find that MAML and FO-MAML do enable the graph neural network to outperform models based on fingerprints, providing a path to using graph neural networks even in settings with severely restricted data availability. In contrast to previous work, we find ANIL performs worse that other meta-learning approaches in this molecule setting. Our results suggest two reasons: molecular machine learning tasks may require significant task-specific adaptation, and distribution shifts in test tasks relative to train tasks may contribute to worse ANIL performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源