论文标题
Gumbel-Softmax选择性网络
Gumbel-Softmax Selective Networks
论文作者
论文摘要
ML模型通常在较大的系统的上下文中运行,该系统可以在ML模型不确定时可以调整其响应,例如倒在安全默认值或循环中的人。这种通常遇到的操作上下文要求训练ML模型的原则性技术,可以选择放弃何时不确定。选择性的神经网络经过了戒除的集成选项的培训,使他们能够学会识别并优化可以对其进行自信预测的数据分布的子集。但是,由于二进制选择函数的非差异性(是预测还是弃权的离散决定),优化选择性网络是具有挑战性的。本文提出了一种培训选择性网络的通用方法,该方法利用Gumbel-Softmax重新聚集技巧,以在端到端可区分培训框架内进行选择。公共数据集上的实验证明了Gumbel-Softmax选择性网络的潜在选择性回归和分类。
ML models often operate within the context of a larger system that can adapt its response when the ML model is uncertain, such as falling back on safe defaults or a human in the loop. This commonly encountered operational context calls for principled techniques for training ML models with the option to abstain from predicting when uncertain. Selective neural networks are trained with an integrated option to abstain, allowing them to learn to recognize and optimize for the subset of the data distribution for which confident predictions can be made. However, optimizing selective networks is challenging due to the non-differentiability of the binary selection function (the discrete decision of whether to predict or abstain). This paper presents a general method for training selective networks that leverages the Gumbel-softmax reparameterization trick to enable selection within an end-to-end differentiable training framework. Experiments on public datasets demonstrate the potential of Gumbel-softmax selective networks for selective regression and classification.