论文标题
深度学习注意力机制的一般调查
A General Survey on Attention Mechanisms in Deep Learning
论文作者
论文摘要
注意是一种重要的机制,可以用于许多不同领域和任务的各种深度学习模型。这项调查概述了文献中提出的最重要的注意机制。各种注意机制是通过一个框架来解释的,该框架由一般注意力模型,均匀的符号和全面的注意机制分类法组成。此外,审查了评估注意力模型的各种措施,并讨论了基于提议的框架的注意力模型结构的方法。最后,考虑了注意力模型的未来工作。
Attention is an important mechanism that can be employed for a variety of deep learning models across many different domains and tasks. This survey provides an overview of the most important attention mechanisms proposed in the literature. The various attention mechanisms are explained by means of a framework consisting of a general attention model, uniform notation, and a comprehensive taxonomy of attention mechanisms. Furthermore, the various measures for evaluating attention models are reviewed, and methods to characterize the structure of attention models based on the proposed framework are discussed. Last, future work in the field of attention models is considered.