论文标题
复发引起的隐式n-grams
Implicit N-grams Induced by Recurrence
论文作者
论文摘要
尽管基于自我注意力的模型(例如变形金刚)在自然语言处理(NLP)任务上取得了巨大的成功,但最近的研究表明,它们对建模顺序转换有局限性(Hahn,2020),这可能会促使重新审查复发性神经网络(RNN),这在处理顺序数据上表现出了令人印象深刻的结果。尽管先前尝试了解释RNN的尝试,但其内部机制尚未完全理解,并且关于它们如何确切捕获顺序特征的问题仍然在很大程度上不清楚。在这项工作中,我们提出了一项研究,该研究表明实际上存在一些可解释的组件,这些组件位于隐藏状态中,这让人想起经典的n-grams特征。我们在下游情感分析任务上评估了从训练有素的RNN中提取的可解释的特征,发现它们可用于建模有趣的语言现象,例如否定和强化。此外,我们研究了单独使用此类N-Gram组件作为编码器在诸如情感分析和语言建模之类的任务上的功效,从而表明它们可能在为RNN的整体绩效做出贡献中起重要作用。我们希望我们的发现可以为RNN体系结构增添解释性,并为提出顺序数据的新体系结构提供灵感。
Although self-attention based models such as Transformers have achieved remarkable successes on natural language processing (NLP) tasks, recent studies reveal that they have limitations on modeling sequential transformations (Hahn, 2020), which may prompt re-examinations of recurrent neural networks (RNNs) that demonstrated impressive results on handling sequential data. Despite many prior attempts to interpret RNNs, their internal mechanisms have not been fully understood, and the question on how exactly they capture sequential features remains largely unclear. In this work, we present a study that shows there actually exist some explainable components that reside within the hidden states, which are reminiscent of the classical n-grams features. We evaluated such extracted explainable features from trained RNNs on downstream sentiment analysis tasks and found they could be used to model interesting linguistic phenomena such as negation and intensification. Furthermore, we examined the efficacy of using such n-gram components alone as encoders on tasks such as sentiment analysis and language modeling, revealing they could be playing important roles in contributing to the overall performance of RNNs. We hope our findings could add interpretability to RNN architectures, and also provide inspirations for proposing new architectures for sequential data.