论文标题

基于拓扑梯度的竞争学习

Topological Gradient-based Competitive Learning

论文作者

Barbiero, Pietro, Ciravegna, Gabriele, Randazzo, Vincenzo, Cirrincione, Giansalvo

论文摘要

拓扑学习是一个广泛的研究领域,旨在发现集合元素之间的相互空间关系。一些最常见和最古老的方法涉及使用无监督的竞争神经网络。但是,这些方法并非基于梯度优化,该梯度优化已被证明可以在无监督学习中提供特征提取的惊人结果。不幸的是,通过主要关注算法效率和准确性,深层聚类技术由过度复杂的特征提取器组成,同时使用顶层的琐碎算法。这项工作的目的是提出一种新颖的综合理论,旨在通过基于梯度的学习来弥合竞争学习,从而允许使用极强大的深度神经网络进行特征提取和投影,并结合具有竞争性学习的显着灵活性和表现力。在本文中,我们完全证明了两个基于梯度的新型竞争层的理论等效性。初步实验表明,双重方法如何接受输入矩阵的转置训练,即$ x^t $,导致在低维情况下,在低维情况下会导致更快的收敛速率和更高的训练精度。

Topological learning is a wide research area aiming at uncovering the mutual spatial relationships between the elements of a set. Some of the most common and oldest approaches involve the use of unsupervised competitive neural networks. However, these methods are not based on gradient optimization which has been proven to provide striking results in feature extraction also in unsupervised learning. Unfortunately, by focusing mostly on algorithmic efficiency and accuracy, deep clustering techniques are composed of overly complex feature extractors, while using trivial algorithms in their top layer. The aim of this work is to present a novel comprehensive theory aspiring at bridging competitive learning with gradient-based learning, thus allowing the use of extremely powerful deep neural networks for feature extraction and projection combined with the remarkable flexibility and expressiveness of competitive learning. In this paper we fully demonstrate the theoretical equivalence of two novel gradient-based competitive layers. Preliminary experiments show how the dual approach, trained on the transpose of the input matrix i.e. $X^T$, lead to faster convergence rate and higher training accuracy both in low and high-dimensional scenarios.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源