论文标题
胶囊网络不需要建模所有内容
Capsule Networks Do Not Need to Model Everything
论文作者
论文摘要
胶囊网络是生物学启发的神经网络,将神经元分组为称为胶囊的向量,每个神经元明确表示对象或其部分之一。路由机制连接连续的胶囊,形成零件和对象之间的层次结构,也称为解析树。胶囊网络通常会尝试对图像中的所有元素进行建模,这需要大型网络大小来处理复杂的背景或无关的对象等复杂性。但是,这种全面的建模导致参数计数增加和计算效率低下。我们的目标是使胶囊网络仅专注于感兴趣的对象,从而减少解析树的数量。我们通过REM(路由熵最小化)来完成此操作,这是一种将类似树木的结构的熵最小化的技术。 REM通过修剪机制将模型参数分布驱动到低熵配置,从而大大降低了阶层内解析树的产生。这使胶囊能够以更少的参数和可忽略的性能损失学习更多稳定和简洁的表示。
Capsule networks are biologically inspired neural networks that group neurons into vectors called capsules, each explicitly representing an object or one of its parts. The routing mechanism connects capsules in consecutive layers, forming a hierarchical structure between parts and objects, also known as a parse tree. Capsule networks often attempt to model all elements in an image, requiring large network sizes to handle complexities such as intricate backgrounds or irrelevant objects. However, this comprehensive modeling leads to increased parameter counts and computational inefficiencies. Our goal is to enable capsule networks to focus only on the object of interest, reducing the number of parse trees. We accomplish this with REM (Routing Entropy Minimization), a technique that minimizes the entropy of the parse tree-like structure. REM drives the model parameters distribution towards low entropy configurations through a pruning mechanism, significantly reducing the generation of intra-class parse trees. This empowers capsules to learn more stable and succinct representations with fewer parameters and negligible performance loss.