论文标题
从Boltzmann机器到神经网络再返回
From Boltzmann Machines to Neural Networks and Back Again
论文作者
论文摘要
图形模型是建模高维数据的强大工具,但是在存在潜在变量的情况下,学习图形模型是很困难的。在这项工作中,我们为学习受限制的Boltzmann机器提供了新的结果,这可能是最深入的潜在变量模型。我们的结果基于在$ \ ell _ {\ infty} $界输入下学习两层神经网络的新连接;对于这两个问题,我们在噪声稀疏的奇偶校验中给出了几乎最佳的结果。使用RBMS和FeedForward网络之间的连接,我们还启动了$监督的〜RBMS $ [Hinton,2012]的理论研究,这是一种神经网络学习的版本,将其与未知功能类别的架构相结合的图形模型诱导的分布假设。然后,我们给出了一种算法,用于学习具有比没有分配假设的相关网络类别的自然监督RBM的运行时间更好。
Graphical models are powerful tools for modeling high-dimensional data, but learning graphical models in the presence of latent variables is well-known to be difficult. In this work we give new results for learning Restricted Boltzmann Machines, probably the most well-studied class of latent variable models. Our results are based on new connections to learning two-layer neural networks under $\ell_{\infty}$ bounded input; for both problems, we give nearly optimal results under the conjectured hardness of sparse parity with noise. Using the connection between RBMs and feedforward networks, we also initiate the theoretical study of $supervised~RBMs$ [Hinton, 2012], a version of neural-network learning that couples distributional assumptions induced from the underlying graphical model with the architecture of the unknown function class. We then give an algorithm for learning a natural class of supervised RBMs with better runtime than what is possible for its related class of networks without distributional assumptions.