论文标题

神经网络与正则化的一致性

Consistency of Neural Networks with Regularization

论文作者

Shen, Xiaoxi, Lin, Jinghang

论文摘要

神经网络由于其在自然语言处理和计算机视觉等应用方面的成功而引起了很多关注。对于大规模数据,由于神经网络中的参数数量众多,过度拟合是训练神经网络的一个问题。为了避免过度拟合,一种常见的方法是惩罚参数,尤其是神经网络中的权重。尽管神经网络在许多应用中都表现出了其优势,但受到惩罚神经网络的理论基础尚未得到良好的建立。我们本文的目标是提出具有正规化神经网络的一般框架,并证明其一致性。在某些条件下,随着样本量的增加,估计的神经网络将融合到真正的潜在功能。筛子的方法和最小神经网络的理论用于克服参数无法识别的问题。已经考虑了两种类型的激活函数:双曲线切线函数(TANH)和整流线性单元(relu)。进行了模拟以验证一致性定理的验证。

Neural networks have attracted a lot of attention due to its success in applications such as natural language processing and computer vision. For large scale data, due to the tremendous number of parameters in neural networks, overfitting is an issue in training neural networks. To avoid overfitting, one common approach is to penalize the parameters especially the weights in neural networks. Although neural networks has demonstrated its advantages in many applications, the theoretical foundation of penalized neural networks has not been well-established. Our goal of this paper is to propose the general framework of neural networks with regularization and prove its consistency. Under certain conditions, the estimated neural network will converge to true underlying function as the sample size increases. The method of sieves and the theory on minimal neural networks are used to overcome the issue of unidentifiability for the parameters. Two types of activation functions: hyperbolic tangent function(Tanh) and rectified linear unit(ReLU) have been taken into consideration. Simulations have been conducted to verify the validation of theorem of consistency.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源