论文标题
系统地设计具有神经算术逻辑单元的细胞图像上的更好实例计数模型
Systematically designing better instance counting models on cell images with Neural Arithmetic Logic Units
论文作者
论文摘要
训练以计数实例的神经网络模型的最大问题是,每当测试范围进行高训练范围概括误差时,即它们不是训练范围之外的好概括。考虑与训练数据中使用的图像相比,通常会遇到更较高的细胞计数的更密集的图像。通过对较高的细胞计数范围进行更好的预测,我们旨在为细胞计数创建更好的概括系统。通过用于算术操作的神经算术逻辑单元(NALU)的架构建议,计数的任务对于更高的数字范围而言是可行的,这些范围不包括在培训数据中以更好的精度包括在培训数据中。作为我们研究的一部分,我们使用了这些单元和不同的其他激活功能来学习细胞计数任务,并使用两个不同的体系结构,即完全卷积回归网络和U-NET。这些数值偏见的单元是以残留串联层的形式添加到原始体系结构的形式,并对这些新提出的更改进行了比较实验研究。这项比较研究是用这些模型优化回归损失问题来描述的,这些模型接受了广泛的数据增强技术训练。我们能够在细胞计数任务的实验中取得更好的结果,并以残留层串联连接的形式引入这些数值偏见的单元,以现有的架构。我们的结果证实,上述数字偏见的单元确实有助于模型学习数字数量以获得更好的概括结果。
The big problem for neural network models which are trained to count instances is that whenever test range goes high training range generalization error increases i.e. they are not good generalizers outside training range. Consider the case of automating cell counting process where more dense images with higher cell counts are commonly encountered as compared to images used in training data. By making better predictions for higher ranges of cell count we are aiming to create better generalization systems for cell counting. With architecture proposal of neural arithmetic logic units (NALU) for arithmetic operations, task of counting has become feasible for higher numeric ranges which were not included in training data with better accuracy. As a part of our study we used these units and different other activation functions for learning cell counting task with two different architectures namely Fully Convolutional Regression Network and U-Net. These numerically biased units are added in the form of residual concatenated layers to original architectures and a comparative experimental study is done with these newly proposed changes. This comparative study is described in terms of optimizing regression loss problem from these models trained with extensive data augmentation techniques. We were able to achieve better results in our experiments of cell counting tasks with introduction of these numerically biased units to already existing architectures in the form of residual layer concatenation connections. Our results confirm that above stated numerically biased units does help models to learn numeric quantities for better generalization results.