论文标题
多模式的深度展开,用于指导图像超分辨率
Multimodal Deep Unfolding for Guided Image Super-Resolution
论文作者
论文摘要
在低分辨率观察中,高分辨率图像的重建是成像中的一个逆问题。深度学习方法依靠培训数据来学习从低分辨率输入到高分辨率输出的端到端映射。与不包含有关该问题的域知识的现有深层多模型不同,我们提出了一种多模式深度学习设计,该设计结合了稀疏的先验,并允许从另一个图像模式中有效地集成信息到网络体系结构中。我们的解决方案依靠一个新颖的深层展开操作员,执行类似于迭代算法的步骤,用于卷积稀疏编码,并带有侧面信息。因此,提出的神经网络可通过设计解释。深度展开的体系结构用作指导图像超分辨率的多模式框架的核心组成部分。通过采用剩余学习来提高训练效率来研究替代的多模式设计。提出的多模式方法应用于近红外和多光谱图像的超分辨率,以及使用RGB图像作为附带信息的深度进行采样。实验结果表明,我们的模型优于最先进的方法。
The reconstruction of a high resolution image given a low resolution observation is an ill-posed inverse problem in imaging. Deep learning methods rely on training data to learn an end-to-end mapping from a low-resolution input to a high-resolution output. Unlike existing deep multimodal models that do not incorporate domain knowledge about the problem, we propose a multimodal deep learning design that incorporates sparse priors and allows the effective integration of information from another image modality into the network architecture. Our solution relies on a novel deep unfolding operator, performing steps similar to an iterative algorithm for convolutional sparse coding with side information; therefore, the proposed neural network is interpretable by design. The deep unfolding architecture is used as a core component of a multimodal framework for guided image super-resolution. An alternative multimodal design is investigated by employing residual learning to improve the training efficiency. The presented multimodal approach is applied to super-resolution of near-infrared and multi-spectral images as well as depth upsampling using RGB images as side information. Experimental results show that our model outperforms state-of-the-art methods.