论文标题
基于等距样本的学习,自我监督的深度估计
Self-Supervised Depth Estimation with Isometric-Self-Sample-Based Learning
论文作者
论文摘要
在光度损失公式中管理动态区域是处理自我监督深度估计问题的主要问题。大多数以前的方法通过根据从另一个模块估算的掩模来删除光度损耗公式中的动态区域,从而缓解了这个问题,这使得难以充分利用训练图像。在本文中,为了解决这个问题,我们提出了一种基于等距自样本的学习(ISSL)方法,以简单但有效的方式充分利用训练图像。所提出的方法在训练期间使用符合纯静态场景假设的自我生成的图像提供了额外的监督。具体而言,等距自样本发生器通过在估计深度上应用随机的刚性转换来综合每个训练图像。因此,生成的自我样本和相应的训练图像始终遵循静态场景假设。我们表明,将ISSL模块插入几个现有模型会始终如一地提高性能。此外,它还提高了不同类型场景的深度精度,即室外场景(Kitti和Make3D)和室内场景(NYUV2),从而验证了其高效性。
Managing the dynamic regions in the photometric loss formulation has been a main issue for handling the self-supervised depth estimation problem. Most previous methods have alleviated this issue by removing the dynamic regions in the photometric loss formulation based on the masks estimated from another module, making it difficult to fully utilize the training images. In this paper, to handle this problem, we propose an isometric self-sample-based learning (ISSL) method to fully utilize the training images in a simple yet effective way. The proposed method provides additional supervision during training using self-generated images that comply with pure static scene assumption. Specifically, the isometric self-sample generator synthesizes self-samples for each training image by applying random rigid transformations on the estimated depth. Thus both the generated self-samples and the corresponding training image always follow the static scene assumption. We show that plugging our ISSL module into several existing models consistently improves the performance by a large margin. In addition, it also boosts the depth accuracy over different types of scene, i.e., outdoor scenes (KITTI and Make3D) and indoor scene (NYUv2), validating its high effectiveness.