论文标题
RGBD-NET:预测新视图合成的颜色和深度图像
RGBD-Net: Predicting color and depth images for novel views synthesis
论文作者
论文摘要
我们提出了一种新的级联体系结构,用于新型视图合成,称为RGBD-NET,它由两个核心组成部分组成:分层深度回归网络和深度感知的发电机网络。前者通过使用自适应深度缩放来预测目标视图的深度图,而后者则利用了预测的深度和呈现在空间和时间上一致的目标图像。在标准数据集的实验评估中,RGBD-NET不仅要明确的余量优于最先进的,而且还可以很好地推广到没有每场比赛优化的新场景。此外,我们表明可以选择对RGBD-NET进行培训,而无需深度监督,同时仍保留高质量的渲染。多亏了深度回归网络,RGBD-NET也可用于创建比某些最新的多视图立体声方法更准确的密集的3D点云。
We propose a new cascaded architecture for novel view synthesis, called RGBD-Net, which consists of two core components: a hierarchical depth regression network and a depth-aware generator network. The former one predicts depth maps of the target views by using adaptive depth scaling, while the latter one leverages the predicted depths and renders spatially and temporally consistent target images. In the experimental evaluation on standard datasets, RGBD-Net not only outperforms the state-of-the-art by a clear margin, but it also generalizes well to new scenes without per-scene optimization. Moreover, we show that RGBD-Net can be optionally trained without depth supervision while still retaining high-quality rendering. Thanks to the depth regression network, RGBD-Net can be also used for creating dense 3D point clouds that are more accurate than those produced by some state-of-the-art multi-view stereo methods.