Significant progress has been witnessed in learning-based Multi-view Stereo (MVS) of supervised and unsupervised settings. To combine their respective merits in accuracy and completeness, meantime reducing the demand for expensive labeled data, this paper explores a novel semi-supervised setting of learning-based MVS problem that only a tiny part of the MVS data is attached with dense depth ground truth. However, due to huge variation of scenarios and flexible setting in views, semi-supervised MVS problem (Semi-MVS) may break the basic assumption in classic semi-supervised learning, that unlabeled data and labeled data share the same label space and data distribution. To handle these issues, we propose a novel semi-supervised MVS framework, namely SE-MVS. For the simple case that the basic assumption works in MVS data, consistency regularization encourages the model predictions to be consistent between original sample and randomly augmented sample via constraints on KL divergence. For further troublesome case that the basic assumption is conflicted in MVS data, we propose a novel style consistency loss to alleviate the negative effect caused by the distribution gap. The visual style of unlabeled sample is transferred to labeled sample to shrink the gap, and the model prediction of generated sample is further supervised with the label in original labeled sample. The experimental results on DTU, BlendedMVS, GTA-SFM, and Tanks\&Temples datasets show the superior performance of the proposed method. With the same settings in backbone network, our proposed SE-MVS outperforms its fully-supervised and unsupervised baselines.
翻译:在以学习为基础的多视系统(MVS)中,受监管和不受监管环境的多视系统(MVS)取得了显著进展。为了将各自在准确性和完整性方面的优点结合起来,同时减少对昂贵的标签数据的需求,本文探讨了一个新的半监督的基于学习的MVS问题,即MVS数据中只有很小的一部分附着了深厚的地面真相。然而,由于各种假设和观点的灵活设置差异很大,半监督的MVS问题(Semi-MVS)可能会打破典型的半监督学习中的基本假设,即无标签数据和标签数据共享相同的标签空间和数据分布。为了处理这些问题,我们提出了一个新的半监督的基于学习的MVS问题,即SE-MVS半监督的MVS问题。关于MVS数据中基本假设与深深深深深深深的深度数据连接,一致性鼓励模型预测在原始样本和随机增加样本之间保持一致。关于MVS数据中基本假设存在冲突,我们提议的非风格一致性损失将S的S级S与S的Servervl 样转换为完全的S。