Recent works have shown that optical flow can be learned by deep networks from unlabelled image pairs based on brightness constancy assumption and smoothness prior. Current approaches additionally impose an augmentation regularization term for continual self-supervision, which has been proved to be effective on difficult matching regions. However, this method also amplify the inevitable mismatch in unsupervised setting, blocking the learning process towards optimal solution. To break the dilemma, we propose a novel mutual distillation framework to transfer reliable knowledge back and forth between the teacher and student networks for alternate improvement. Concretely, taking estimation of off-the-shelf unsupervised approach as pseudo labels, our insight locates at defining a confidence selection mechanism to extract relative good matches, and then add diverse data augmentation for distilling adequate and reliable knowledge from teacher to student. Thanks to the decouple nature of our method, we can choose a stronger student architecture for sufficient learning. Finally, better student prediction is adopted to transfer knowledge back to the efficient teacher without additional costs in real deployment. Rather than formulating it as a supervised task, we find that introducing an extra unsupervised term for multi-target learning achieves best final results. Extensive experiments show that our approach, termed MDFlow, achieves state-of-the-art real-time accuracy and generalization ability on challenging benchmarks. Code is available at https://github.com/ltkong218/MDFlow.
翻译:最近的作品表明,光学流动可以通过深网络从未贴标签的图像配对中学习,其基础是亮度凝固的假设和之前的光滑。目前的方法进一步对持续自我监督的图像配进一个增强性规范化的术语,这在困难的匹配区域已证明是有效的。然而,这种方法还扩大了在不受监督的环境中不可避免的不匹配现象,阻碍了学习进程走向最佳解决办法。为了打破这一困境,我们提议了一个全新的相互蒸馏框架,在教师和学生网络之间将可靠的知识相继和相继传递,以便进行其他改进。具体地说,将现成的未经监督的方法作为假标签来估计,我们的洞察力是确定一个信任选择机制,以提取相对良好的匹配,然后增加多样化的数据增量,以从教师到学生之间提取充足和可靠的知识。由于我们的方法的模糊性,我们可以选择一个更强大的学生结构来进行充分学习。最后,我们采用了更好的学生预测,在实际部署中不增加成本的情况下,将知识转回给高效的教师。我们发现,而不是作为监督的任务,我们发现,在多目标的低的M-DF方法上引入一个额外的不精确的术语,能够实现最有挑战性的标准。