In this paper, we present a CNN-based fully unsupervised method for motion segmentation from optical flow. We assume that the input optical flow can be represented as a piecewise set of parametric motion models, typically, affine or quadratic motion models. The core idea of our work is to leverage the Expectation-Maximization (EM) framework in order to design in a well-founded manner a loss function and a training procedure of our motion segmentation neural network that does not require either ground-truth or manual annotation. However, in contrast to the classical iterative EM, once the network is trained, we can provide a segmentation for any unseen optical flow field in a single inference step and without estimating any motion models. We investigate different loss functions including robust ones and propose a novel efficient data augmentation technique on the optical flow field, applicable to any network taking optical flow as input. In addition, our method is able by design to segment multiple motions. Our motion segmentation network was tested on four benchmarks, DAVIS2016, SegTrackV2, FBMS59, and MoCA, and performed very well, while being fast at test time.
翻译:在本文中,我们提出了一个完全不受监督的光学流动运动分解方法。我们认为,输入的光学流可以作为一组分光运动模型,典型的直角或二次运动模型来代表。我们工作的核心思想是利用期望-最大化(EM)框架,以便以有充分根据的方式设计一个损失函数和我们运动分解神经网络的培训程序,而这种程序既不需要地面真相,也不需要人工说明。然而,与古典迭代EM不同的是,一旦对网络进行了培训,我们就可以在单一的推论步骤中为任何看不见的光学流场提供分解,而不对任何运动模型进行估计。我们调查不同的损失功能,包括强健的功能,并提议在光学流领域采用新的高效数据增强技术,适用于任何光学流作为投入的网络。此外,我们的方法可以通过设计来进行多部分移动。我们的运动分解网络按四个基准,DAVIS2016、SegtrackV2、FBMS59和MoCA进行测试,并且运行得十分顺利,同时正在快速测试。