There has been a recent surge in methods that aim to decompose and segment scenes into multiple objects in an unsupervised manner, i.e., unsupervised multi-object segmentation. Performing such a task is a long-standing goal of computer vision, offering to unlock object-level reasoning without requiring dense annotations to train segmentation models. Despite significant progress, current models are developed and trained on visually simple scenes depicting mono-colored objects on plain backgrounds. The natural world, however, is visually complex with confounding aspects such as diverse textures and complicated lighting effects. In this study, we present a new benchmark called ClevrTex, designed as the next challenge to compare, evaluate and analyze algorithms. ClevrTex features synthetic scenes with diverse shapes, textures and photo-mapped materials, created using physically based rendering techniques. It includes 50k examples depicting 3-10 objects arranged on a background, created using a catalog of 60 materials, and a further test set featuring 10k images created using 25 different materials. We benchmark a large set of recent unsupervised multi-object segmentation models on ClevrTex and find all state-of-the-art approaches fail to learn good representations in the textured setting, despite impressive performance on simpler data. We also create variants of the ClevrTex dataset, controlling for different aspects of scene complexity, and probe current approaches for individual shortcomings. Dataset and code are available at https://www.robots.ox.ac.uk/~vgg/research/clevrtex.
翻译:最近出现了一些旨在以不受监督的方式将图像分解和片段图像分解成多个对象的方法的激增, 即不受监督的多对象分割。 执行这样的任务是一个长期的计算机视觉目标, 提供在不要求密集的注释的情况下解开目标层次推理来训练分解模型。 尽管取得了显著进展, 当前的模型在视觉简单图像上开发和培训, 描绘着简单背景上的单色对象。 然而, 自然世界是视觉复杂的, 其缺点令人难以理解, 比如不同的纹理和复杂的照明效果。 在这次研究中, 我们提出了一个称为 ClevrTex 的新基准, 作为下一个比较、 评估和分析算法的挑战。 ClevrTex 以多种形状、 纹理和图片拼图材料的合成场景。 它包括50k 示例, 描述背景上排列的3- 10个对象, 使用60种材料的目录创建, 以及一个包含10k 图像的进一步测试集, 使用25种不同材料创建的图像。 我们用一系列不精确的多曲线/ 的多曲线图解路路段方法作为基准, 来比较的图解的图像模型,, 也用于在Clevlevleval- dislex ex 和Crevaltradealmmlex 。