Automatic image cropping algorithms aim to recompose images like human-being photographers by generating the cropping boxes with improved composition quality. Cropping box regression approaches learn the beauty of composition from annotated cropping boxes. However, the bias of annotations leads to quasi-trivial recomposing results, which has an obvious tendency to the average location of training samples. The crux of this predicament is that the task is naively treated as a box regression problem, where rare samples might be dominated by normal samples, and the composition patterns of rare samples are not well exploited. Observing that similar composition patterns tend to be shared by the cropping boundaries annotated nearly, we argue to find the beauty of composition from the rare samples by clustering the samples with similar cropping boundary annotations, ie, similar composition patterns. We propose a novel Contrastive Composition Clustering (C2C) to regularize the composition features by contrasting dynamically established similar and dissimilar pairs. In this way, common composition patterns of multiple images can be better summarized, which especially benefits the rare samples and endows our model with better generalizability to render nontrivial results. Extensive experimental results show the superiority of our model compared with prior arts. We also illustrate the philosophy of our design with an interesting analytical visualization.
翻译:自动图像裁剪算法旨在通过生成成份质量更好的作物盒来重新配置像人类摄影师这样的图像。 裁剪盒回归法从附加说明的作物盒中学习成份的美貌。 但是, 说明的偏差导致半三重再组合结果, 这明显倾向于培训样本的平均位置。 这种困境的症结在于, 任务被天真地当作一个盒式回归问题, 即稀有样本可能由正常样本占上风, 稀有样本的构成模式没有得到很好的利用。 观察类似的成份模式往往被近乎注解的作物框所共有, 我们争论的是, 通过将样品与相似的作物边界说明、 类似、 相似的成份模式组合, 从稀有样本中找到成份的美貌。 我们提出一个新的相矛盾的成份组合( C2C), 通过对比动态建立的相似和不相似的两对组合, 来规范成份的成份。 这样就可以更好地总结多种图像的常见成份模式, 特别有利于稀有的样品和最终的模型, 使我们的模型更具有较普通的通用性, 使非令人厌异的结果成为非端。 我们的实验性的哲学的模型也展示了我们之前的视觉的视觉性模型展示。 我们的模型展示了我们之前的视觉的模型的模型的图像的优越性。