The use of synthetic (or simulated) data for training machine learning models has grown rapidly in recent years. Synthetic data can often be generated much faster and more cheaply than its real-world counterpart. One challenge of using synthetic imagery however is scene design: e.g., the choice of content and its features and spatial arrangement. To be effective, this design must not only be realistic, but appropriate for the target domain, which (by assumption) is unlabeled. In this work, we propose an approach to automatically choose the design of synthetic imagery based upon unlabeled real-world imagery. Our approach, termed Neural-Adjoint Meta-Simulation (NAMS), builds upon the seminal recent meta-simulation approaches. In contrast to the current state-of-the-art methods, our approach can be pre-trained once offline, and then provides fast design inference for new target imagery. Using both synthetic and real-world problems, we show that NAMS infers synthetic designs that match both the in-domain and out-of-domain target imagery, and that training segmentation models with NAMS-designed imagery yields superior results compared to na\"ive randomized designs and state-of-the-art meta-simulation methods.
翻译:合成(或模拟)数据用于培训机器学习模型的使用近年来迅速发展。合成数据往往比真实世界的模拟数据更快、更便宜。使用合成图像的一个挑战是场景设计:例如内容及其特征的选择和空间安排。要有效,这种设计不仅必须现实,而且要适合目标领域,因为(假设)没有标签。在这项工作中,我们建议一种办法,根据未标记的真实世界图像自动选择合成图像的设计。我们的方法,称为神经-Ad联合元模拟(NAMS),以近期的半成像模拟方法为基础。与目前的最新最新元模拟方法相比,我们的方法可以先离线训练一次,然后为新的目标图像提供快速设计推导。我们用合成和现实世界的问题来推断合成图像的设计,既符合内部的和外部的目标图像,又能与培训分解模型相比,与NASS设计成型图像的高级模型和高级成型成型模型相比。