Few-shot image generation aims to generate data of an unseen category based on only a few samples. Apart from basic content generation, a bunch of downstream applications hopefully benefit from this task, such as low-data detection and few-shot classification. To achieve this goal, the generated images should guarantee category retention for classification beyond the visual quality and diversity. In our preliminary work, we present an ``editing-based'' framework Attribute Group Editing (AGE) for reliable few-shot image generation, which largely improves the generation performance. Nevertheless, AGE's performance on downstream classification is not as satisfactory as expected. This paper investigates the class inconsistency problem and proposes Stable Attribute Group Editing (SAGE) for more stable class-relevant image generation. SAGE takes use of all given few-shot images and estimates a class center embedding based on the category-relevant attribute dictionary. Meanwhile, according to the projection weights on the category-relevant attribute dictionary, we can select category-irrelevant attributes from the similar seen categories. Consequently, SAGE injects the whole distribution of the novel class into StyleGAN's latent space, thus largely remains the category retention and stability of the generated images. Going one step further, we find that class inconsistency is a common problem in GAN-generated images for downstream classification. Even though the generated images look photo-realistic and requires no category-relevant editing, they are usually of limited help for downstream classification. We systematically discuss this issue from both the generative model and classification model perspectives, and propose to boost the downstream classification performance of SAGE by enhancing the pixel and frequency components.
翻译:少量图像生成旨在生成一个仅基于少数样本的隐蔽类别的数据。 除了基本内容生成之外, 下游应用程序有望从这一任务中受益, 比如低频检测和少发分类。 为了实现这一目标, 生成的图像应该保证在视觉质量和多样性之外保留分类。 在初步工作中, 我们为可靠的少发图像生成提供“ 编辑基础” 框架属性组编辑( AGE ), 这在很大程度上改善了生成性能。 然而, AGE 在下游分类上的性能不如预期的令人满意。 本文调查了类不一致问题, 并提出了更稳定的类相关图像生成 SAGE 。 SAGE 使用所有给定的少量图像并估算基于类相关属性字典的分类中心。 同时, 根据对与类别相关的属性字典的预测权重, 我们可以从相似的类别中选择与类别相关的属性模型。 因此, SAGAG 将新类的分布推算为更深层空间, 并提出了更稳定的属性组编辑( SAG ) 。 因此, 将常规的图像分类的排序和稳定性分类通常需要从 GAN 的分类中 的分类 的分类 。