Great diversity and photorealism have been achieved by unconditional GAN frameworks such as StyleGAN and its variations. In the meantime, persistent efforts have been made to enhance the semantic controllability of StyleGANs. For example, a dozen of style manipulation methods have been recently proposed to perform attribute-conditioned style editing. Although some of these methods work well in manipulating the style codes along one attribute, the control accuracy when jointly manipulating multiple attributes tends to be problematic. To address these limitations, we propose a Dynamic Style Manipulation Network (DyStyle) whose structure and parameters vary by input samples, to perform nonlinear and adaptive manipulation of latent codes for flexible and precise attribute control. Additionally, a novel easy-to-hard training procedure is introduced for efficient and stable training of the DyStyle network. Extensive experiments have been conducted on faces and other objects. As a result, our approach demonstrates fine-grained disentangled edits along multiple numeric and binary attributes. Qualitative and quantitative comparisons with existing style manipulation methods verify the superiority of our method in terms of the attribute control accuracy and identity preservation without compromising the photorealism. The advantage of our method is even more significant for joint multi-attribute control. The source codes are made publicly available at \href{https://github.com/phycvgan/DyStyle}{phycvgan/DyStyle}.
翻译:无条件的 GAN 框架,如StyleGAN 及其变异,实现了巨大的多样性和摄影现实。与此同时,我们不断努力加强StyleGANs 的语义控制。例如,最近提出了十多种风格操纵方法,以进行属性化的样式编辑。虽然其中一些方法在按照一个属性调控样式代码方面效果良好,但联合操控多个属性时的控制精度往往有问题。为了解决这些局限性,我们提议建立一个动态样式操纵网络(DyStyStyle),其结构和参数因输入样本而异,对潜在代码进行非线性和适应性操纵,以进行灵活和精确的属性控制。此外,为了高效和稳定地培训DyStyStyle网络,引入了一套新颖的简单到硬化的培训程序。对面部和其他对象进行了广泛的实验。结果是,我们的方法显示在多个数字和二进制属性的细微分解的编辑。与现有风格调控法方法的定性和定量比较,以核实我们方法在属性控制方面是否优劣的精准性和适应性。Stencial-realalalalalal-dealal spress 维护了我们现有的多来源。