Face sketch synthesis has been widely used in multi-media entertainment and law enforcement. Despite the recent developments in deep neural networks, accurate and realistic face sketch synthesis is still a challenging task due to the diversity and complexity of human faces. Current image-to-image translation-based face sketch synthesis frequently encounters over-fitting problems when it comes to small-scale datasets. To tackle this problem, we present an end-to-end Memory Oriented Style Transfer Network (MOST-Net) for face sketch synthesis which can produce high-fidelity sketches with limited data. Specifically, an external self-supervised dynamic memory module is introduced to capture the domain alignment knowledge in the long term. In this way, our proposed model could obtain the domain-transfer ability by establishing the durable relationship between faces and corresponding sketches on the feature level. Furthermore, we design a novel Memory Refinement Loss (MR Loss) for feature alignment in the memory module, which enhances the accuracy of memory slots in an unsupervised manner. Extensive experiments on the CUFS and the CUFSF datasets show that our MOST-Net achieves state-of-the-art performance, especially in terms of the Structural Similarity Index(SSIM).
翻译:在多媒体娱乐和执法中广泛使用了面部素描合成。尽管在深神经网络中最近有所发展,但准确和现实的面部素描合成仍是一项艰巨的任务,因为人类面孔的多样性和复杂性。当前图像到图像的面部素描合成在小规模数据集方面经常遇到过大的问题。为了解决这个问题,我们提出了一个端到端的内存导向样式传输网络(MOST-Net),用于面部素描合成,该面部素描可产生高不洁的面部素描草。具体地说,引入了外部自我监督的动态记忆模块,以长期捕捉域对齐知识。通过这种方式,我们提议的模型可以通过在面部和相应的图谱层上建立持久的关系,获得域转移能力。此外,我们设计了一个新的记忆模块的内存精度调整损失(MRML),以不超缩的方式提高记忆位置的准确性能。关于CUFS和CUFSFSF数据集的广泛实验显示,特别是我们的MOST-Net状态性能指数。