Automatic generation of high-quality Chinese fonts from a few online training samples is a challenging task, especially when the amount of samples is very small. Existing few-shot font generation methods can only synthesize low-resolution glyph images that often possess incorrect topological structures or/and incomplete strokes. To address the problem, this paper proposes FontTransformer, a novel few-shot learning model, for high-resolution Chinese glyph image synthesis by using stacked Transformers. The key idea is to apply the parallel Transformer to avoid the accumulation of prediction errors and utilize the serial Transformer to enhance the quality of synthesized strokes. Meanwhile, we also design a novel encoding scheme to feed more glyph information and prior knowledge to our model, which further enables the generation of high-resolution and visually-pleasing glyph images. Both qualitative and quantitative experimental results demonstrate the superiority of our method compared to other existing approaches in the few-shot Chinese font synthesis task.
翻译:从几个在线培训样本中自动生成高质量的中国字体是一项艰巨的任务,尤其是在样本数量非常小的情况下。现有的微小字体生成方法只能合成往往具有不正确的地形结构或/和不完全划线的低分辨率胶片。为了解决这个问题,本文提议了Font Transformexed,这是一个新颖的微小的学习模型,用于使用堆叠变形器对中国高分辨率胶片进行合成。关键的想法是应用平行变形器避免预测错误的累积,并利用序列变形器提高合成划线的质量。同时,我们还设计了一个新颖的编码方案,为模型提供更多的胶片信息和先前的知识,从而进一步使得能够生成高分辨率和直观的胶片图像。定性和定量实验结果都表明我们的方法优于少数中国字体合成任务中的其他现有方法。