General image super-resolution techniques have difficulties in recovering detailed face structures when applying to low resolution face images. Recent deep learning based methods tailored for face images have achieved improved performance by jointly trained with additional task such as face parsing and landmark prediction. However, multi-task learning requires extra manually labeled data. Besides, most of the existing works can only generate relatively low resolution face images (e.g., $128\times128$), and their applications are therefore limited. In this paper, we introduce a novel SPatial Attention Residual Network (SPARNet) built on our newly proposed Face Attention Units (FAUs) for face super-resolution. Specifically, we introduce a spatial attention mechanism to the vanilla residual blocks. This enables the convolutional layers to adaptively bootstrap features related to the key face structures and pay less attention to those less feature-rich regions. This makes the training more effective and efficient as the key face structures only account for a very small portion of the face image. Visualization of the attention maps shows that our spatial attention network can capture the key face structures well even for very low resolution faces (e.g., $16\times16$). Quantitative comparisons on various kinds of metrics (including PSNR, SSIM, identity similarity, and landmark detection) demonstrate the superiority of our method over current state-of-the-arts. We further extend SPARNet with multi-scale discriminators, named as SPARNetHD, to produce high resolution results (i.e., $512\times512$). We show that SPARNetHD trained with synthetic data cannot only produce high quality and high resolution outputs for synthetically degraded face images, but also show good generalization ability to real world low quality face images. Codes are available at \url{https://github.com/chaofengc/Face-SPARNet}.
翻译:普通图像超分辨率技术在应用低分辨率网面图像时很难恢复详细的面部结构。最近为脸部图像专门设计的深层次学习方法通过联合培训,通过面部剖析和里程碑式预测等额外任务,提高了业绩。然而,多任务学习需要额外的手工标签数据。此外,大多数现有作品只能产生相对较低的分辨率图像(例如12美元12\time128美元),因此其应用有限。在本文中,我们推出一个新颖的低斜度关注网络(SPARNet),建在我们新提议的面部关注股(FAUs)上,用于面部超分辨率。具体来说,我们为香草残留区引入了空间关注机制。这使得变异层能够适应与关键面部结构相关的布局功能功能功能,而较少注意那些功能丰富区域。这使得培训效果和效率更高,因为关键面部结构仅代表了很小的面部图像。我们现有的空间关注网络无法捕捉到关键面部结构,但即使分辨率非常低的面部(e.g. g., 16\ mestal realal assalalalalalal ressal laveal laveal dal dal dal dal dal laveald) 也展示了我们的S-de 数据。