Image synthesis and image recognition have witnessed remarkable progress, but often at the expense of computationally expensive training and inference. Learning lightweight yet expressive deep model has emerged as an important and interesting direction. Inspired by the well-known split-transform-aggregate design heuristic in the Inception building block, this paper proposes a Skip-Layer Inception Module (SLIM) that facilitates efficient learning of image synthesis models, and a same-layer variant (dubbed as SLIM too) as a stronger alternative to the well-known ResNeXts for image recognition. In SLIM, the input feature map is first split into a number of groups (e.g., 4).Each group is then transformed to a latent style vector(via channel-wise attention) and a latent spatial mask (via spatial attention). The learned latent masks and latent style vectors are aggregated to modulate the target feature map. For generative learning, SLIM is built on a recently proposed lightweight Generative Adversarial Networks (i.e., FastGANs) which present a skip-layer excitation(SLE) module. For few-shot image synthesis tasks, the proposed SLIM achieves better performance than the SLE work and other related methods. For one-shot image synthesis tasks, it shows stronger capability of preserving images structures than prior arts such as the SinGANs. For image classification tasks, the proposed SLIM is used as a drop-in replacement for convolution layers in ResNets (resulting in ResNeXt-like models) and achieves better accuracy in theImageNet-1000 dataset, with significantly smaller model complexity
翻译:图像合成和图像识别取得了显著的进展,但往往以计算成本昂贵的培训和推断为代价。学习轻量但表现深刻的模型已经成为一个重要和有趣的方向。在受感应建筑块中众所周知的分变式聚合设计超常性激素的启发下,本文件提议了一个跳过-拉叶感应模块(SLIM),该模块有助于高效学习图像合成模型,以及一个同级变量(与 SLIM 一同浸泡),作为人们熟知的 ResNeX 分类识别图像的更强替代品。在 SLIM 中,输入特征图首先分为若干组(例如, 4 ) 。然后,Each 组转换为潜伏风格矢量式( 通过频道关注) 和潜伏空间遮罩(通过空间关注 ) 。 学到的潜伏遮罩和潜伏风格矢量矢量矢量矩阵将组合组合成一个比S-LEAR 图像更强的图像替换模型(例如, FastGANs ) 。 SL 的拟议SL 将S- hold 的S- relishal imal relishall 的图像转换模型显示一个比SL 之前的SL 相关的SL 格式化模型更强的SL) 。