This paper aims to dive more deeply into various models available, including InceptionV3, InceptionResNetV2, MobileNetV2, and EfficientNetB7, using transfer learning to classify Japanese animation-style character faces. This paper has shown that EfficientNet-B7 has the highest accuracy rate with 85.08% top-1 Accuracy, followed by MobileNetV2, having a slightly less accurate result but with the benefits of much lower inference time and fewer number of required parameters. This paper also uses a few-shot learning framework, specifically Prototypical Networks, which produces decent results that can be used as an alternative to traditional transfer learning methods.
翻译:本文旨在更深入地探讨各种现有模型,包括InpenionV3、InptionionResNetV2、MobileNetV2、高效NetB7,利用转移学习对日本动画风格的性格进行分类,该文件表明,高效Net-B7的精确率最高,最高为85.08%的顶层-1准确度,其次是PaliveNetV2, 其结果略微不准确,但从低得多的推论时间和较少的所需参数中受益。本文还利用了几个短片的学习框架,特别是Protogram 网络,这些网络产生了体面的结果,可以替代传统的转移学习方法。