Learning discriminative representation using large-scale face datasets in the wild is crucial for real-world applications, yet it remains challenging. The difficulties lie in many aspects and this work focus on computing resource constraint and long-tailed class distribution. Recently, classification-based representation learning with deep neural networks and well-designed losses have demonstrated good recognition performance. However, the computing and memory cost linearly scales up to the number of identities (classes) in the training set, and the learning process suffers from unbalanced classes. In this work, we propose a dynamic class queue (DCQ) to tackle these two problems. Specifically, for each iteration during training, a subset of classes for recognition are dynamically selected and their class weights are dynamically generated on-the-fly which are stored in a queue. Since only a subset of classes is selected for each iteration, the computing requirement is reduced. By using a single server without model parallel, we empirically verify in large-scale datasets that 10% of classes are sufficient to achieve similar performance as using all classes. Moreover, the class weights are dynamically generated in a few-shot manner and therefore suitable for tail classes with only a few instances. We show clear improvement over a strong baseline in the largest public dataset Megaface Challenge2 (MF2) which has 672K identities and over 88% of them have less than 10 instances. Code is available at https://github.com/bilylee/DCQ
翻译:使用野生大规模脸形数据集进行学习的歧视性表现方式对于现实世界应用至关重要,但它仍然具有挑战性。 困难在于许多方面, 这项工作的重点是计算资源限制和长尾类分布。 最近, 与深神经网络和设计完善的损失进行基于分类的代表性学习, 显示了良好的认知性表现。 然而, 计算和记忆成本线度的尺度, 以培训组的身份( 类) 数量为限, 以及学习过程存在不平衡的等级。 在这项工作中, 我们提议一个动态的类排( DCQ) 来应对这两个问题。 具体地说, 培训期间的反复出现, 一组确认课是动态选择的, 其类重量是动态地在飞行上生成的, 储存在队列中。 由于每个循环网络只选择了一组课程, 计算要求就减少了。 通过使用一个没有模型平行的单一服务器, 我们从经验中核实了大规模数据集, 10%的班级足以实现与所有班级相似的性能。 此外, 班级重量是动态生成的, 以几发式方式产生的, 类重量是动态产生的, 类重量是动态产生的, 在飞行上产生的, 因此, 类重量是清晰的排序中, 级的排序 级的排序 格式只有多少个基准 。