In human learning, an effective learning methodology is small-group learning: a small group of students work together towards the same learning objective, where they express their understanding of a topic to their peers, compare their ideas, and help each other to trouble-shoot problems. In this paper, we aim to investigate whether this human learning method can be borrowed to train better machine learning models, by developing a novel ML framework -- small-group learning (SGL). In our framework, a group of learners (ML models) with different model architectures collaboratively help each other to learn by leveraging their complementary advantages. Specifically, each learner uses its intermediately trained model to generate a pseudo-labeled dataset and re-trains its model using pseudo-labeled datasets generated by other learners. SGL is formulated as a multi-level optimization framework consisting of three learning stages: each learner trains a model independently and uses this model to perform pseudo-labeling; each learner trains another model using datasets pseudo-labeled by other learners; learners improve their architectures by minimizing validation losses. An efficient algorithm is developed to solve the multi-level optimization problem. We apply SGL for neural architecture search. Results on CIFAR-100, CIFAR-10, and ImageNet demonstrate the effectiveness of our method.
翻译:在人类学习中,有效的学习方法是小群体学习:一小群学生为同一学习目标而共同努力:一小群学生向同龄人表示对一个主题的理解,比较他们的想法,并互相帮助解决难题。在本文件中,我们的目标是研究是否可以借用这种人类学习方法来训练更好的机器学习模式,方法是开发一个全新的ML框架 -- -- 小型群体学习(SGL)。在我们的框架内,一组具有不同模型结构的学习者(ML模型)通过利用他们互补的优势相互协作帮助学习。具体地说,每个学习者利用其中级培训模型,利用由其他学习者生成的假标签数据集来生成一个假标签数据集,并重新编织模型。在三个学习阶段组成一个多层次的优化框架:每个学习者独立地培训一个模型,并使用这个模型进行假标签。在我们的框架中,每个学习者都用假标签的数据集来培养另一个模型;每个学习者通过尽量减少验证损失来改进他们的结构。一个高效的算法是用来解决多层次的优化问题。我们用FAR、CIS-10的图像搜索方法。