The current dominant paradigm when building a machine learning model is to iterate over a dataset over and over until convergence. Such an approach is non-incremental, as it assumes access to all images of all categories at once. However, for many applications, non-incremental learning is unrealistic. To that end, researchers study incremental learning, where a learner is required to adapt to an incoming stream of data with a varying distribution while preventing forgetting of past knowledge. Significant progress has been made, however, the vast majority of works focus on the fully supervised setting, making these algorithms label-hungry thus limiting their real-life deployment. To that end, in this paper, we make the first attempt to survey recently growing interest in label-efficient incremental learning. We identify three subdivisions, namely semi-, few-shot- and self-supervised learning to reduce labeling efforts. Finally, we identify novel directions that can further enhance label-efficiency and improve incremental learning scalability. Project website: {https://github.com/kilickaya/label-efficient-il.
翻译:目前,在建立机器学习模式时,主流模式是反复重复一个数据集,直至趋同。这种方法是非渐进式的,因为它可以同时获取所有类别的所有图像。然而,在许多应用中,非循环式学习是不现实的。为此,研究人员研究渐进式学习,需要一位学习者适应不同分布的数据流,同时防止忘记过去的知识。然而,已经取得重大进展,绝大多数工作都集中在完全监督下的设置上,使这些算法标签饥饿限制其真实生活的部署。为此,我们首次尝试调查最近对标签效率增量学习的兴趣日益增加的情况。我们确定了三个子部门,即半、少照和自我监督的学习,以减少标签工作。最后,我们确定了能够进一步提高标签效率和增加增量学习能力的新方向。项目网站: {https://github.com/kilikaya/lab-effective-il。