Self-supervised representation learning methods aim to provide powerful deep feature learning without the requirement of large annotated datasets, thus alleviating the annotation bottleneck that is one of the main barriers to practical deployment of deep learning today. These methods have advanced rapidly in recent years, with their efficacy approaching and sometimes surpassing fully supervised pre-training alternatives across a variety of data modalities including image, video, sound, text and graphs. This article introduces this vibrant area including key concepts, the four main families of approach and associated state of the art, and how self-supervised methods are applied to diverse modalities of data. We further discuss practical considerations including workflows, representation transferability, and compute cost. Finally, we survey the major open challenges in the field that provide fertile ground for future work.
翻译:自我监督的代表学习方法旨在提供强大的深层次特征学习,而不需要大量附加说明的数据集,从而减轻说明瓶颈,这是当今实际部署深层学习的主要障碍之一,近年来,这些方法取得了迅速进展,其效力接近,有时超过充分监督的培训前选择,涉及各种数据模式,包括图像、视频、声音、文本和图表。本文章介绍了这一充满活力的领域,包括关键概念、方法的四大组合和相关的艺术状态,以及如何将自我监督的方法应用于不同的数据模式。我们进一步讨论了包括工作流程、代表性可转移性和成本计算在内的实际考虑。最后,我们调查了为未来工作提供肥沃土壤的实地主要公开挑战。