The author of this work proposes an overview of the recent semi-supervised learning approaches and related works. Despite the remarkable success of neural networks in various applications, there exist few formidable constraints including the need for a large amount of labeled data. Therefore, semi-supervised learning, which is a learning scheme in which the scarce labels and a larger amount of unlabeled data are utilized to train models (e.g., deep neural networks) is getting more important. Based on the key assumptions of semi-supervised learning, which are the manifold assumption, cluster assumption, and continuity assumption, the work reviews the recent semi-supervised learning approaches. In particular, the methods in regard to using deep neural networks in a semi-supervised learning setting are primarily discussed. In addition, the existing works are first classified based on the underlying idea and explained, and then the holistic approaches that unify the aforementioned ideas are detailed.
翻译:这项工作的作者建议了对最近的半监督学习方法和相关作品的概述。尽管神经网络在各种应用中取得了显著的成功,但几乎没有什么巨大的制约因素,包括需要大量贴标签的数据。因此,半监督学习计划,即利用稀缺标签和大量未贴标签的数据来培训模型(例如深层神经网络),根据半监督学习的关键假设,即多重假设、集群假设和连续性假设,工作审查了最近的半监督学习方法。特别是,主要讨论了在半监督学习环境中使用深神经网络的方法。此外,现有工作首先根据基本理念进行分类,并作出解释,然后详细列出统一上述理念的整体方法。