We investigate the generalization properties of a self-training algorithm with halfspaces. The approach learns a list of halfspaces iteratively from labeled and unlabeled training data, in which each iteration consists of two steps: exploration and pruning. In the exploration phase, the halfspace is found sequentially by maximizing the unsigned-margin among unlabeled examples and then assigning pseudo-labels to those that have a distance higher than the current threshold. The pseudo-labeled examples are then added to the training set, and a new classifier is learned. This process is repeated until no more unlabeled examples remain for pseudo-labeling. In the pruning phase, pseudo-labeled samples that have a distance to the last halfspace greater than the associated unsigned-margin are then discarded. We prove that the misclassification error of the resulting sequence of classifiers is bounded and show that the resulting semi-supervised approach never degrades performance compared to the classifier learned using only the initial labeled training set. Experiments carried out on a variety of benchmarks demonstrate the efficiency of the proposed approach compared to state-of-the-art methods.
翻译:我们用半空来调查半空自我培训算法的一般特性。 这种方法从标签和无标签的培训数据中反复学习半空的列表, 其中每个迭代由两个步骤组成: 勘探和修剪。 在勘探阶段, 通过在未贴标签的示例中尽量扩大未签名的边距, 并随后为那些距离高于当前阈值的人分配假标签。 然后将假标签示例添加到培训集中, 并学习一个新的分类器。 这一过程会重复, 直到没有更多未贴标签的例子留在伪标签的培训数据中。 在运行阶段, 假标签样本与上半空的距离大于相关未签名的边距, 然后被丢弃。 我们证明, 由此产生的分类器序列的分类错误是被捆绑的, 并表明, 由此产生的半超标法方法不会降低与仅使用初始标签培训集的分类器相比的性能。 在各种基准上进行的实验显示了拟议方法相对于状态图方法的效率 。