In this work, we propose Reciprocal Distribution Alignment (RDA) to address semi-supervised learning (SSL), which is a hyperparameter-free framework that is independent of confidence threshold and works with both the matched (conventionally) and the mismatched class distributions. Distribution mismatch is an often overlooked but more general SSL scenario where the labeled and the unlabeled data do not fall into the identical class distribution. This may lead to the model not exploiting the labeled data reliably and drastically degrade the performance of SSL methods, which could not be rescued by the traditional distribution alignment. In RDA, we enforce a reciprocal alignment on the distributions of the predictions from two classifiers predicting pseudo-labels and complementary labels on the unlabeled data. These two distributions, carrying complementary information, could be utilized to regularize each other without any prior of class distribution. Moreover, we theoretically show that RDA maximizes the input-output mutual information. Our approach achieves promising performance in SSL under a variety of scenarios of mismatched distributions, as well as the conventional matched SSL setting. Our code is available at: https://github.com/NJUyued/RDA4RobustSSL.
翻译:在这项工作中,我们提议对等分配调整,以解决半监督的学习(SSL)问题,这是一个独立于信任阈值的超度无参数框架,与匹配(常规)和不匹配类分布的匹配(常规)和不匹配类分配同时工作。分配不匹配是一个经常被忽视但更一般的 SSL假设,标签和未贴标签数据不归同类分配。这可能导致模型不利用标签数据和标签数据可靠地大幅降低SLS方法的性能,而传统分配比对无法挽救这种性能。在RDA中,我们对预测伪标签和补充标签的两种分类者预测的分布实行对等调整。这两个包含补充信息的分配可以用来在不事先分配类分配的情况下相互规范。此外,我们理论上表明RDA最大限度地利用输入-输出的相互信息。我们的方法在各种不匹配的分布假设情景下,以及在常规匹配的SSLF设置下,SLS(常规匹配的SLV)取得了有希望的业绩。我们的代码可以查到: https://giuth.