Multi-instance learning (MIL) deals with tasks where data is represented by a set of bags and each bag is described by a set of instances. Unlike standard supervised learning, only the bag labels are observed whereas the label for each instance is not available to the learner. Previous MIL studies typically follow the i.i.d. assumption, that the training and test samples are independently drawn from the same distribution. However, such assumption is often violated in real-world applications. Efforts have been made towards addressing distribution changes by importance weighting the training data with the density ratio between the training and test samples. Unfortunately, models often need to be trained without seeing the test distributions. In this paper we propose possibly the first framework for addressing distribution change in MIL without requiring access to the unlabeled test data. Our framework builds upon identifying a novel connection between MIL and the potential outcome framework in causal effect estimation. Experimental results on synthetic distribution change datasets, real-world datasets with synthetic distribution biases and real distributional biased image classification datasets validate the effectiveness of our approach.
翻译:多因子学习(MIL) 涉及由一组包包代表数据的任务,每个包包由一组实例描述。 与标准监督的学习不同, 只能观察到包包标签, 而每个包的标签没有提供给学习者。 以前的MIL研究通常遵循i. i. id. 假设, 即培训和测试样本是从同一分布中独立抽取的。 然而, 在现实应用中,这种假设常常被违反。 已经作出努力,通过对培训数据与培训样本和测试样本之间的密度比率进行重要加权, 解决分布变化。 不幸的是, 模型往往需要培训而不看到测试分布。 在本文件中,我们可能提出处理MIL分配变化的第一个框架, 不需要查阅无标签的测试数据。 我们的框架建立在确定MIL与因果关系估计潜在结果框架之间的新联系上。 合成分发变化数据集、 带有合成分布偏差和真实分布偏差图像分类数据集的实验结果证实了我们的方法的有效性。