We propose generative multitask learning (GMTL), a simple and scalable approach to causal representation learning for multitask learning. Our approach makes a minor change to the conventional multitask inference objective, and improves robustness to target shift. Since GMTL only modifies the inference objective, it can be used with existing multitask learning methods without requiring additional training. The improvement in robustness comes from mitigating unobserved confounders that cause the targets, but not the input. We refer to them as \emph{target-causing confounders}. These confounders induce spurious dependencies between the input and targets. This poses a problem for conventional multitask learning, due to its assumption that the targets are conditionally independent given the input. GMTL mitigates target-causing confounding at inference time, by removing the influence of the joint target distribution, and predicting all targets jointly. This removes the spurious dependencies between the input and targets, where the degree of removal is adjustable via a single hyperparameter. This flexibility is useful for managing the trade-off between in- and out-of-distribution generalization. Our results on the Attributes of People and Taskonomy datasets reflect an improved robustness to target shift across four multitask learning methods.
翻译:我们提出基因化多任务学习(GMTL),这是为多任务学习而进行因果代表学习的一种简单且可扩缩的方法。我们的方法对常规的多任务推导目标稍有改变,提高了目标转移的力度。由于GMTL只改变推论目标,因此可以与现有的多任务学习方法一起使用,而不需要额外的培训。强度的提高来自减轻导致目标的未观察到的混乱因素,而不是投入。我们称之为输入和目标之间的不明显依赖性,通过单一的超分度来调整清除的程度。这给传统的多任务学习带来了问题,因为它假定根据投入的情况,目标是有条件的。 GMTL通过消除联合目标分布的影响,并联合预测所有目标,减轻目标之间的不明显依赖性。这可以消除投入和目标之间的不明显依赖性,通过单一的超分度来调整清除程度。这种灵活性对于传统的多任务学习,因为根据其假设,目标是有条件的,可以反映我们在四大任务分配中的变化。