Representation learning constructs low-dimensional representations to summarize essential features of high-dimensional data. This learning problem is often approached by describing various desiderata associated with learned representations; e.g., that they be non-spurious, efficient, or disentangled. It can be challenging, however, to turn these intuitive desiderata into formal criteria that can be measured and enhanced based on observed data. In this paper, we take a causal perspective on representation learning, formalizing non-spuriousness and efficiency (in supervised representation learning) and disentanglement (in unsupervised representation learning) using counterfactual quantities and observable consequences of causal assertions. This yields computable metrics that can be used to assess the degree to which representations satisfy the desiderata of interest and learn non-spurious and disentangled representations from single observational datasets.
翻译:代表制学习是用来总结高维数据基本特征的低维代表制,这种学习问题往往通过描述与学习表现相关的各种分层问题来解决,例如,它们不是纯净的、高效的或分解的,但是,将这些直观的分层转变为正式的标准,根据观察到的数据加以衡量和加强,可能具有挑战性;在本文件中,我们从因果角度看待代表性学习,将非纯洁和效率正规化(在监督的代表制学习中),并利用反事实数量和因果关系的可观察后果来说明分歧(未经监督的代言学习),从而得出可计算指标,用以评估陈述满足利益分层的程度,并从单一的观察数据集中学习非纯洁和分解的表述。