Contextualized word representations, such as ELMo and BERT, were shown to perform well on various semantic and syntactic tasks. In this work, we tackle the task of unsupervised disentanglement between semantics and structure in neural language representations: we aim to learn a transformation of the contextualized vectors, that discards the lexical semantics, but keeps the structural information. To this end, we automatically generate groups of sentences which are structurally similar but semantically different, and use metric-learning approach to learn a transformation that emphasizes the structural component that is encoded in the vectors. We demonstrate that our transformation clusters vectors in space by structural properties, rather than by lexical semantics. Finally, we demonstrate the utility of our distilled representations by showing that they outperform the original contextualized representations in a few-shot parsing setting.
翻译:ELMO 和 BERT 等背景化文字表达方式在各种语义和合成任务上表现良好。 在这项工作中,我们处理神经语言表达方式的语义和结构之间未经监督的分解任务:我们的目标是学习背景化矢量的转变,这种转变抛弃了词汇性语义,但保留结构信息。为此,我们自动生成在结构上相似但在语义上有所不同的一组句子,并使用衡量学习方法来学习一种强调矢量中编码的结构组成部分的转变。我们通过结构属性而不是词汇性语义来证明我们在空间的变异集群矢量。最后,我们通过显示这些变异的表达方式在微小的剖析环境中比原始背景化的表达方式要强。