Automatic transfer of text between domains has become popular in recent times. One of its aims is to preserve the semantic content of text being translated from source to target domain. However, it does not explicitly maintain other attributes between the source and translated text, for e.g., text length and descriptiveness. Maintaining constraints in transfer has several downstream applications, including data augmentation and de-biasing. We introduce a method for such constrained unsupervised text style transfer by introducing two complementary losses to the generative adversarial network (GAN) family of models. Unlike the competing losses used in GANs, we introduce cooperative losses where the discriminator and the generator cooperate and reduce the same loss. The first is a contrastive loss and the second is a classification loss, aiming to regularize the latent space further and bring similar sentences across domains closer together. We demonstrate that such training retains lexical, syntactic, and domain-specific constraints between domains for multiple benchmark datasets, including ones where more than one attribute change. We show that the complementary cooperative losses improve text quality, according to both automated and human evaluation measures.
翻译:近期来,在不同域间自动传送文本变得十分普遍,其目的之一是保持从源到目标域翻译文本的语义内容,但没有明确保持源与翻译文本之间的其他属性,例如文字长度和描述性。保持对传输的限制有几种下游应用,包括数据增强和减少偏差。我们采用了一种方法,通过对模型的基因对抗网络(GAN)系列引入两个补充性损失来限制这种不受监督的文本样式转移。与在GANs中使用的相竞损失不同,我们引入了歧视者和生成者合作并减少相同损失的合作性损失。第一个是对比性损失,第二个是分类性损失,目的是进一步规范潜在空间,使跨域的类似句子更加接近。我们证明,这种培训在多个基准数据集的域间保留了词汇、合成和特定域内的限制,包括不止一个属性变化的域间。我们表明,根据自动化和人文评估措施,补充性合作损失提高了文本质量。