Learning from multimodal data is an important research topic in machine learning, which has the potential to obtain better representations. In this work, we propose a novel approach to generative modeling of multimodal data based on generative adversarial networks. To learn a coherent multimodal generative model, we show that it is necessary to align different encoder distributions with the joint decoder distribution simultaneously. To this end, we construct a specific form of the discriminator to enable our model to utilize data efficiently, which can be trained constrastively. By taking advantage of contrastive learning through factorizing the discriminator, we train our model on unimodal data. We have conducted experiments on the benchmark datasets, whose promising results show that our proposed approach outperforms the-state-of-the-art methods on a variety of metrics. The source code will be made publicly available.
翻译:从多式联运数据中学习是机器学习的一个重要研究课题,它有可能得到更好的表述。在这项工作中,我们提出了一种基于基因对抗网络的多式联运数据基因模型化新颖方法。为了学习一种连贯的多式联运基因模型,我们表明有必要同时将不同的编码器分布与联合解码器分布相协调。为此,我们构建了一种歧视者的具体形式,使我们的模型能够有效地利用数据,这些数据可以经过严格的培训。通过将歧视者因素化来利用对比性学习,我们用单式数据来培训我们的模型。我们已经在基准数据集上进行了实验,这些实验有希望的结果表明,我们所提议的方法超越了各种计量标准的最新方法。源代码将予以公布。