Since its inception, the neural estimation of mutual information (MI) has demonstrated the empirical success of modeling expected dependency between high-dimensional random variables. However, MI is an aggregate statistic and cannot be used to measure point-wise dependency between different events. In this work, instead of estimating the expected dependency, we focus on estimating point-wise dependency (PD), which quantitatively measures how likely two outcomes co-occur. We show that we can naturally obtain PD when we are optimizing MI neural variational bounds. However, optimizing these bounds is challenging due to its large variance in practice. To address this issue, we develop two methods (free of optimizing MI variational bounds): Probabilistic Classifier and Density-Ratio Fitting. We demonstrate the effectiveness of our approaches in 1) MI estimation, 2) self-supervised representation learning, and 3) cross-modal retrieval task.
翻译:自建立以来,对相互信息的神经估计(MI)表明,模拟高维随机变量之间预期依赖性的经验性成功。然而,MI是一个综合统计,不能用来衡量不同事件之间的点度依赖性。在这项工作中,我们重点不是估计预期依赖性,而是估计点向依赖性(PD),从数量上衡量两种结果的可能性。我们显示,当我们优化了神经质变界限时,我们自然可以获得PD。然而,优化这些界限因其在实践上的巨大差异而具有挑战性。为了解决这一问题,我们制定了两种方法(无优化MI变异界限):概率分类和密度-拉蒂奥调适。我们展示了我们的方法的有效性,1)MI估计,2)自我超强的代议式学习,3)跨模式的检索任务。