Mutual Information (MI) and Conditional Mutual Information (CMI) are multi-purpose tools from information theory that are able to naturally measure the statistical dependencies between random variables, thus they are usually of central interest in several statistical and machine learning tasks, such as conditional independence testing and representation learning. However, estimating CMI, or even MI, is infamously challenging due the intractable formulation. In this study, we introduce DINE (Diffeomorphic Information Neural Estimator)-a novel approach for estimating CMI of continuous random variables, inspired by the invariance of CMI over diffeomorphic maps. We show that the variables of interest can be replaced with appropriate surrogates that follow simpler distributions, allowing the CMI to be efficiently evaluated via analytical solutions. Additionally, we demonstrate the quality of the proposed estimator in comparison with state-of-the-arts in three important tasks, including estimating MI, CMI, as well as its application in conditional independence testing. The empirical evaluations show that DINE consistently outperforms competitors in all tasks and is able to adapt very well to complex and high-dimensional relationships.
翻译:相互信息(MI)和有条件的相互信息(CMI)是信息理论的多用途工具,能够自然地测量随机变量之间的统计依赖性,因此这些变量通常对若干统计和机器学习任务具有核心意义,例如有条件的独立测试和代表性学习;然而,估计CMI,甚至MI,由于这一棘手的表述方式,是难以为人知的。在本研究中,我们引入了DINE(Diffomodical Information Enoralstimator)-一种新颖的方法,用以估计CMI的连续随机变量,这是受CMI对地貌相貌图的异性所启发的。我们表明,在各种任务中,DINE始终超越竞争者,能够很好地适应复杂和高维度的关系。