Partial information decomposition (PID) of the multivariate mutual information describes the distinct ways in which a set of source variables contains information about a target variable. The groundbreaking work of Williams and Beer has shown that this decomposition cannot be determined from classic information theory without making additional assumptions, and several candidate measures have been proposed, often drawing on principles from related fields such as decision theory. None of these measures is differentiable with respect to the underlying probability mass function. We here present a novel measure that satisfies this property, emerges solely from information-theoretic principles, and has the form of a local mutual information. We show how the measure can be understood from the perspective of exclusions of probability mass, a principle that is foundational to the original definition of the mutual information by Fano. Since our measure is well-defined for individual realizations of the random variables it lends itself for example to local learning in artificial neural networks. We also show that it has a meaningful M\"{o}bius inversion on a redundancy lattice and obeys a target chain rule. We give an operational interpretation of the measure based on the decisions that an agent should take if given only the shared information.
翻译:多变量相互信息的局部信息分解(PID)描述了一组源变量包含目标变量信息的不同方式。威廉斯和比尔的开创性工作表明,这种分解无法在不做额外假设的情况下从经典信息理论中确定,并提出了若干备选措施,这些措施往往借鉴了决策理论等相关领域的原则。对于潜在的概率质量功能而言,这些措施无一不同。我们在这里提出了一个满足这一属性的新措施,它仅来自信息理论原则,并具有一种本地相互信息的形式。我们从概率质量排除的角度来说明如何理解该措施,这一原则是法诺对相互信息最初定义的基础。由于我们的措施是为了个别了解随机变量,例如它有助于在人工神经网络中进行本地学习。我们还表明,它具有有意义的 M\"{o}bius 反向冗余的特性,并遵循了目标链规则。我们根据代理人应当作出的决定对措施进行操作性解释,如果仅提供共享的信息,则仅根据共享的信息。