Markov categories are a novel framework to describe and treat problems in probability and information theory. In this work we combine the categorical formalism with the traditional quantitative notions of entropy, mutual information, and data processing inequalities. We show that several quantitative aspects of information theory can be captured by an enriched version of Markov categories, where the spaces of morphisms are equipped with a divergence or even a metric. As it is customary in information theory, mutual information can be defined as a measure of how far a joint source is from displaying independence of its components. More strikingly, Markov categories give a notion of determinism for sources and channels, and we can define entropy exactly by measuring how far a source or channel is from being deterministic. This recovers Shannon and R\'enyi entropies, as well as the Gini-Simpson index used in ecology to quantify diversity, and it can be used to give a conceptual definition of generalized entropy.
翻译:Markov 分类是描述和处理概率和信息理论问题的新框架。 在这项工作中,我们把绝对的正规主义与英特罗比、相互信息和数据处理不平等等传统的定量概念结合起来。我们表明,信息理论的若干定量方面可以通过一个更丰富的Markov 类别来捕捉,在这些类别中,形态主义空间具有差异甚至衡量标准。由于信息理论的习惯性,可以将相互信息定义为衡量一个共同来源离显示其组成部分的独立性有多远的尺度。更明显的是,Markov 类别给源和渠道提供了一个确定论的概念,我们可以通过测量一个来源或渠道离确定性有多远来精确地定义昆特罗比。这又恢复了香农和R'enyi 的异性,以及在生态学中使用的吉尼-辛普森指数来量化多样性,并且可以用来给通用的昆虫下一个概念定义。