论文标题

信息分解图超出了香农熵:HU定理的概括

Information Decomposition Diagrams Applied beyond Shannon Entropy: A Generalization of Hu's Theorem

论文作者

Lang, Leon, Baudot, Pierre, Quax, Rick, Forré, Patrick

论文摘要

在信息理论中,一个主要目标是找到有用的功能,以总结几个随机变量相互作用中包含的信息量。具体而言,人们可以询问经典的香农熵,互相信息和更高的交互信息相互关联。 This is answered by Hu's theorem, which is widely known in the form of information diagrams: it relates shapes in a Venn diagram to information functions, thus establishing a bridge from set theory to information theory.在这项工作中,我们将随机变量与关节操作一起视为一种单体,该变量是通过根据信息功能来调节的,而熵作为满足信息链规则的函数。这种抽象的观点允许证明对HU定理的概括。它适用于香农和Tsallis熵,(Tsallis)Kullback-Leibler Divergence,跨透明镜,Kolmogorov复杂性,子模型信息功能以及机器学习中的概括误差。我们的结果暗示了Chaitin的Kolmogorov复杂性,所有程度的相互作用复杂性都接近Shannon相互作用信息。 For well-behaved probability distributions on increasing sequence lengths, this shows that the per-bit expected interaction complexity and information asymptotically coincide, thus showing a strong bridge between algorithmic and classical information theory.

In information theory, one major goal is to find useful functions that summarize the amount of information contained in the interaction of several random variables. Specifically, one can ask how the classical Shannon entropy, mutual information, and higher interaction information relate to each other. This is answered by Hu's theorem, which is widely known in the form of information diagrams: it relates shapes in a Venn diagram to information functions, thus establishing a bridge from set theory to information theory. In this work, we view random variables together with the joint operation as a monoid that acts by conditioning on information functions, and entropy as a function satisfying the chain rule of information. This abstract viewpoint allows to prove a generalization of Hu's theorem. It applies to Shannon and Tsallis entropy, (Tsallis) Kullback-Leibler Divergence, cross-entropy, Kolmogorov complexity, submodular information functions, and the generalization error in machine learning. Our result implies for Chaitin's Kolmogorov complexity that the interaction complexities of all degrees are in expectation close to Shannon interaction information. For well-behaved probability distributions on increasing sequence lengths, this shows that the per-bit expected interaction complexity and information asymptotically coincide, thus showing a strong bridge between algorithmic and classical information theory.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源