论文标题
使用信息瓶颈训练归一流的流量以进行竞争性生成分类
Training Normalizing Flows with the Information Bottleneck for Competitive Generative Classification
论文作者
论文摘要
信息瓶颈(IB)目标使用信息理论来制定任务 - 性能与稳健性权衡。它已成功应用于标准判别分类设置。我们提出了一个问题,即IB是否也可以用于训练生成可能性模型,例如标准化流量。由于标准化流量使用可逆网络体系结构(INNS),因此它们通过构造提供信息。这似乎与瓶颈的想法相矛盾。在这项工作中,首先,我们开发了IB-Inns的理论和方法,这是一类有条件地归一化的流量,其中使用IB目标对Inns进行了训练:引入少量{\ em Controled}信息损失,允许渐近地制定IB的IB,同时保持Inn Inn的生成能力完整。其次,我们在实验上研究了这些模型的特性,特别是用作生成分类器。该模型类具有改善不确定性量化和分布外检测等优点,但是传统的生成分类器解决方案在分类准确性方面遭受了巨大影响。我们发现IB中的权衡参数控制着生成功能和准确性接近标准分类器的混合。从经验上讲,我们在这种混合制度中的不确定性估计与常规生成和歧视性分类器相比有利。
The Information Bottleneck (IB) objective uses information theory to formulate a task-performance versus robustness trade-off. It has been successfully applied in the standard discriminative classification setting. We pose the question whether the IB can also be used to train generative likelihood models such as normalizing flows. Since normalizing flows use invertible network architectures (INNs), they are information-preserving by construction. This seems contradictory to the idea of a bottleneck. In this work, firstly, we develop the theory and methodology of IB-INNs, a class of conditional normalizing flows where INNs are trained using the IB objective: Introducing a small amount of {\em controlled} information loss allows for an asymptotically exact formulation of the IB, while keeping the INN's generative capabilities intact. Secondly, we investigate the properties of these models experimentally, specifically used as generative classifiers. This model class offers advantages such as improved uncertainty quantification and out-of-distribution detection, but traditional generative classifier solutions suffer considerably in classification accuracy. We find the trade-off parameter in the IB controls a mix of generative capabilities and accuracy close to standard classifiers. Empirically, our uncertainty estimates in this mixed regime compare favourably to conventional generative and discriminative classifiers.