论文标题
ReactNet:朝着具有广义激活功能的精确二进制神经网络
ReActNet: Towards Precise Binary Neural Network with Generalized Activation Functions
论文作者
论文摘要
在本文中,我们提出了几种想法,以增强二进制网络,以缩小实际价值网络的准确性差距,而不会产生任何额外的计算成本。我们首先通过使用无参数的快捷方式修改和对紧凑型实价网络进行修改和二线化,绕过所有中间卷积层,包括下采样层,构建基线网络。该基线网络在准确性和效率之间取得了良好的权衡,比大多数现有的二进制网络以大约一半的计算成本达到了卓越的性能。通过广泛的实验和分析,我们观察到二进制网络的性能对激活分布变化敏感。基于这一重要的观察,我们建议概括传统的符号和PRELU函数,称为各个广义功能的RSIGN和RPRELU,以显式学习分布重塑和以接近零的额外成本的转移。最后,我们采用分配损失来进一步执行二进制网络,以学习与现实价值网络的输出分布相似的输出分布。我们表明,在整合了所有这些想法之后,提议的ReactNet以很大的利润优于所有最先进的东西。具体而言,对于TOP-1的精度,它的表现分别优于实际到二进制网和Meliusnet29,分别优于4.0%和3.6%,并且在ImagEnet数据集上的差距将其差距缩小到3.0%的TOP-1准确性。代码和模型可在以下网址提供:https://github.com/liuzechun/reaectnet。
In this paper, we propose several ideas for enhancing a binary network to close its accuracy gap from real-valued networks without incurring any additional computational cost. We first construct a baseline network by modifying and binarizing a compact real-valued network with parameter-free shortcuts, bypassing all the intermediate convolutional layers including the downsampling layers. This baseline network strikes a good trade-off between accuracy and efficiency, achieving superior performance than most of existing binary networks at approximately half of the computational cost. Through extensive experiments and analysis, we observed that the performance of binary networks is sensitive to activation distribution variations. Based on this important observation, we propose to generalize the traditional Sign and PReLU functions, denoted as RSign and RPReLU for the respective generalized functions, to enable explicit learning of the distribution reshape and shift at near-zero extra cost. Lastly, we adopt a distributional loss to further enforce the binary network to learn similar output distributions as those of a real-valued network. We show that after incorporating all these ideas, the proposed ReActNet outperforms all the state-of-the-arts by a large margin. Specifically, it outperforms Real-to-Binary Net and MeliusNet29 by 4.0% and 3.6% respectively for the top-1 accuracy and also reduces the gap to its real-valued counterpart to within 3.0% top-1 accuracy on ImageNet dataset. Code and models are available at: https://github.com/liuzechun/ReActNet.