论文标题
神经网络重量的位培训
Bit-wise Training of Neural Network Weights
论文作者
论文摘要
我们介绍了一种算法,其中可以学习代表神经网络权重的单个位。这种方法允许在任意位深度上具有整数值的训练权重,并且自然地发现稀疏网络,而没有其他约束或正则化技术。与卷积和剩余网络的标准培训相比,我们显示出比具有完全连接的网络的标准培训技术更好的结果。通过以选择性的方式训练位,我们发现前三个最重要的位给出了实现高精度的最大贡献,而其余的则提供了内在的正则化。因此,超过90%的网络可用于存储任意代码而不会影响其准确性。这些代码可能是随机噪声,二进制文件,甚至是先前训练的网络的权重。
We introduce an algorithm where the individual bits representing the weights of a neural network are learned. This method allows training weights with integer values on arbitrary bit-depths and naturally uncovers sparse networks, without additional constraints or regularization techniques. We show better results than the standard training technique with fully connected networks and similar performance as compared to standard training for convolutional and residual networks. By training bits in a selective manner we found that the biggest contribution to achieving high accuracy is given by the first three most significant bits, while the rest provide an intrinsic regularization. As a consequence more than 90\% of a network can be used to store arbitrary codes without affecting its accuracy. These codes may be random noise, binary files or even the weights of previously trained networks.