论文标题
微化:基于GAN的压缩的知识蒸馏
Microdosing: Knowledge Distillation for GAN based Compression
论文作者
论文摘要
最近,在学习的图像和视频压缩中取得了重大进展。尤其是生成对抗网络的使用导致低比特率制度带来了令人印象深刻的结果。但是,模型大小仍然是当前最新建议中的重要问题,现有解决方案需要在解码方面进行大量计算工作。这限制了它们在现实的场景中的用法,并扩展到视频压缩。在本文中,我们演示了如何利用知识蒸馏以在原始参数数量的一部分中获得同等功能的图像解码器。我们研究解决方案的几个方面,包括序列专业化,以及用于图像编码的侧面信息。最后,我们还展示了如何将获得的好处转移到视频压缩的设置中。总体而言,这使我们能够将模型尺寸减少20倍,并在解码时间减少50%。
Recently, significant progress has been made in learned image and video compression. In particular the usage of Generative Adversarial Networks has lead to impressive results in the low bit rate regime. However, the model size remains an important issue in current state-of-the-art proposals and existing solutions require significant computation effort on the decoding side. This limits their usage in realistic scenarios and the extension to video compression. In this paper, we demonstrate how to leverage knowledge distillation to obtain equally capable image decoders at a fraction of the original number of parameters. We investigate several aspects of our solution including sequence specialization with side information for image coding. Finally, we also show how to transfer the obtained benefits into the setting of video compression. Overall, this allows us to reduce the model size by a factor of 20 and to achieve 50% reduction in decoding time.