TY - GEN
T1 - Scalable Learned Image Compression with A Recurrent Neural Networks-Based Hyperprior
AU - Su, Rige
AU - Cheng, Zhengxue
AU - Sun, Heming
AU - Katto, Jiro
N1 - Funding Information:
This work is supported by JST, PRESTO Grant Number JP-MJPR19M5, Japan.
Publisher Copyright:
© 2020 IEEE.
PY - 2020/10
Y1 - 2020/10
N2 - Recently learned image compression has achieved many great progresses, such as representative hyperprior and its variants based on convolutional neural networks (CNNs). However, CNNs are not fit for scalable coding and multiple models need to be trained separately to achieve variable rates. In this paper, we incorporate differentiable quantization and accurate entropy models into recurrent neural networks (RNNs) architectures to achieve a scalable learned image compression. First, we present an RNN architecture with quantization and entropy coding. To realize the scalable coding, we allocate the bits to multiple layers, by adjusting the layer-wise lambda values in Lagrangian multiplier-based rate-distortion optimization function. Second, we add an RNN-based hyperprior to improve the accuracy of entropy models for multiple-layer residual representations. Experimental results demonstrate that our performance can be comparable with recent CNN-based hyperprior methods on Kodak dataset. Besides, our method is a scalable and flexible coding approach, to achieve multiple rates using one single model, which is very appealing.
AB - Recently learned image compression has achieved many great progresses, such as representative hyperprior and its variants based on convolutional neural networks (CNNs). However, CNNs are not fit for scalable coding and multiple models need to be trained separately to achieve variable rates. In this paper, we incorporate differentiable quantization and accurate entropy models into recurrent neural networks (RNNs) architectures to achieve a scalable learned image compression. First, we present an RNN architecture with quantization and entropy coding. To realize the scalable coding, we allocate the bits to multiple layers, by adjusting the layer-wise lambda values in Lagrangian multiplier-based rate-distortion optimization function. Second, we add an RNN-based hyperprior to improve the accuracy of entropy models for multiple-layer residual representations. Experimental results demonstrate that our performance can be comparable with recent CNN-based hyperprior methods on Kodak dataset. Besides, our method is a scalable and flexible coding approach, to achieve multiple rates using one single model, which is very appealing.
KW - RNN-based hyperprior
KW - RNN-based image compression
KW - entropy coding
KW - quantization
UR - http://www.scopus.com/inward/record.url?scp=85098648579&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85098648579&partnerID=8YFLogxK
U2 - 10.1109/ICIP40778.2020.9190704
DO - 10.1109/ICIP40778.2020.9190704
M3 - Conference contribution
AN - SCOPUS:85098648579
T3 - Proceedings - International Conference on Image Processing, ICIP
SP - 3369
EP - 3373
BT - 2020 IEEE International Conference on Image Processing, ICIP 2020 - Proceedings
PB - IEEE Computer Society
T2 - 2020 IEEE International Conference on Image Processing, ICIP 2020
Y2 - 25 September 2020 through 28 September 2020
ER -