2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00462
|View full text |Cite
|
Sign up to set email alerts
|

Conditional Probability Models for Deep Image Compression

Abstract: Deep Neural Networks trained as image auto-encoders have recently emerged as a promising direction for advancing the state-of-the-art in image compression. The key challenge in learning such networks is twofold: To deal with quantization, and to control the trade-off between reconstruction error (distortion) and entropy (rate) of the latent image representation. In this paper, we focus on the latter challenge and propose a new technique to navigate the ratedistortion trade-off for an image compression auto-enc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
511
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 482 publications
(512 citation statements)
references
References 29 publications
1
511
0
Order By: Relevance
“…For example, an uncompressed grayscale image has eight bpp. We first compare the proposed CCNs with mask convolutional networks (MCNs) [24], [25], PixelCNN++ [22], and side information networks (SINs) [19] for entropy modeling. As a special case of CCNs, MCNs specify the raster coding order without using any code dividing technique (see Fig.…”
Section: A Lossless Image Compressionmentioning
confidence: 99%
“…For example, an uncompressed grayscale image has eight bpp. We first compare the proposed CCNs with mask convolutional networks (MCNs) [24], [25], PixelCNN++ [22], and side information networks (SINs) [19] for entropy modeling. As a special case of CCNs, MCNs specify the raster coding order without using any code dividing technique (see Fig.…”
Section: A Lossless Image Compressionmentioning
confidence: 99%
“…Each residual block comprises two convolutional layers (normalized using the batch norm) and uses PRelu as the activation function. Inspired by [28], we also use an identical shortcut connecting the input and output of the residual blocks, which improves the performance as revealed by the experiments. Let H = f f−de ( M, Θ de ) denote the output of the joint decoder, parameterized by Θ de , and M denote the estimate of M provided by the entropy decoder.…”
Section: A Feature Encoder and Decodermentioning
confidence: 99%
“…There has been a plethora of work on learned, singleimage, lossy image compression [44,45,4,5,43,31,35,33,24]. These works generally use nonlinear transforms through convolutional neural network (CNN) layers to encode an image into a latent space, which is then quantized into discrete symbols.…”
Section: A Review On Deep Image Compressionmentioning
confidence: 99%