|
End-to-end optimized image compression with competition of prior distributions |
|
Benoit Brummer |
|
intoPIX |
|
Mont-Saint-Guibert, Belgium |
|
[email protected] De Vleeschouwer |
|
Universit ´e catholique de Louvain |
|
Louvain-la-Neuve, Belgium |
|
[email protected] |
|
Abstract |
|
Convolutional autoencoders are now at the forefront of |
|
image compression research. To improve their entropy cod- |
|
ing, encoder output is typically analyzed with a second |
|
autoencoder to generate per-variable parametrized prior |
|
probability distributions. We instead propose a compression |
|
scheme that uses a single convolutional autoencoder and |
|
multiple learned prior distributions working as a competition |
|
of experts. Trained prior distributions are stored in a static |
|
table of cumulative distribution functions. During inference, |
|
this table is used by an entropy coder as a look-up-table |
|
to determine the best prior for each spatial location. Our |
|
method offers rate-distortion performance comparable to |
|
that obtained with a predicted parametrized prior with only |
|
a fraction of its entropy coding and decoding complexity. |
|
1. Introduction |
|
Image compression typically consists of a transforma- |
|
tion step (including quantization) and an entropy coding |
|
step that attempts to capture the probability distribution of |
|
a transformed context to generate a smaller compressed bit- |
|
stream. Entropy coding ranges in complexity from simple |
|
non-adaptive encoders [ 26,24] to complex arithmetic coders |
|
with adaptive context models [ 15,23]. The entropy cod- |
|
ing strategy has been revised to address the specificities of |
|
learned compression. More specifically, for recent works |
|
that make use of a convolutional autoencoder [ 12] (AE) as |
|
the all-inclusive transformation and quantization step, the en- |
|
tropy coder relies on a cumulative probability model (CPM) |
|
trained alongside the AE [ 5]. This model estimates the cumu- |
|
lative distribution function (CDF) of each channel coming |
|
out of the AE and passes these learned CDFs to an entropy |
|
coder such as range encoding [16]. |
|
Such a simple method outperforms traditional codecs |
|
like JPEG2000 but work is still needed to surpass complex |
|
codecs like BPG. Johannes Ball ´e et al. (2018) [ 6] proposed |
|
analyzing the output of the convolutional encoder with an- |
|
other AE to generate a floating-point scale parameter thatdiffers for every variable that needs to be encoded by the |
|
entropy coder, thus for every location in every channel. This |
|
method has been widely used in subsequent works but in- |
|
troduces substantial complexity in the entropy coding step |
|
because a different CDF is needed to encode every variable |
|
in the latent representation of the image, whereas the single |
|
AE method by Ball ´e et al. (2017) [ 5] reused the same CDF |
|
table for every latent spatial location. |
|
Our work uses the principle of competition of experts |
|
[22,14] to get the best out of both worlds. Multiple prior |
|
distributions compete for the lowest bit cost on every spatial |
|
location in the quantized latent representation. During train- |
|
ing, only the best prior distribution is updated in each spatial |
|
location, further improving the prior distributions special- |
|
ization. CDF tables are fixed at the end of training. Hence, |
|
at testing, the CDF table resulting in the lowest bitcost is |
|
assigned to each spatial location of the latent representation. |
|
The rate-distortion (RD) performance obtained is compa- |
|
rable to that obtained with a parametrized distribution [ 6], |
|
yet the entropy coding process is greatly simplified since it |
|
does not require a per-variable CDF and can build on look- |
|
up-tables (LUT) rather than the computation of analytical |
|
distributions. |
|
2. Background |
|
Entropy coders such as range encoding [ 16] require cdfs |
|
where, for each variable to be encoded, the probability that a |
|
smaller or equal value appears is defined for every allowable |
|
value in the latent representation space. Johannes Ball ´e et |
|
al.’s seminal work (2017) [ 5] consists of an AE, computing a |
|
latent image representation consisting in CLchannels of size |
|
HLWL, and a CPM, consisting of one CDF per latent out- |
|
put channel, which are trained conjointly. The latent repre- |
|
sentation coming out of the encoder is quantized then passed |
|
through the CPM. The CPM defines, in a parametrized and |
|
differentiable manner, a CDF per channel. At the end of |
|
training, the CPM is evaluated at every possible value1to |
|
generate the static CDF table. The CDF table is not differen- |
|
tiable, but going from a differentiable CPM to a static CDF |
|
table speeds up the encoding and decoding process. The |
|
1arXiv:2111.09172v1 [eess.IV] 17 Nov 2021CDF table is used to compress latent representations with an |
|
entropy coder, the approximate bit cost of a symbol is the |
|
binary logarithm of its probability. |
|
Ball´e et al. (2018) improved the RD efficiency by re- |
|
placing the unique CDF table with a Gaussian distribution |
|
parametrized with a hyperprior (HP) sub-network [ 6]. The |
|
HP generates a scale parameter, and in turn a different CDF, |
|
for every variable to be encoded. Thus, complexity is added |
|
by exploiting the parametrized Gaussian prior during the |
|
entropy coding process, since a different CDF is required for |
|
each variable in the channel and spatial dimensions. |
|
Minnen et al. proposed a scheme where one of multi- |
|
ple probability distributions is chosen to adapt the entropy |
|
model locally [ 21]. However, these distributions are defined |
|
a posteriori, given the encoder trained with a global entropy |
|
model. Thus [ 21] does not perform as well as the HP scheme |
|
[6] per [ 19, Fig. 2a]. In contrast, the present method jointly |
|
optimizes the local entropy models and the AE in an end-to- |
|
end fashion that results in greater performance. Minnen et al. |
|
[19] later proposed to improve RD with the use of an autore- |
|
gressive sequential context model. However, as highlighted |
|
in [13], this is obtained at the cost of increased runtime |
|
by several orders of magnitude. Subsequent works have |
|
attempted to reduce complexity of the neural network archi- |
|
tecture [ 10] and to bridge the RD gap with Minnen’s work |
|
[13], but entropy coding complexity has remained largely |
|
unaddressed and has instead evolved towards increased com- |
|
plexity [ 19,7,20] compared to [ 6]. The present work builds |
|
on Ball ´e et al. (2017) [ 5] and achieves the performance of |
|
Ball´e et al. (2018) [ 6] without the complexity introduced |
|
by a per-variable parametrized probability distribution. We |
|
chose Ball ´e et al. (2017) as a baseline because it corresponds |
|
to the basic unit adopted as a common reference and starting |
|
point for most models proposed in the recent literature to im- |
|
prove compression quality [ 6,19,13,20]. Due to its generic |
|
nature, our contribution remains relevant for the newer, often |
|
computationally more complex, incremental improvements |
|
on Ball ´e et al. (2017). |
|
3. Competition of prior distributions |
|
Our proposed method introduces competitions of expert |
|
[22,14] prior distributions: a single AE transforms the image |
|
and a set of prior distributions are trained to model the CDF |
|
of the latent representation in each spatial location. For each |
|
latent spatial dimension the CDF table which minimizes |
|
bit cost is selected; that prior is either further optimized |
|
on the features it won in the training mode, or its index is |
|
stored for decoding in the inference mode. This scheme is |
|
illustrated in Figure 1, a set of 16 optimized CDF tables is |
|
shown in Figure 2, and three sample images are segmented |
|
by “winning” CDF table in Figure 3. |
|
All prior distributions are estimated in parallel by consid- |
|
ering NCDFCDF tables, and selecting, as a function of the |
|
CDF0... |
|
CDFNCDF...Cumulative probability model |
|
Input image |
|
Encoder |
|
Decoder |
|
Reconstruction |
|
Distortion |
|
(eg: MSE, MS-SSIM, discriminator)x̂ |
|
bitstream[ ŷk,l]ik,lbitstream |
|
bitCost( ŷk,l, )BackpropagateBackpropagate |
|
Entropy coder |
|
ik,l=argmin p |
|
[bitCost( ŷk,l, CDF p)]...y x |
|
ŷ |
|
CDFik,l |
|
Quantizer |
|
Noise (train), |
|
round (test) ŷk,lFor all k,l |
|
CDFik,l |
|
CDFik,lFigure 1. AE compression scheme with competition of prior distri- |
|
butions. The AE architecture is detailed in [ 6, Fig. 4]. The indices |
|
idenote the indices of CDF tables that minimizes the bitcount for |
|
each latent spatial dimension. Loss = Distortion + bitCost. |
|
0.00.51.0 |
|
0.00.51.0 |
|
0.00.51.0 |
|
100 |
|
0 1000.00.51.0 |
|
100 |
|
0 100 100 |
|
0 100 100 |
|
0 100 |
|
Figure 2. We observe some diversity among the 16 cumulative |
|
distribution functions learned by a network trained with MSE loss |
|
and= 4096 . Each box presents a CDF table and each colored |
|
line corresponds to the cdfof one of 256 latent channels. The best |
|
fitting CDF table is selected for each latent spatial location. |
|
encoded latent spatial location, the one that minimizes the |
|
entropy coder bitcount. The CDF table index is determined |
|
for each spatial location by evaluating each CDF table in |
|
inference. This can be done in a vectorized operation given |
|
sufficient memory. During training the CPM is evaluated |
|
instead of CDF tables such that the probabilities are up to |
|
date and the model is differentiable, and the bit cost is re- |
|
turned as it contributes to the loss function. The cost of CDF |
|
table indices has been shown to be neglectable due to the |
|
reasonably small number of priors, which in turns results |
|
from the fact that little gain in latent code entropy has been |
|
obtained by increasing the number of priors. |
|
In all our experiments , the AE architecture follows the |
|
one in Ball ´e et al. (2018) [ 6], without the HP, since we found |
|
that the AE from [ 6] offers better RD than the one describedFigure 3. Segmentation of three test images [ 1]: each distinct color |
|
represents one of 64 CDF tables used to encode a latent spatial |
|
location ( 1616pixels patch) |
|
in Ball ´e et al. (2017) [ 5], even with a single CDF table. A |
|
functional training loop is described in Algorithm 1. |
|
Algorithm 1 Training loop |
|
y model.Encoder( x) |
|
ˆy quantize( y) |
|
ˆx clip(model.Decoder( ˆy), 0, 1) |
|
distortion visualLossFunction( ˆx,x) |
|
for0k< HLand0l<WLdo |
|
bitCost [k;l] min i<NCDF log2 |
|
CPM i(ˆy[k;l] + 0:5) CPM i(ˆy[k;l] 0:5) |
|
end for .CPM is the differentiable version of CDF |
|
Loss distortion+jbitCostj |
|
Loss.backward() |
|
4. Experiments |
|
4.1. Method |
|
These experiments are based on the PyTorch implemen- |
|
tation of Ball ´e et al. (2018) [ 6] published by Liu Jia- |
|
heng [ 9,13]. To implement our proposed method, the |
|
HP is omitted in favor of competition of expert prior dis- |
|
tributions. The CPM is that defined in [ 9] with an addi- |
|
tional NCDFdimension to compute all CDF tables in par- |
|
allel. Theoretical results are verified using the torchac |
|
range coder [ 18,17,16]. A functional training loop is de- |
|
scribed in Algorithm 1, and source code is provided on |
|
https://github.com/trougnouf/Manypriors . |
|
To ensure that all priors get an opportunity to train, the prior |
|
distributions that have not been used for at least fifty stepsare randomly assigned to spatial locations with largest bit- |
|
counts, to be forced to train. The Adam optimizer [ 11] is |
|
used with a starting learning rate (LR) of 0.0001 for the AE |
|
and 0.001 for the CPM. Performance is tested every 2500 |
|
steps in inference mode on the validation set, and the LR |
|
is decayed by a factor of 0.99 if the performance have not |
|
improved for two tests. Reported performance is the one of |
|
the model taht minimizes (visualLoss+bitCost )on the |
|
validation set at the end of training. Base models are trained |
|
for six million steps at = 4096 with the mean squared er- |
|
ror (MSE) loss. Smaller values and MS-SSIM models are |
|
trained for four million steps starting from the base model |
|
with their LR and optimizer reset. All models use CH= 192 |
|
(hidden layers channels) and CL= 256 (output channels) |
|
such that a single base model is needed for each prior con- |
|
figuration. The training and validation dataset is made of |
|
free-license images from Wikimedia Commons [ 3]; mainly |
|
“Category:Featured pictures on Wikimedia Common” which |
|
consists of 13928 images of the highest quality. The images |
|
are cropped into 10242pixels patches on disk to speed up |
|
further resizing, then they are resized on-the-fly by a random |
|
factor down to 2562pixels during training. A batch size of 4 |
|
patches is used. The kodak set [ 2] is used as a validation set |
|
and the CLIC professional test dataset [ 4] is used for testing. |
|
The RD curve of our “multiprior” model is compared |
|
with that of the HP model [ 6], which is trained from scratch |
|
using Liu Jiaheng’s PyTorch implementation [ 9,13]. Liu |
|
Jiaheng’s code differs slightly from the paper’s definition [ 6] |
|
in that a Laplace distribution is used in place of the normal |
|
distribution to stabilize training. Complexity is measured as |
|
the number of GMac (billion multiply-accumulate operation) |
|
using the ptflops counter [ 25] and the number of memory |
|
lookup operations is calculated manually. |
|
4.2. Results |
|
The PSNR RD curve measured on the CLIC professional |
|
test set [ 4] is shown on top of Figure 4. The performance |
|
of a 64-priors model is in line with that of the HP model |
|
: they both perform slightly better than BPG at high bpp, |
|
and achieve significantly better RD than the single-prior |
|
model. In the middle, the RD value at = 4096 , the highest |
|
bitrate, is shown for 1, 2, 4, 8, 16, 32, 64, and 128 prior |
|
distributions. 128-priors offer marginal gains and costs an |
|
increased training time (1.5) and encoding time. MS-SSIM |
|
performance of fine-tuned models is shown in the bottom |
|
of Figure 4; the 64-priors model still performs similarly to |
|
[6], and both learned compression models benefit from this |
|
more perceptual metric compared with traditional codecs. A |
|
visual comparison of images compressed with the MSE loss |
|
(= 512 ) and the equivalent bitrate settings in conventional |
|
codecs is shown in Figure 5. |
|
Computational complexity of our Manypriors has been |
|
compared to the one of the HP model [ 6]). This complex-0.0 0.2 0.4 0.6 0.8 1.0323334353637383940PSNR |
|
1-prior |
|
2-priors |
|
4-priors |
|
8-priors |
|
16-priors |
|
32-priors |
|
64-priors |
|
128-priors |
|
hyperprior |
|
BPG |
|
JPEG |
|
0.60 0.65 0.70 0.75 0.80 0.8538.538.638.738.838.939.039.139.2PSNR |
|
1-prior |
|
2-priors |
|
4-priors |
|
8-priors |
|
16-priors |
|
32-priors |
|
64-priors |
|
128-priors |
|
hyperprior |
|
BPG |
|
JPEG |
|
0.0 0.2 0.4 0.6 0.8 1.0 |
|
bpp |
|
0.960.970.980.991.00MS-SSIM |
|
1-prior |
|
64-priors |
|
hyperprior |
|
BPG |
|
JPEGFigure 4. Top: PSNR RD curve of a 64-priors model on the CLIC |
|
pro. test set, compared with the HP model [ 6], and the BPG and |
|
JPEG codecs. Middle : Zoom in on models with 1, 2, 4, 8, 16, 32, |
|
and 64 priors. Bottom: MS-SSIM RD curve. |
|
Table 1. Complexity of the HP model [ 6]) compared to Manypriors |
|
(ours), expressed in GMac for the neural network parts and number |
|
of memory lookup operations (* or parametrized Laplace CDF |
|
generations in full-precision) for the CDF tables generation, to |
|
process a 4K image. |
|
(#) Hyperprior Manypriors ratio MPHP |
|
EncodingGMacmain encoder 769.82 769.82 |
|
hyper encoder 23.75 |
|
hyper decoder 23.86 |
|
total 817.43 769.82 0.942 |
|
Lookupsindices 530.84 M |
|
CDF 829.44 K * 32.400 K |
|
total 829.44 K* 530.87 M N CDF= 64 |
|
DecodingGMachyper decoder 23.154 |
|
main decoder 769.60 769.60 |
|
total 792.75 769.60 0.971 |
|
Lookups CDF (total) 829.44 K* 32.400 K1 |
|
CL= 0:004 |
|
ity is expressed in GMac for the neural network parts and |
|
number of memory lookup operations. It is summarized in |
|
Table 1. The lack of a HP AE saves 3 % to 6 % GMac, de- |
|
pending on whether only the HP decoder (image decoding) |
|
Figure 5. Visual comparison of Larry the cat [ 1] compressed with |
|
learned (= 512 ) and conventional methods. Top-left: uncom- |
|
pressed, top-middle: JPEG (PSNR: 29.3, 0.224 bpp), top-right: |
|
BPG (PSNR: 32.9, 0.217 bpp), bottom-left: 1-prior (PSNR: 32.4, |
|
0.252 bpp), bottom-middle: hyperprior (PSNR: 32.8, 0.217 bpp), |
|
bottom-right: 64-priors/ours (PSNR: 32.9, 0.218 bpp) |
|
or the whole HP codec (image encoding) is used. Decoding |
|
with the Manypriors scheme is greatly simplified compared |
|
to [6] because the CDF tables generation process takes the |
|
optimal indices stored as side-information and looks up one |
|
static CDF table per latent spatial dimension, that is CL(typ- |
|
ically 256) fewer lookups than with a HP. During encoding, |
|
the Manypriors scheme must lookup every latent variable |
|
with every CDF table in order to determine the most cost |
|
effective CDF tables. This results in NCDF(typically 64) |
|
times more lookup operations than the HP scheme overall, |
|
although these lookup operations are relatively cheap be- |
|
cause only two values are needed (variable 0.5), whereas |
|
each CDF table lookup in [ 6] returns Lprobabilities. More- |
|
over, it is challenging to make an accurate CDF LUT for the |
|
HP scheme, because quantizing the distribution scale param- |
|
eter reduces the accuracy of the resulting CDFs, negatively |
|
impacting the bitrate. This challenge is exacerbated when |
|
the distribution has multiple parameters [ 19] or a mixture |
|
of distributions [ 7] is used. In Figure 4, LUT are replaced |
|
by accurate but complex Laplace distribution computation |
|
for the HP scheme in order to maximize the reported RD |
|
performance. |
|
Time complexity is measured for every step on CPU, |
|
where it can be reliably profiled due to synchroneous execu- |
|
tion. It is summarized in Table 2 with the following distinct |
|
sub-categories: NN (neural network) is the time spent in |
|
the AE, CDF generation is the time spent building the CDF |
|
tables for a specific image, and entropy is the bitstream gener- |
|
ation. All operations are done using the PyTorch framework |
|
in python, except for entropy encoding which makes use of |
|
the torchac range coding library [ 18,17], written in C++, andTable 2. Breaking down the image encoding and decoding time, in |
|
seconds. Image: 4.5 MP snail [ 1]. CPU: AMD Ryzen 7 2700X. |
|
Time avg. of 50 runs. |
|
(#)Hyperprior |
|
(Ball ´e2018)64-priors |
|
(ours)ratio |
|
(oursHP) |
|
EncodingNN encode: main + hyperprior 3.81 + 0.41 3.79 + 0.00 0.90 |
|
entropy encode, main + hyperprior 0.15 + 0.02 0.15 + 0.00 |
|
CDF : select indices + gather tables 0.00 +FP: 15.95 |
|
LUT: 5.661.90 + 0.81FP: 0.17 |
|
LUT: 0.48 |
|
encode (total)FP: 20.33 |
|
LUT: 10.046.65FP: 0.32 |
|
LUT: 0.66 |
|
DecodingNN decode : main + hyperprior 10.66 + 0.34 10.50 0.95 |
|
CDF : gather tablesFP: 15.95 |
|
LUT: 5.660.81FP: 0.05 |
|
LUT: 0.14 |
|
entropy decode : main + hyperprior 0.24 + 0.02 0.24 0.92 |
|
decode (total)FP: 27.21 |
|
LUT: 16.9211.54FP: 0.42 |
|
LUT: 0.68 |
|
the prior indices are compressed using the LZMA library [ 8]. |
|
The total encoding time of the 64-priors model is 0.32 time |
|
that of the HP model and the decoding time is 0.42 times that |
|
of the HP model. The timing is more significant when it is |
|
broken down by sub-category because each component has |
|
a different response time depending on the hardware (and |
|
software) architecture in place. The AE (“NN”) encoding |
|
time is 0.90 that of the HP scheme and decoding time is |
|
0.95 time as much as the HP. Both the hyper-encoder and |
|
hyper-decoder are called during encoding, thus it appears |
|
that each part of the HP sub-network costs 5 % of the AE |
|
time. The time taken to build the CDF tables for the HP |
|
model was measured both by estimating the per-variable |
|
Laplace distributions (“full-precision”) and with a quantized |
|
scale parameter LUT. In any case, finding the best indices of |
|
a 64-priors model appears to be relatively inexpensive and |
|
the total CDF tables generation time is only 0.17 to 0.48 that |
|
of the HP model (depending on whether the HP model uses |
|
full-precision or LUT) for encoding. During decoding, the |
|
64-priors model spends 0.05 to 0.14 as much time building |
|
the CDF tables as the HP model, because the optimal CDF |
|
table indices have already been determined during encoding |
|
and they are included in the bitstream. |
|
5. Conclusion |
|
Convolutional autoencoders trained for compression are |
|
optimized for both rate and distortion. Rate is estimated with |
|
a cumulative probability model, which in turns generates a |
|
CDF for every latent variable to be encoded. A single CDF |
|
per latent channel is not sufficient to capture the statistics at |
|
the output of the encoder, nor to allow the encoder to express |
|
a wide variety of features. To support multiple statistics, the |
|
hyperprior [ 6] parametrizes a standard distribution, but this |
|
introduces a great deal of complexity in the entropy coding |
|
stage because the CDF differs for every latent variable to be |
|
encoded. The proposed method uses multiple prior distri- |
|
butions working as a competition of experts to capture the |
|
relevant features which they specialize on. This approach is |
|
advantageous because the learned CDF tables are stored in |
|
a static LUT once training is finished, and a model trainedwith 64 prior distributions performs with a similar RD as |
|
one trained with a HP sub-network. Moreover, a learned |
|
CDF table includes the CDF for all channels in the latent |
|
code. Hence, accessing the CDF table for a spatial location |
|
provides the CDF for each of its channels and the number of |
|
lookups is reduced to the number of latent spatial locations. |
|
In our experiments, CDF tables generation in the encoding |
|
step takes 0.17 to 0.48 as much time with a 64-priors model |
|
as it does with the HP model (depending on the precision |
|
of the HP model). This ratio is lowered to 0.05 to 0.14 dur- |
|
ing decoding because the prior indices have already been |
|
determined during the encoding. |
|
6. Acknowledgements |
|
This research has been funded by the Walloon Region. |
|
Computational resources have been provided by the super- |
|
computing facilities of the Universit ´e catholique de Lou- |
|
vain (CISM/UCL) and the Consortium des ´Equipements de |
|
Calcul Intensif en F ´ed´eration Wallonie Bruxelles (C ´ECI) |
|
funded by the Fond de la Recherche Scientifique de Bel- |
|
gique (F.R.S.-FNRS) under convention 2.5020.11 and by the |
|
Walloon Region. |
|
References |
|
[1]Commons test photographs. https : / / commons . |
|
wikimedia . org / wiki / Category : Commons _ |
|
Test_Photographs . Accessed: 2020-10-22. 3, 4, 5 |
|
[2]True color kodak images. http://r0k.us/graphics/ |
|
kodak/ . Accessed: 2020-01-20. 3 |
|
[3]Wikimedia commons. https://commons.wikimedia. |
|
org. Accessed: 2020-04-03. 3 |
|
[4]Challenge on learned image compression. http:// |
|
challenge.compression.cc/tasks/ , 2020. 3 |
|
[5]Johannes Ball ´e, Valero Laparra, and Eero Simoncelli. End- |
|
to-end optimized image compression. In 5th International |
|
Conference on Learning Representations, ICLR 2017 , 2017. |
|
1, 2, 3 |
|
[6]Johannes Ball ´e, David Minnen, Saurabh Singh, Sung Jin |
|
Hwang, and Nick Johnston. Variational image compression |
|
with a scale hyperprior. In International Conference on Learn- |
|
ing Representations , 2018. 1, 2, 3, 4, 5 |
|
[7]Zhengxue Cheng, Heming Sun, Masaru Takeuchi, and Jiro |
|
Katto. Learned image compression with discretized gaussian |
|
mixture likelihoods and attention modules. In Proceedings of |
|
the IEEE/CVF Conference on Computer Vision and Pattern |
|
Recognition (CVPR) , June 2020. 2, 4 |
|
[8] Lasse Collin and Igor Pavlov. Xz utils, Jul 2009. 5 |
|
[9]Liu Jiaheng. compression. https://github.com/ |
|
liujiaheng/compression , 2020. 3 |
|
[10] Nick Johnston, Elad Eban, Ariel Gordon, and Johannes Ball ´e. |
|
Computationally efficient neural image compression. CoRR , |
|
abs/1912.08771, 2019. 2 |
|
[11] Diederik P. Kingma and Jimmy Ba. Adam: A method for |
|
stochastic optimization. In Yoshua Bengio and Yann LeCun,editors, 3rd International Conference on Learning Represen- |
|
tations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, |
|
Conference Track Proceedings , 2015. 3 |
|
[12] Mark A. Kramer. Nonlinear principal component analy- |
|
sis using autoassociative neural networks. AIChE Journal , |
|
37(2):233–243, 1991. 1 |
|
[13] Jiaheng Liu, Guo Lu, Zhihao Hu, and Dong Xu. A unified |
|
end-to-end framework for efficient deep image compression. |
|
arXiv preprint arXiv:2002.03370 , 2020. 2, 3 |
|
[14] Shunta Maeda. Fast and flexible image blind denoising |
|
via competition of experts. In 2020 IEEE/CVF Conference |
|
on Computer Vision and Pattern Recognition Workshops |
|
(CVPRW) , pages 2239–2247, 2020. 1, 2 |
|
[15] Detlev Marpe, Heiko Schwarz, and Thomas Wiegand. |
|
Context-based adaptive binary arithmetic coding in the |
|
h.264/avc video compression standard. IEEE Transactions on |
|
Circuits and Systems for Video Technology , 13(7):620–636, |
|
2003. 1 |
|
[16] Gloria Mart ´ın. Range encoding: an algorithm for removing |
|
redundancy from a digitised message. In Video and Data |
|
Recording Conference, Southampton, 1979 , pages 24–27, |
|
1979. 1, 3 |
|
[17] Fabian Mentzer. torchac. https://github.com/fab- |
|
jul/L3C-PyTorch/tree/master/src/torchac , |
|
2020. 3, 4 |
|
[18] Fabian Mentzer, Eirikur Agustsson, Michael Tschannen, |
|
Radu Timofte, and Luc Van Gool. Practical full resolution |
|
learned lossless image compression. In Proceedings of the |
|
IEEE Conference on Computer Vision and Pattern Recogni- |
|
tion (CVPR) , 2019. 3, 4 |
|
[19] David Minnen, Johannes Ball ´e, and George D Toderici. Joint |
|
autoregressive and hierarchical priors for learned image com- |
|
pression. In S. Bengio, H. Wallach, H. Larochelle, K. Grau- |
|
man, N. Cesa-Bianchi, and R. Garnett, editors, Advances |
|
in Neural Information Processing Systems 31 , pages 10771– |
|
10780. Curran Associates, Inc., 2018. 2, 4 |
|
[20] D. Minnen and S. Singh. Channel-wise autoregressive en- |
|
tropy models for learned image compression. In 2020 IEEE |
|
International Conference on Image Processing (ICIP) , pages |
|
3339–3343, 2020. 2 |
|
[21] D. Minnen, G. Toderici, S. Singh, S. J. Hwang, and M. Covell. |
|
Image-dependent local entropy models for learned image |
|
compression. In 2018 25th IEEE International Conference |
|
on Image Processing (ICIP) , pages 430–434, 2018. 2 |
|
[22] Giambattista Parascandolo, Niki Kilbertus, Mateo Rojas- |
|
Carulla, and Bernhard Sch ¨olkopf. Learning independent |
|
causal mechanisms. volume 80 of Proceedings of Machine |
|
Learning Research , pages 4036–4044, Stockholmsm ¨assan, |
|
Stockholm Sweden, 10–15 Jul 2018. PMLR. 1, 2 |
|
[23] Alexander Rhatushnyak, Jan Wassenberg, Jon Sneyers, Jyrki |
|
Alakuijala, Lode Vandevenne, Luca Versari, Robert Obryk, |
|
Zoltan Szabadka, Evgenii Kliuchnikov, Iulia-Maria Comsa, |
|
Krzysztof Potempa, Martin Bruse, Moritz Firsching, Renata |
|
Khasanova, Ruud van Asseldonk, Sami Boukortt, Sebastian |
|
Gomez, and Thomas Fischbacher. Committee draft of jpeg xl |
|
image coding system, 2019. 1[24] Thomas Richter, Joachim Keinert, Antonin Descampe, and |
|
Gael Rouvroy. Entropy coding and entropy coding improve- |
|
ments of jpeg xs. In 2018 Data Compression Conference , |
|
pages 87–96, 2018. 1 |
|
[25] Vladislav Sovrasov. Flops counter for convolutional net- |
|
works in pytorch framework. https://github.com/ |
|
sovrasov/flops-counter.pytorch , 2020. 3 |
|
[26] Gregory K Wallace. The jpeg still picture compression stan- |
|
dard. IEEE transactions on consumer electronics , 38(1):xviii– |
|
xxxiv, 1992. 1 |