2022 IEEE International Conference on Image Processing (ICIP) 2022
DOI: 10.1109/icip46576.2022.9897240
|View full text |Cite
|
Sign up to set email alerts
|

AIVC: Artificial Intelligence Based Video Codec

Abstract: This paper introduces AIVC, an end-to-end neural video codec. It is based on two conditional autoencoders MNet and CNet, for motion compensation and coding. AIVC learns to compress videos using any coding configurations through a single end-to-end rate-distortion optimization. Furthermore, it offers performance competitive with the recent video coder HEVC under several established test conditions. A comprehensive ablation study is performed to evaluate the benefits of the different modules composing AIVC. The … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 16 publications
0
6
0
Order By: Relevance
“…where ŷik|j is the latent coefficient at the position i of level k of description j and c ik|j ∈ Z C is the set of the C decoded coefficients neighboring ŷik|j , representing decoding context. The auto-regressive model p ψ , estimated through a MLP, uses the Laplace distribution as described in [11] to approximate the real conditional probability of the latent space and by using the factorized model defined in (7). The rate for each description ŷj can be expressed as:…”
Section: Autoregressive Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…where ŷik|j is the latent coefficient at the position i of level k of description j and c ik|j ∈ Z C is the set of the C decoded coefficients neighboring ŷik|j , representing decoding context. The auto-regressive model p ψ , estimated through a MLP, uses the Laplace distribution as described in [11] to approximate the real conditional probability of the latent space and by using the factorized model defined in (7). The rate for each description ŷj can be expressed as:…”
Section: Autoregressive Modelmentioning
confidence: 99%
“…In recent works on image compression using neural networks, the so-called Implicit Neural Representation (INR), learns to represent an image implicitly through the weights of the INR model, a coordinate map, and possibly a latent space [9,10]. More recently, the Coordinate-based Low Complexity Hierarchical Image Codec (COOL-CHIC) framework [11] has achieved superior performance compared to traditional image compression methods. The first MD scheme using INR (INR-MDSQC) has been proposed in [12] with the following advantages: generalized model training is unnecessary, high performance and flexible redundancy tuning.…”
Section: Introductionmentioning
confidence: 99%
“…As an additional experiment we apply HNeR-S for compression, driven by the demand to store daily-growing high-resolution climate and weather datasets (Huang & Hoefler, 2023). To this end, we combine our method with the existing HNeR-based compression scheme (Ladune et al, 2023) to apply image-wise compression, i.e., we train one neural network for the compression of each image. As a non-machine learning baseline, we consider a modular composable SZ3 (Liang et al, 2022) framework.…”
Section: Weather and Climate Datamentioning
confidence: 99%
“…The "base coder" with no spatial rate allocation strategy is the main anchor of the experiments in this section. To this end, an open-source LVC called AIVC which has shown a competitive performance in the CLIC challenge of the past few years, has been used [16].…”
Section: A Anchormentioning
confidence: 99%