2020
DOI: 10.48550/arxiv.2006.13125
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DeepQTMT: A Deep Learning Approach for Fast QTMT-based CU Partition of Intra-mode VVC

Tianyi Li,
Mai Xu,
Runzhi Tang
et al.

Abstract: The latest standard Versatile Video Coding (VVC) significantly improves the coding efficiency over its ancestor standard High Efficiency Video Coding (HEVC), but at the expense of sharply increased complexity. In VVC, the quadtree plus multi-type tree (QTMT) structure of coding unit (CU) partition accounts for most of encoding time, due to the bruteforce search for recursive rate-distortion (RD) optimization. Instead of the brute-force QTMT search, this paper proposes a deep learning approach to predict the QT… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 36 publications
0
5
0
Order By: Relevance
“…Early Exit [12,118,126,161,194,244,245,249,271,282,313] Model Selection [159,191,271,314] Result Cache [13,39,53,92,93,96,108,112,114,123,209,268,293,319] 3.3.1 Model Compression: Model compression techniques facilitate the deployment of resource-hungry AI models into resourceconstrained EDGE servers by reducing the complexity of the DNN. Model compression exploits the sparse nature of gradients' and computation involved while training the DNN model.…”
Section: Model Compressionmentioning
confidence: 99%
See 1 more Smart Citation
“…Early Exit [12,118,126,161,194,244,245,249,271,282,313] Model Selection [159,191,271,314] Result Cache [13,39,53,92,93,96,108,112,114,123,209,268,293,319] 3.3.1 Model Compression: Model compression techniques facilitate the deployment of resource-hungry AI models into resourceconstrained EDGE servers by reducing the complexity of the DNN. Model compression exploits the sparse nature of gradients' and computation involved while training the DNN model.…”
Section: Model Compressionmentioning
confidence: 99%
“…With the multiple early exit points, it can also be considered as an enabler for localized inference using shallow DNN models [271]. In [126], the authors proposed DeepQTMT to lower the encoding time spent on video compression. In the DeepQTMT, the authors utilised a multi-stage early exit mechanism to accommodate the high encoding time.…”
Section: Conditionalmentioning
confidence: 99%
“…This high encoding effort boosted the development of some works focusing on reducing the VVC encoding time. Several of these works [14]- [21] focused on reducing the computational cost of the QTMT partitioning structure, especially for intra-frame prediction. These solutions include fast CU decisions based on statistical analysis [14]- [17] and machine learning techniques [18]- [21]; all these solutions focus only on luminance blocks.…”
Section: Introductionmentioning
confidence: 99%
“…Several of these works [14]- [21] focused on reducing the computational cost of the QTMT partitioning structure, especially for intra-frame prediction. These solutions include fast CU decisions based on statistical analysis [14]- [17] and machine learning techniques [18]- [21]; all these solutions focus only on luminance blocks. Lei et al [14] developed a fast solution to avoid unnecessary block partition evaluations, where a subset of directional intra-frame prediction modes is evaluated to estimate the horizontal and vertical partitioning cost of the current block.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation