2020
DOI: 10.48550/arxiv.2011.09190
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CVEGAN: A Perceptually-inspired GAN for Compressed Video Enhancement

Abstract: We propose a new Generative Adversarial Network for Compressed Video quality Enhancement (CVEGAN). The CVEGAN generator benefits from the use of a novel Mul 2 Res block (with multiple levels of residual learning branches), an enhanced residual non-local block (ERNB) and an enhanced convolutional block attention module (ECBAM). The ERNB has also been employed in the discriminator to improve the representational capability. The training strategy has also been re-designed specifically for video compression applic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

3
4

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 91 publications
(178 reference statements)
0
7
0
Order By: Relevance
“…In order to capture pixel movements at multiple scales, we devise a U-Net style feature extractor, U-MultiScaleResNext (UMSResNext), consisting of eight MSResNext blocks (illustrated in Figure 3). Each MSRes-Next block employs two ResNext blocks [59] in parallel with different kernel sizes in the middle layer, 3×3 and 7×7, which further increases the network cardinality [37,59]. The outputs of these two ResNext blocks are then concatenated and connected to a channel attention module [24], which learns adaptive weighting of the feature maps extracted by the two ResNext blocks.…”
Section: Multi-interflow Networkmentioning
confidence: 99%
“…In order to capture pixel movements at multiple scales, we devise a U-Net style feature extractor, U-MultiScaleResNext (UMSResNext), consisting of eight MSResNext blocks (illustrated in Figure 3). Each MSRes-Next block employs two ResNext blocks [59] in parallel with different kernel sizes in the middle layer, 3×3 and 7×7, which further increases the network cardinality [37,59]. The outputs of these two ResNext blocks are then concatenated and connected to a channel attention module [24], which learns adaptive weighting of the feature maps extracted by the two ResNext blocks.…”
Section: Multi-interflow Networkmentioning
confidence: 99%
“…The spatio-temporal filter adaptive network was developed for video deblurring [45] and optical flow estimation was jointly trained with processing component for video enhancement [41]. Recently, generative adversarial network (GAN) also attracted considerable attention due to its favorable performance especially for extremely low bit-rate compression [5,23].…”
Section: Related Workmentioning
confidence: 99%
“…Since TMIV is based on a 2D video coding framework, we hypothesise that it would benefit from resolution adaptation methods [9,10,11,12,13] and post-processing methodologies [14,15] that have proved successful in conventional coding. In this context we propose a novel approach that exploits resolution adaptation and post-processing of synthesized views, realised using a Convolutional Neural Network (CNN).…”
Section: Introductionmentioning
confidence: 99%