2019
DOI: 10.1109/tmm.2019.2920603
|View full text |Cite
|
Sign up to set email alerts
|

Progressive Spatial Recurrent Neural Network for Intra Prediction

Abstract: Intra prediction is an important component of modern video codecs, which is able to efficiently squeeze out the spatial redundancy in video frames. With preceding pixels as the context, traditional intra prediction schemes generate linear predictions based on several predefined directions (i.e. modes) for blocks to be encoded. However, these modes are relatively simple and their predictions may fail when facing blocks with complex textures, which leads to additional bits encoding the residue. In this paper, we… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 63 publications
(25 citation statements)
references
References 51 publications
0
25
0
Order By: Relevance
“…One would intuitively expect that coding performance can be further improved if better predictions can be produced. Therefore, there have been a number of attempts to leverage the powerful capacity of stacked DNNs for better intrapredictor generation, including the CNN-based predictor refinement suggested in [113] to reduce prediction residual, additional learned mode trained using FCN models reported in [114] and [115], using RNNs in [116], using CNNs in [108], even using GANs in [117], and so on. These approaches have actively utilized the neighbor pixels or blocks and/or other context information (e.g., mode) if applicable, in order to accurately represent the local structures for better prediction.…”
Section: A Modularized Neural Video Codingmentioning
confidence: 99%
“…One would intuitively expect that coding performance can be further improved if better predictions can be produced. Therefore, there have been a number of attempts to leverage the powerful capacity of stacked DNNs for better intrapredictor generation, including the CNN-based predictor refinement suggested in [113] to reduce prediction residual, additional learned mode trained using FCN models reported in [114] and [115], using RNNs in [116], using CNNs in [108], even using GANs in [117], and so on. These approaches have actively utilized the neighbor pixels or blocks and/or other context information (e.g., mode) if applicable, in order to accurately represent the local structures for better prediction.…”
Section: A Modularized Neural Video Codingmentioning
confidence: 99%
“…In [16,17], a fully connected neural networks was used to construct the relation between boundary reconstructed pixels and original pixels in current block, and about 2%-4% bit-rate saving could be achieved. The researchers in [18] applied the recurrent neural network to progressively generate the prediction signal of current block, and improved the coding efficiency obviously. At almost the same time, the attention mechanism have become one hotspot in the literature [19][20][21][22].…”
Section: Introductionmentioning
confidence: 99%
“…Dumas et al [5] stated that using convolutional neural network (CNN) performs better than FC for blocks larger than 8×8. Hu et al [6] presented a new structure based on recurrent neural network (RNN). Sun et al [7] studied different combination schemes of traditional modes (TM) and neural network modes (NM) for the fixed block 8×8.…”
Section: Introductionmentioning
confidence: 99%
“…First is that TM still remains in the coding framework. In [4], [5] and [6], one or two NMs were provided. In [7], at most seven NMs were exploited.…”
Section: Introductionmentioning
confidence: 99%