2020
DOI: 10.1109/access.2020.3046194
|View full text |Cite
|
Sign up to set email alerts
|

Learning-Based JND-Directed HDR Video Preprocessing for Perceptually Lossless Compression With HEVC

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 21 publications
0
5
0
Order By: Relevance
“…The above network is trained once for the QP27 case and applied to other QPs. On the contrary, in HDR-JNDNet [17], Ki et al generate the training data and train the JND prefiltering network for each QP individually, which is time-consuming and inconvenient for practical use. In this work, from our experiences, we found that the best scale for the same area will be similar in the IQA-guided selection under different QPs, which implies that we could train only one model in the base QP and apply it on different QPs directly.…”
Section: B Model Consideration For Different Qpsmentioning
confidence: 99%
See 2 more Smart Citations
“…The above network is trained once for the QP27 case and applied to other QPs. On the contrary, in HDR-JNDNet [17], Ki et al generate the training data and train the JND prefiltering network for each QP individually, which is time-consuming and inconvenient for practical use. In this work, from our experiences, we found that the best scale for the same area will be similar in the IQA-guided selection under different QPs, which implies that we could train only one model in the base QP and apply it on different QPs directly.…”
Section: B Model Consideration For Different Qpsmentioning
confidence: 99%
“…With this, they trained a CNN-based JND prefiltering model with a one-QP one-model approach for different QPs, called CNN-JNQD. They also extended ERJND and applied it to HDR video (HDR-JNDNet) [17]. However, their 8x8 block-based method could lead to blocking effects.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…First, we use a multiprocess to improve the average response time and calculate the number of pictures cut out of the video at the speed of 30 fps. Then, based on equal difference sampling, the tolerance is calculated when the cardinality is fixed, and the image is selected from the starting frame [30]. Through isometric selection, we not only obtain the core motion information of sign language, but also reduce the dataset and the impact of unnecessary data on the experiment.…”
Section: Rgb Video Preprocessingmentioning
confidence: 99%
“…Each image is converted into a feature vector of [1,512]. The 30 images are combined into a temporal feature vector [30,1,1512], which is then used as the input of the Bi-LSTM network. The initial learning rate is 0.001, and the CrossEntropy loss is used.…”
Section: Network Pre-trainingmentioning
confidence: 99%