2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2023
DOI: 10.1109/wacv56688.2023.00502
|View full text |Cite
|
Sign up to set email alerts
|

Enhanced Bi-directional Motion Estimation for Video Frame Interpolation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
18
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(18 citation statements)
references
References 34 publications
0
18
0
Order By: Relevance
“…Existing mainstream methods of video frame interpolation can be generally categorized into flow-based methods [7][8][9]11,12,[15][16][17][18][19][20][21][22][23][24][25][26], kernel-based methods [1,5,10,[27][28][29][30], phase-based methods [31], and hallucination-based methods [2,3,14,32].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Existing mainstream methods of video frame interpolation can be generally categorized into flow-based methods [7][8][9]11,12,[15][16][17][18][19][20][21][22][23][24][25][26], kernel-based methods [1,5,10,[27][28][29][30], phase-based methods [31], and hallucination-based methods [2,3,14,32].…”
Section: Related Workmentioning
confidence: 99%
“…VFIformer [23] refines the warped frames with a Transformer [34] structure. UPR-Net [24] uses a unified pyramid recurrent network to estimate optical flow and synthesize frames. AMT [25] leverages bidirectional correlation volumes for all pairs of pixels in flow estimation and feature update.…”
Section: Related Workmentioning
confidence: 99%
“…where k ≤ c o and m, ñ indicate subtracting half of the kernel size, i.e., m = m − ⌊ k h 2 ⌋. To reduce the model size and improve generalization, some models reuse the same convolution block multiple times (Jin et al 2023;Sim, Oh, and Kim 2021;Teed and Deng 2020). However, the standard convolution operation requires the input and output to always have the same predefined number of channels, which hinders the ability to handle diverse inputs.…”
Section: Partial Kernel Convolution (Pkconv)mentioning
confidence: 99%
“…However, recurrent encoders in the spatial dimension have yet to be wellexplored for optical flow. Some works adopted it for video interpolation to generate multi-scale features, but that required either producing features of a fixed size (Sim, Oh, and Kim 2021) or treating the entire optical flow network as a recurrent unit (Jin et al 2023;Zhang, Zhao, and Wang 2020), which decreases the flexibility of the model.…”
Section: Recurrent Models For Optical Flowmentioning
confidence: 99%
“…Previous crossbreed coding schemes depended on each pixel-space operation to reduce repetitive actions in space and time [22]. Still, these systems could need help with motion estimation or motion pay [23]. The work built an element-fieldcoding video network (FCV) by modelling every significant operation in the component space (motion estimation, motion compression, motion remuneration, and lingering compression).…”
Section: Related Workmentioning
confidence: 99%