2019
DOI: 10.1109/access.2019.2961383
|View full text |Cite
|
Sign up to set email alerts
|

Mixpred: Video Prediction Beyond Optical Flow

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 22 publications
0
4
0
Order By: Relevance
“…Convolutional networks, considered as feed-forward neural networks, are also commonly used in future prediction problems. A multi-model is defined in [28] to model dynamic patterns and learn image representation by combining temporal and spatial sub-networks. In [29], Deep Voxel Flow (DVF) is trained to synthesize future frames by flowing pixel values directly from input frames.…”
Section: Related Workmentioning
confidence: 99%
“…Convolutional networks, considered as feed-forward neural networks, are also commonly used in future prediction problems. A multi-model is defined in [28] to model dynamic patterns and learn image representation by combining temporal and spatial sub-networks. In [29], Deep Voxel Flow (DVF) is trained to synthesize future frames by flowing pixel values directly from input frames.…”
Section: Related Workmentioning
confidence: 99%
“…This method shows the effectiveness of modelling coherent spatio-temporal fields. Mixpred neural network has been proposed to model the dynamic pattern and learn appearance representations based on given video frames in [ 25 ]. A 3D CNN is utilized into RNN in [ 26 ], which extends representations in temporal dimension and makes the memory unit store better long-term representations.…”
Section: Literature Reviewmentioning
confidence: 99%
“…A 3D CNN is utilized into RNN in [ 26 ], which extends representations in temporal dimension and makes the memory unit store better long-term representations. However, convolutional operations [ 24 , 25 , 26 ] account for short-range intraframe dependencies due to their limited receptive fields and the lack of explicit inter-frame modeling capabilities. The generative adversarial networks [ 8 ] is another approach for spatiotemporal prediction.…”
Section: Literature Reviewmentioning
confidence: 99%
“…The other approach in future frame prediction is a multi-model composed of the above-mentioned approaches. Yan et al [26] studied future frame prediction with a multi-model composed of the temporal and spatial sub-network. Kim et al [9] detected key-points in input frames, and translated the frame according to key-points motion.…”
Section: Related Workmentioning
confidence: 99%