2020
DOI: 10.1007/978-3-030-58574-7_16
|View full text |Cite
|
Sign up to set email alerts
|

Aligning Videos in Space and Time

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 21 publications
(14 citation statements)
references
References 53 publications
0
14
0
Order By: Relevance
“…The learned representation has shown to be useful for multiple downstream recogni-tion and geometry estimation tasks. Instead of learning a general representation, there is a line of recent research on self-supervised learning specifically for finding correspondence [70,74,71,41,40,43,34,59]. For example, Vondrick et al [70] propose to propagate the current frame color to predict the future frame color as a pretext task to learn fine-grained correspondence between the current and future frame.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The learned representation has shown to be useful for multiple downstream recogni-tion and geometry estimation tasks. Instead of learning a general representation, there is a line of recent research on self-supervised learning specifically for finding correspondence [70,74,71,41,40,43,34,59]. For example, Vondrick et al [70] propose to propagate the current frame color to predict the future frame color as a pretext task to learn fine-grained correspondence between the current and future frame.…”
Section: Related Workmentioning
confidence: 99%
“…For example, Vondrick et al [70] propose to propagate the current frame color to predict the future frame color as a pretext task to learn fine-grained correspondence between the current and future frame. Other tasks including tracking objects [74,71] and patches [34,59] are also designed to explicitly find different levels of correspondences. But is explicit tracking task the only way to learn correspondence?…”
Section: Related Workmentioning
confidence: 99%
“…Cycle-consistency learning. Our work is influenced by cycle-consistency learning in different computer vision applications including 3D scene understanding [31,73,20,71], image alignment and translation [74,76,75,77], and space-time alignment in videos [4,62,38,17,59,32,49,37]. For example, Wang et al [62] propose to perform forward and backward tracking in time to achieve a cycleconsistency for learning temporal correspondence.…”
Section: Related Workmentioning
confidence: 99%
“…Dwibedi et al [17] formulates a temporal cycle consistency loss which aligns frames from one video to another between a pair of videos, and achieves good performance in video frame alignemnt tasks. Building on these two works, Purushwalkam et al [49] propose to track object patches inside a video and align them across videos at the same time. While these results are encouraging, both approaches learning from video pairs [17,49] require human annota-tors to provide ground-truth pairs (video-level) in training with a small scale of videos.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation