2019
DOI: 10.1007/978-3-030-34869-4_16
|View full text |Cite
|
Sign up to set email alerts
|

Sustained Self-Supervised Pretraining for Temporal Order Verification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 17 publications
0
8
0
Order By: Relevance
“…During the initial days of self-supervised learning, a lot of research was done on handcrafting pre-training tasks, also known as pretext tasks. These handcrafted tasks include geometric transformation prediction [1,2,3], context prediction [4,5], jigsaw puzzle solving [6,7,8,9], temporal order related tasks for videos [10,11,12,13,14], pace prediction in videos [15], image colorization [16], etc. These pretext tasks are aimed at learning representations that are invariant to transformations, context, etc.…”
Section: Literature Surveymentioning
confidence: 99%
“…During the initial days of self-supervised learning, a lot of research was done on handcrafting pre-training tasks, also known as pretext tasks. These handcrafted tasks include geometric transformation prediction [1,2,3], context prediction [4,5], jigsaw puzzle solving [6,7,8,9], temporal order related tasks for videos [10,11,12,13,14], pace prediction in videos [15], image colorization [16], etc. These pretext tasks are aimed at learning representations that are invariant to transformations, context, etc.…”
Section: Literature Surveymentioning
confidence: 99%
“…Conditioning on generated samples is called auto-conditioning. Joint probability of the distribution Y can be expressed as the product of ordered conditionals, refer (5). Auto-conditioning improves on maximum-likelihood estimation (MLE) based training of RNN by allowing generated distribution Y to drift away from the ground-truth distribution X .…”
Section: Proposed Modelmentioning
confidence: 99%
“…It is a core problem since it facilitates the understanding of computational aspects of human-machine interactions, cybernetic systems, and has multiple applications ranging from human behavior analysis, security systems, augmented and virtual reality. It has drawn significant attention from the action recognition community [1]- [5]. However, due to lingering challenges of action prediction in visual domain [6], the trend is shifting from simple action classification [7]- [9] to motion prediction in disentangled domain [10]- [16].…”
Section: Introductionmentioning
confidence: 99%
“…The self-supervised learning models learn meaningful features by means of solving various tasks called Pretext tasks. Several types of pretext tasks such as image inpainting [14], temporal order correction or verification [15][16][17][18][19], clip order prediction [20], image coloring [21], geometric transformation prediction [22,23], relative patch prediction based on context-aware features [24], etc. have been proposed.…”
Section: Introductionmentioning
confidence: 99%