2021
DOI: 10.1145/3478513.3480533
|View full text |Cite
|
Sign up to set email alerts
|

Camera keyframing with style and control

Abstract: HAL is a multi-disciplinary open access archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers. L'archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d'enseignement et de recherche français ou étrangers, des labor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 16 publications
(13 citation statements)
references
References 30 publications
0
13
0
Order By: Relevance
“…Such a system encodes an example sequence to a latent style code with a gating network and generates a camera sequence through generative experts that account for both the encoded style code and keyframe constraints. To perform our comparison (see Tab 1), we first took ground truth camera sequences as inputs to the SoTA system (LSTM+Gating [JCW*21]), to serve as a reference. Then for the other models, we used the textual descriptions linked to the ground truth sequences.…”
Section: Results and Experimentsmentioning
confidence: 99%
“…Such a system encodes an example sequence to a latent style code with a gating network and generates a camera sequence through generative experts that account for both the encoded style code and keyframe constraints. To perform our comparison (see Tab 1), we first took ground truth camera sequences as inputs to the SoTA system (LSTM+Gating [JCW*21]), to serve as a reference. Then for the other models, we used the textual descriptions linked to the ground truth sequences.…”
Section: Results and Experimentsmentioning
confidence: 99%
“…Later works tend to directly imitate exemplar videos to generate similar camera trajectories with SfM [42]. Jiang et al [22] use deep neural networks to extract camera behaviors from real movie clips based on toric space [26], which is helpful in imitating camera motion [58] and keyframing [21]. However, the inaccuracy of SfM and toric space estimation may severely affect their performance.…”
Section: Related Workmentioning
confidence: 99%
“…Still, two significant problems need to be further addressed in order to fulfill this task with enough high quality. First, most existing works [6,21,22,44] only study the camera behavior under fixed plots and environments that require heavy manual force to modify. Secondly, the full action space of camera movements is too large to allow effective per-frame decisions [49].…”
Section: Introductionmentioning
confidence: 99%
“…The same authors extend the work in [9] with an algorithm that classifies and recreates the style that defines the relative pose between camera and subject from an input sequence, using multiple layers of imitation learning. The work in [10] presents a framework that uses machine learning to move a drone to reproduce the style of a reference movie clip, passing through a set of given camera viewpoints. The paper [11] considers an imitation learning framework where a network predicts camera motions using a dataset of human-based videos recorded by professional cinematographers.…”
Section: Related Workmentioning
confidence: 99%