The 34th Annual ACM Symposium on User Interface Software and Technology 2021
DOI: 10.1145/3472749.3474778
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Instructional Video Creation from a Markdown-Formatted Tutorial

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(2 citation statements)
references
References 37 publications
0
2
0
Order By: Relevance
“…Several automatic approaches facilitated the video editing process by automatically applying edits based on predetermined markers [25], placing transitions and cuts in interview videos [9], adding visuals to audio travel podcasts [103], selecting appropriate clips for dialogue-driven scenes [60], adding lyric text to music videos [73], and placing cuts by matching the user's voice-over annotations with relevant segments of the videos [97]. Other systems bootstrapped the editing process by generating videos from documents [22,23], web pages [24,52], text-based instructions [108], recipe texts [98], and articles [61] or synthesized talking head videos of puppets [32] and used deep learning methods to automatically generate speech animations [95]. However, these automated approaches restrict the editor's control over the editing process by providing only predetermined input formats for interactions (e.g., markers, annotations), which in turn inhibits the expressiveness.…”
Section: Video Editing Systemsmentioning
confidence: 99%
“…Several automatic approaches facilitated the video editing process by automatically applying edits based on predetermined markers [25], placing transitions and cuts in interview videos [9], adding visuals to audio travel podcasts [103], selecting appropriate clips for dialogue-driven scenes [60], adding lyric text to music videos [73], and placing cuts by matching the user's voice-over annotations with relevant segments of the videos [97]. Other systems bootstrapped the editing process by generating videos from documents [22,23], web pages [24,52], text-based instructions [108], recipe texts [98], and articles [61] or synthesized talking head videos of puppets [32] and used deep learning methods to automatically generate speech animations [95]. However, these automated approaches restrict the editor's control over the editing process by providing only predetermined input formats for interactions (e.g., markers, annotations), which in turn inhibits the expressiveness.…”
Section: Video Editing Systemsmentioning
confidence: 99%
“…Some researchers focus on several key steps, such as frame composition [60], shot selection [23,25], shot cut suggestion [35]. Others tackle high-level automatic procedures with simple user interactions and take multiple videos captured by different cameras to produce a coherent video in different application scenarios [1,24,48] using different data sources [5,31,39,54]. Our system also belongs to high-level automatic creation that takes story/camera scripts as input.…”
Section: Related Workmentioning
confidence: 99%