2021
DOI: 10.5626/ktcp.2021.27.1.7
|View full text |Cite
|
Sign up to set email alerts
|

DramaQA: Character-Centered Video Story Understanding with Hierarchical QA

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 11 publications
(12 citation statements)
references
References 0 publications
0
12
0
Order By: Relevance
“…Three QA tasks are used for the verification of the proposed model: RACE [36], QuAC [37], and DramaQA [9]. RACE is a data set for the multiple choice QA where a question is composed of a questionary sentence, an associated passage, and a set of candidate answers.…”
Section: Experimental Settingmentioning
confidence: 99%
See 2 more Smart Citations
“…Three QA tasks are used for the verification of the proposed model: RACE [36], QuAC [37], and DramaQA [9]. RACE is a data set for the multiple choice QA where a question is composed of a questionary sentence, an associated passage, and a set of candidate answers.…”
Section: Experimental Settingmentioning
confidence: 99%
“…In this section, we solve the question answering with a predicted difficulty level to verify that the performance of question answering is improved with the difficulty level. We choose the multi-level context matching model [9] as a question answering model, since is currently the state-of-the-art model for the DramaQA QA task. The multi-level context matching model is designed to understand the multimodal story of a drama.…”
Section: Performance Of Question Answering With Difficulty Levelmentioning
confidence: 99%
See 1 more Smart Citation
“…Thanks to this, visionlanguage tasks such as image captioning (Xu et al, 2015), visual question answering (VQA) (Antol et al, 2015;Goyal et al, 2017), and visual commonsense reasoning (VCR) (Zellers et al, 2019) have been introduced to the research community, along with some benchmark datasets. In particular, video question answering (video QA) tasks (Xu et al, 2016;Jang et al, 2017;Yu et al, 2019;Choi et al, 2020) have been proposed with the goal of reasoning over higher-level visionlanguage interactions. In contrast to QA tasks based on static images, the questions presented in the video QA dataset vary from frame-level questions regarding the appearance of objects (e.g., what is the color of the hat?)…”
Section: Introductionmentioning
confidence: 99%
“…Recently, researches on text-to-video retrieval, video captioning, and video question answering (videoQA) have been actively conducted to improve video understanding intelligence. In addition, large-scale datasets have been built and publicly available to facilitate the researches (Alamri et al 2019;Lei et al 2018Lei et al , 2020Choi et al 2021). Studies using these datasets usually apply automatic evaluation metrics to measure the performances of AI agents.…”
Section: Introductionmentioning
confidence: 99%