2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR) 2019
DOI: 10.1109/mipr.2019.00109
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Indicators of Humor in Videos

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 14 publications
0
5
0
Order By: Relevance
“…Kayatani et al [16] also uses the same TV drama series as their testbed and presents a model to predict whether an utterance of a character causes laughter based on subtitles as well as facial features and the identity of the character. Yang et al [18] obtains humor labels in videos based on user comments together with visual and audio features. The ground-truth humor labels in these methods are mainly associated with texts and a prediction is made for a sentence.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Kayatani et al [16] also uses the same TV drama series as their testbed and presents a model to predict whether an utterance of a character causes laughter based on subtitles as well as facial features and the identity of the character. Yang et al [18] obtains humor labels in videos based on user comments together with visual and audio features. The ground-truth humor labels in these methods are mainly associated with texts and a prediction is made for a sentence.…”
Section: Related Workmentioning
confidence: 99%
“…In recent years, some methods have been proposed to predict humor using both single modality and multiple modalities, which are often accompanied by a dedicated dataset [7][8][9][10][11][12]. Single modal humor prediction mainly uses the linguistic modality [13][14][15], while multiple modal humor prediction combines the information from different modalities [6,[16][17][18]. The ground-truth labels of these methods are usually associated with blocks of text, like sentences and dialogues, while signals from other modalities are often treated as supplementary.…”
Section: Introductionmentioning
confidence: 99%
“…MaSaC [40] consists of Hindi-English code-mixed sitcom dialogues manually annotated for the presence of humour as well as sarcasm. A different approach is presented in [41], where the authors obtained humour labels by exploiting time-aligned user comments for videos on the Chinese video platform Bilibili. Hasan et al [17] compile their dataset UR-Funny from TED talk recordings, using laughter markup in the provided transcripts to automatically label punchline sentences in the recorded talks.…”
Section: Multimodal Humour Recognitionmentioning
confidence: 99%
“…For Open Mic, Mittal et al [42] collected standup comedy recordings and used the audience's laughter to create annotations indicating the degree of humour on a scale from zero to four. Similar to text-only datasets, most multimodal datasets are in English, notable exceptions being the already mentioned MUMOR-ZH [19], MaSaC [40], the Chinese dataset used in [41] and M2H2 [43], which is based on a Hindi TV show.…”
Section: Multimodal Humour Recognitionmentioning
confidence: 99%
See 1 more Smart Citation