Proceedings of the 8th Workshop on Argument Mining 2021
DOI: 10.18653/v1/2021.argmining-1.8
|View full text |Cite
|
Sign up to set email alerts
|

M-Arg: Multimodal Argument Mining Dataset for Political Debates with Audio and Transcripts

Abstract: Argumentation mining aims at extracting, analysing and modelling people's arguments, but large, high-quality annotated datasets are limited, and no multimodal datasets exist for this task. In this paper, we present M-Arg, a multimodal argument mining dataset with a corpus of US 2020 presidential debates, annotated through crowd-sourced annotations. This dataset allows models to be trained to extract arguments from natural dialogue such as debates using information like the intonation and rhythm of the speaker.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(9 citation statements)
references
References 24 publications
0
9
0
Order By: Relevance
“…The results were positive towards the addition of audio, although the performance was modest due to the small size of the dataset and the limitations of the text and audio representations. The only other work that considered multimodal aspects used the M-arg dataset (Mestre et al, 2021). There, the authors analyzed argumentative relations in the 2016 US presidential debates using text and audio, building an argumentation mining pipeline based on BERT embeddings for text and a combination of a Bi-LSTM and a CNN for the audio.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…The results were positive towards the addition of audio, although the performance was modest due to the small size of the dataset and the limitations of the text and audio representations. The only other work that considered multimodal aspects used the M-arg dataset (Mestre et al, 2021). There, the authors analyzed argumentative relations in the 2016 US presidential debates using text and audio, building an argumentation mining pipeline based on BERT embeddings for text and a combination of a Bi-LSTM and a CNN for the audio.…”
Section: Related Workmentioning
confidence: 99%
“…Here, we build upon a previous dataset (US-ElecDeb60To16) presented by Haddadan et al (2019), which contained English transcripts of the US presidential debates from 1960 to 2016 labelled with more than 29k annotations of argument components and their boundaries. We used the original videos from the debates to obtain aligned timestamps at the sentence level following the work of Mestre et al (2021), thus enabling the task of multimodal AM with a total of 28,850 aligned and annotated sentences. Concurrently to the submission of our work, Mancini et al (2022) also presented and released a multimodal dataset, using the same videos and alignment process, with 26,791 sentences.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations