Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-2006
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Task Learning for Argumentation Mining in Low-Resource Settings

Abstract: We investigate whether and where multitask learning (MTL) can improve performance on NLP problems related to argumentation mining (AM), in particular argument component identification. Our results show that MTL performs particularly well (and better than single-task learning) when little training data is available for the main task, a common scenario in AM. Our findings challenge previous assumptions that conceptualizations across AM datasets are divergent and that MTL is difficult for semantic or higher-level… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
75
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 56 publications
(80 citation statements)
references
References 24 publications
4
75
1
Order By: Relevance
“…Related to the study of diagnostic reasoning is argumentative reasoning, which has recently received growing attention from the NLP community. The focus has been on identifying argument components (Lippi and Torroni 2015;Schulz et al 2018) or whole arguments, made of components (such as premises and claims) as well as attacking and supporting relations between them (Menini et al 2018;Habernal and Gurevych 2017). Like us, Stab and Gurevych (2014) and Nguyen and Litman (2018) investigate arguments in an educational setting by automatically identifying arguments in students' persuasive essays.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Related to the study of diagnostic reasoning is argumentative reasoning, which has recently received growing attention from the NLP community. The focus has been on identifying argument components (Lippi and Torroni 2015;Schulz et al 2018) or whole arguments, made of components (such as premises and claims) as well as attacking and supporting relations between them (Menini et al 2018;Habernal and Gurevych 2017). Like us, Stab and Gurevych (2014) and Nguyen and Litman (2018) investigate arguments in an educational setting by automatically identifying arguments in students' persuasive essays.…”
Section: Related Workmentioning
confidence: 99%
“…Challenges C1 and C2 imply that we are dealing with a multiclass sequence labeling task, as for example encountered in the related task of argument component identification (Schulz et al 2018). This is commonly modeled by assigning a label to each token that expresses both the type of segment, here the type of epistemic activity A = {HG, EG, EE, DC}, and the segment boundaries in terms of BIO-labels S = {B, I, O}, indicating the beginning (B), continuation (I), or absence (O) of a segment.…”
Section: Modeling the Taskmentioning
confidence: 99%
“…Users in social media platforms usually express emotions or quick messages with very little argumentation, however the introduction of argumentative features can enhance other NLP tasks [60,66]. Both micro [93] and macro [94] analysis have the attention of the research community, whereas they have been approaches that combine them [87,88]. Another research topic that has gained the interest of the research community is the reconstruction of implicit warrants, although the existing research papers [8,9,29] do not utilize social media as source.…”
Section: Relations Identificationmentioning
confidence: 99%
“…Deep learning techniques are able to handle a great volume of data in an unsupervised or semi-supervised way and they have achieved break-trough results in NLP field. Deep learning has been applied [93,39,9] in AM, but does not seem to overpass other ML algorithms, mainly because of the limited available datasets, however more research should take place in order safe conclusion to be drawn.…”
Section: Future Directions: Semi-supervision and Background Knowledgementioning
confidence: 99%
“…However, since T is finite, overfitting the training set might lead to poor generalization performance. One way to avoid fitting Equation 1 too # train # dev ES Bollmann et al (2018) 5k 12k-46k Yes 400-700 100-200 Yes Makarov and Clematide (2018) 100 1k Yes Sharma et al (2018) 100 100 Yes Schulz et al (2018) 1k-21k 9k N/A Upadhyay et al (2018) 500 1k Yes closely is early stopping: a separate development or validation set is used to end training as soon as the loss on the development set L D (θ) starts increasing or model performance on the development set D starts decreasing. The best set of parameters θ is used in the final model.…”
Section: Introductionmentioning
confidence: 99%