Proceedings of the Fourteenth Workshop on Semantic Evaluation 2020
DOI: 10.18653/v1/2020.semeval-1.228
|View full text |Cite
|
Sign up to set email alerts
|

Hitachi at SemEval-2020 Task 11: An Empirical Study of Pre-Trained Transformer Family for Propaganda Detection

Abstract: In this paper, we show our system for SemEval-2020 task 11, where we tackle propaganda span identification (SI) and technique classification (TC). We investigate heterogeneous pre-trained language models (PLMs) such as BERT, GPT-2, XLNet, XLM, RoBERTa, and XLM-RoBERTa for SI and TC fine-tuning, respectively. In large-scale experiments, we found that each of the language models has a characteristic property, and using an ensemble model with them is promising. Finally, the ensemble model was ranked 1st amongst 3… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 21 publications
0
8
0
Order By: Relevance
“…Span prediction identifies the start and end offsets of the span (Chhablani et al 2021). Sequence labelling classifies each member of a sequence, for example, identify whether each token is toxic (Chhablani et al 2021) or use BIO encoding (i.e., mark the token as (B) if it is at the beginning, (I) if it is inside or (O) if it is outside of the span) (Morio et al 2020).…”
Section: Span Extractionmentioning
confidence: 99%
See 2 more Smart Citations
“…Span prediction identifies the start and end offsets of the span (Chhablani et al 2021). Sequence labelling classifies each member of a sequence, for example, identify whether each token is toxic (Chhablani et al 2021) or use BIO encoding (i.e., mark the token as (B) if it is at the beginning, (I) if it is inside or (O) if it is outside of the span) (Morio et al 2020).…”
Section: Span Extractionmentioning
confidence: 99%
“…Since SE tasks require highly nuanced semantic understanding, most solutions leveraged large language models pre-trained using transformers, including BERT (Devlin et al 2019) and other types of transformers (Morio et al 2020;Chhablani et al 2021). These models are pre-trained on billions of words of English text data and can be easily fine-tuned to adapt to new tasks.…”
Section: Span Extractionmentioning
confidence: 99%
See 1 more Smart Citation
“…As previously introduced, current systems address these tasks relying on word embedding models (e.g., BERT-embedding) and standard features (e.g., PoS, name-entity, n-grams), as representations to feed various RNN architectures (Morio et al, 2020;Chernyavskiy et al, 2020). Recently, the language model BERT (Devlin et al, 2019) has been widely utilized to optimize the performances of classification tasks, but there is still room for improvement, in particular when applied to propaganda detection (Da San Martino et al, 2020a.…”
Section: Related Workmentioning
confidence: 99%
“…The proposed approaches for the sub-task can be broadly classified into Span Prediction or Token Classification. Most teams use multi-granular transformer-based systems for token classification/sequence tagging (Khosla et al, 2020;Morio et al, 2020;Patil et al, 2020). Inspired by Souza et al (2019), Jurkiewicz et al (2020) use RoBERTa-CRF based systems.…”
Section: Introductionmentioning
confidence: 99%