Proceedings of the Fourteenth Workshop on Semantic Evaluation 2020
DOI: 10.18653/v1/2020.semeval-1.199
|View full text |Cite
|
Sign up to set email alerts
|

ANDES at SemEval-2020 Task 12: A Jointly-trained BERT Multilingual Model for Offensive Language Detection

Abstract: This paper describes our participation in SemEval-2020 Task 12: Multilingual Offensive Language Detection. We jointly-trained a single model by fine-tuning Multilingual BERT to tackle the task across all the proposed languages: English, Danish, Turkish, Greek and Arabic. Our single model had competitive results, with a performance close to top-performing systems in spite of sharing the same parameters across all languages. Zero-shot and few-shot experiments were also conducted to analyze the transference perfo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 22 publications
0
7
0
Order By: Relevance
“…Pre Trained LLM: After the preprocessing the dataset three deep pretrained models for spanish language were used: i) Py-sentimiento, a Robertuito and RoBERTa-based model trained in spanish tweets (48)(49)(50), ii) a customized version of VADER (Valence Aware Dictionary and sEntiment Reasoner) wich translate comments and classify the text polarity ( 51) and iii) Distilbert-based model, a small, fast and light Transformer by distilling BERT base (52,53).…”
Section: Methodsmentioning
confidence: 99%
“…Pre Trained LLM: After the preprocessing the dataset three deep pretrained models for spanish language were used: i) Py-sentimiento, a Robertuito and RoBERTa-based model trained in spanish tweets (48)(49)(50), ii) a customized version of VADER (Valence Aware Dictionary and sEntiment Reasoner) wich translate comments and classify the text polarity ( 51) and iii) Distilbert-based model, a small, fast and light Transformer by distilling BERT base (52,53).…”
Section: Methodsmentioning
confidence: 99%
“…For each retrieved tweet, we analyzed the hate speech content using pysentimiento (Pérez et al, 2021), a Python toolkit providing advanced classification algorithms in opinion mining tasks including sentiment analysis, emotion detection, and hate speech detection. Powered by pre-trained language models for several languages (Cañete et al, 2020;Devlin et al, 2019;Pérez et al, 2022) and the the Transformer architecture (Vaswani et al, 2017), the models are executed and distributed through the huggingface library (Wolf et al, 2020).…”
Section: Hate Speech Classification Methodsmentioning
confidence: 99%
“…This adaptation is rooted in RoBERTa, an improved version of BERT distinguished by augmented data and extended training iterations. Pérez et al (2022) introduced Robertuito to enhance sentiment analysis within the Pysentimento framework. This comprehensive toolkit is designed to provide users with the capability to perform sentiment analysis on their proprietary data or make use of alternative models available in Pysentimento or the Hugging Face Transformers library.…”
Section: Pysentimentomentioning
confidence: 99%