Proceedings of the Sixth Workshop on Noisy User-Generated Text (W-Nut 2020) 2020
DOI: 10.18653/v1/2020.wnut-1.67
|View full text |Cite
|
Sign up to set email alerts
|

EdinburghNLP at WNUT-2020 Task 2: Leveraging Transformers with Generalized Augmentation for Identifying Informativeness in COVID-19 Tweets

Abstract: Twitter has become an important communication channel in times of emergency. The ubiquitousness of smartphones enables people to announce an emergency they're observing in real-time. Because of this, more agencies are interested in programatically monitoring Twitter (disaster relief organizations and news agencies) and therefore recognizing the informativeness of a tweet can help filter noise from large volumes of data. In this paper, we present our submission for WNUT-2020 Task 2: Identification of informativ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(1 citation statement)
references
References 15 publications
0
1
0
Order By: Relevance
“…This dataset was created for a shared task in WNUT-2020 on classifying tweets as "informative" or "uninformative", in providing information about "recovered, suspected, confirmed and death cases as well as location or travel history of the cases" [63]. Many of the models submitted in the competition were based on pre-trained transformer-based language models [3,12,33,33,53,61,61,67,87,91], including the first place model which was an ensemble of CT-BERT and RoBERTa [42], and the tied first place model which was a carefully tuned version of CT-BERT [58]. A 70-10-20 split was provided by the task organizers.…”
Section: 23mentioning
confidence: 99%
“…This dataset was created for a shared task in WNUT-2020 on classifying tweets as "informative" or "uninformative", in providing information about "recovered, suspected, confirmed and death cases as well as location or travel history of the cases" [63]. Many of the models submitted in the competition were based on pre-trained transformer-based language models [3,12,33,33,53,61,61,67,87,91], including the first place model which was an ensemble of CT-BERT and RoBERTa [42], and the tied first place model which was a carefully tuned version of CT-BERT [58]. A 70-10-20 split was provided by the task organizers.…”
Section: 23mentioning
confidence: 99%