Proceedings of the 9th Workshop on Slavic Natural Language Processing 2023 (SlavicNLP 2023) 2023
DOI: 10.18653/v1/2023.bsnlp-1.3
|View full text |Cite
|
Sign up to set email alerts
|

TrelBERT: A pre-trained encoder for Polish Twitter

Wojciech Szmyd,
Alicja Kotyla,
Michał Zobniów
et al.

Abstract: Pre-trained Transformer-based models have become immensely popular amongst NLP practitioners. We present TrelBERT -the first Polish language model suited for application in the social media domain. TrelBERT is based on an existing general-domain model and adapted to the language of social media by pre-training it further on a large collection of Twitter data. We demonstrate its usefulness by evaluating it in the downstream task of cyberbullying detection, in which it achieves state-of-the-art results, outperfo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
references
References 15 publications
0
0
0
Order By: Relevance