Proceedings of the Conference Recent Advances in Natural Language Processing - Deep Learning for Natural Language Processing Me 2021
DOI: 10.26615/978-954-452-072-4_090
|View full text |Cite
|
Sign up to set email alerts
|

Developing a Clinical Language Model for Swedish: Continued Pretraining of Generic BERT with In-Domain Data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 9 publications
0
3
0
Order By: Relevance
“…4, two different PLMs are used. One-SweDeClin-BERT-that has been trained using pseudonymized pre-training data [22], and another model-SweClin-BERT-that was trained on the unaltered version of the same dataset [51]. Both models were initialized using weights from the Swedish general-domain KB-BERT model [52] and were adapted to the clinical domain by pre-training for three epochs over the Health Bank corpus.…”
Section: Clinical Bert Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…4, two different PLMs are used. One-SweDeClin-BERT-that has been trained using pseudonymized pre-training data [22], and another model-SweClin-BERT-that was trained on the unaltered version of the same dataset [51]. Both models were initialized using weights from the Swedish general-domain KB-BERT model [52] and were adapted to the clinical domain by pre-training for three epochs over the Health Bank corpus.…”
Section: Clinical Bert Modelsmentioning
confidence: 99%
“…4 This study uses two different clinical BERT models created in earlier studies. SweClin-BERT is trained on a sensitive version of the Health Bank corpus [51], whereas SweDeClin-BERT is trained on a version that has been automatically pseudonymized [22]. Both models are initialized with the weights of KB-BERT [52] which is comparable to the 3.3 billion words used to train KB-BERT [3].…”
Section: Clinical Bert Modelsmentioning
confidence: 99%
“…Since the data is from a Swedish hospital, we use a clinical BERT model for Swedish. Clinical KB-BERT [10] is a domain-adapted version of a generic language model for Swedish that has been further pre-trained on 17.8 GB of Swedish clinical text. Representations from Clinical KB-BERT are then concatenated with the structured EHR data in a fullyconnected neural network.…”
Section: B Multimodal Mortality Prediction Modelmentioning
confidence: 99%