BioNLP 2017 2017
DOI: 10.18653/v1/w17-2338
|View full text |Cite
|
Sign up to set email alerts
|

Adapting Pre-trained Word Embeddings For Use In Medical Coding

Abstract: Word embeddings are a crucial component in modern NLP. Pre-trained embeddings released by different groups have been a major reason for their popularity. However, they are trained on generic corpora, which limits their direct use for domain specific tasks. In this paper, we propose a method to add task specific information to pre-trained word embeddings. Such information can improve their utility. We add information from medical coding data, as well as the first level from the hierarchy of ICD-10 medical code … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
12
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 18 publications
(13 citation statements)
references
References 15 publications
1
12
0
Order By: Relevance
“…From these results we can conclude that pre-trained embeddings can improve the language modeling quality, but only if training content is aligned with the task at hand. Similar results are reported by Kim [29] and Patel et al [30]: refining the Google embeddings for the described task and data improves results.…”
Section: Resultssupporting
confidence: 87%
“…From these results we can conclude that pre-trained embeddings can improve the language modeling quality, but only if training content is aligned with the task at hand. Similar results are reported by Kim [29] and Patel et al [30]: refining the Google embeddings for the described task and data improves results.…”
Section: Resultssupporting
confidence: 87%
“…This is done by back‐propagating the training errors to the embedding level, as suggested in Collobert et al (2011). For example, for the task of medical coding in Patel et al (2017), embeddings were improved with the addition of information from ICD‐10 codes.…”
Section: Discussionmentioning
confidence: 99%
“…For example, the original BERT took 4 days to pre-train on 4 TPUs. Furthermore, there are few limitations of using a non-medical corpus to train a model for medical tasks (Patel et al, 2017). There are medical-specific terms that do not usually exist in general corpora such as news or Wikipedia.…”
Section: Problem Statementmentioning
confidence: 99%