2021
DOI: 10.2196/preprints.32523
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unifying Heterogenous Electronic Health Records Systems via Text-Based Code Embedding: Study of Predictive Modeling (Preprint)

Abstract: BACKGROUND Substantial increase in the use of Electronic Health Records (EHRs) has opened new frontiers for predictive healthcare. However, while EHR systems are nearly ubiquitous, they lack a unified code system for representing medical concepts. Heterogeneous formats of EHR present a substantial barrier for the training and deployment of state-of-the-art deep learning models at scale. OBJECTIVE … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(14 citation statements)
references
References 26 publications
0
14
0
Order By: Relevance
“…DescEmb [25] proposed to resolve this problem by suggesting a text-based embedding, where hospital-specific feature values are first converted to textual descriptions (e.g., "401.9" → "unspecified essential hypertension"), then a text encoder paired with a sub-word tokenizer is used to obtain m i [37]. With Fig.…”
Section: B General Healthcare Predictive Frameworkmentioning
confidence: 99%
See 2 more Smart Citations
“…DescEmb [25] proposed to resolve this problem by suggesting a text-based embedding, where hospital-specific feature values are first converted to textual descriptions (e.g., "401.9" → "unspecified essential hypertension"), then a text encoder paired with a sub-word tokenizer is used to obtain m i [37]. With Fig.…”
Section: B General Healthcare Predictive Frameworkmentioning
confidence: 99%
“…Additionally, in multi-source learning, our framework is not constrained by the features that are present in each schema since both the name n k i and the value v k i of the feature are used. A formal comparison of the conventional approach, DescEmb [25] and our approach for obtaining m i is provided below:…”
Section: Employing the Entire Features Of Ehrmentioning
confidence: 99%
See 1 more Smart Citation
“…In conditions where there is a lack of data, it is possible to enhance the performance of the model [214] . The EHRsrelated tasks include prediction [33,126,[214][215][216][217][218][219][220][221][222] , information extraction from clinic notes [223][224][225][226] , the international classification of disease (ICD) coding [227,228] , medication recommendation [229,230] , etc.…”
Section: Ehrs In Pre-trainingmentioning
confidence: 99%
“…Xu et al [216] introduced the medical knowledge graph combined with self-supervised pre-training to deal with the sparsity and high-dimensional issue of EHR data. Lu et al [221] utilised a pre-trained model to detect disease complications and compute the contributions of particular diseases and admissions. Using the self-supervised learning method, the pre-trained model was trained based on the hidden disease representation.…”
Section: Eegmentioning
confidence: 99%