2019
DOI: 10.1371/journal.pone.0212488
|View full text |Cite
|
Sign up to set email alerts
|

Using natural language processing and machine learning to classify health literacy from secure messages: The ECLIPPSE study

Abstract: Limited health literacy is a barrier to optimal healthcare delivery and outcomes. Current measures requiring patients to self-report limitations are time-consuming and may be considered intrusive by some. This makes widespread classification of patient health literacy challenging. The objective of this study was to develop and validate “literacy profiles” as automated indicators of patients’ health literacy to facilitate a non-intrusive, economic and more comprehensive characterization of health literacy among… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
26
0

Year Published

2019
2019
2025
2025

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 29 publications
(26 citation statements)
references
References 73 publications
0
26
0
Order By: Relevance
“…More recently, ML has successfully been applied to more unstructured written health notes and dialogues [35,36]. One study was able to phenotype depression using unstructured notes in the medical records with a sensitivity of 93.5% and specificity of 68% [35].…”
Section: Principal Findingsmentioning
confidence: 99%
See 1 more Smart Citation
“…More recently, ML has successfully been applied to more unstructured written health notes and dialogues [35,36]. One study was able to phenotype depression using unstructured notes in the medical records with a sensitivity of 93.5% and specificity of 68% [35].…”
Section: Principal Findingsmentioning
confidence: 99%
“…One study was able to phenotype depression using unstructured notes in the medical records with a sensitivity of 93.5% and specificity of 68% [35]. Natural language processing has also been used to review secure online discussions between patients and health professionals, in which linguistic features were assessed to predict health literacy levels with an accuracy of 60.55% (C-statistic 0.63) [36]. Common to these studies, there was one ML model which performed better than the other models, and the model varied across the studies dependent on the primary data and outcome they were measuring, and this is also reflected in our results.…”
Section: Principal Findingsmentioning
confidence: 99%
“…However, with the aim of reducing the false negatives to the lowest possible number, the ensemble model was found to be superior (1.43% false negatives and 16.2% false positives), giving a sensitivity of 93.5% for identifying the SMS text messages requiring review, and a specificity of 81.3%. Using the ensemble model with heuristics, the results indicated that health professionals would have to review 36 3).…”
Section: Accuracy For Determining If Staff Review Is Requiredmentioning
confidence: 99%
“…Health information understandability can be achieved by using familiar language and good writing practice that highlights health information directness, clearness of the desired health outcome, easy-to-follow informational organization, and discourse explicitness, that is, clear explanation of health and medical knowledge using simple, plain, and purposeful language [2][3][4][5]. Approaches to health information evaluation can be divided into 2 large categories, that is, expert-led qualitative evaluation based on clinical experiences [6][7][8][9] and automated health information analyzers using medical readability formulas or natural language processing tools [10][11][12][13]. The strengths and limitations of both approaches are well-known [14][15][16].…”
Section: Introductionmentioning
confidence: 99%