2024
DOI: 10.2196/56655
|View full text |Cite
|
Sign up to set email alerts
|

Quality of Answers of Generative Large Language Models Versus Peer Users for Interpreting Laboratory Test Results for Lay Patients: Evaluation Study

Zhe He,
Balu Bhasuran,
Qiao Jin
et al.

Abstract: Background Although patients have easy access to their electronic health records and laboratory test result data through patient portals, laboratory test results are often confusing and hard to understand. Many patients turn to web-based forums or question-and-answer (Q&A) sites to seek advice from their peers. The quality of answers from social Q&A sites on health-related questions varies significantly, and not all responses are accurate or reliable. Large language models (LLMs) such as Ch… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(1 citation statement)
references
References 40 publications
0
1
0
Order By: Relevance
“…Previous studies have noted that the responses generated by GPT-3.5 are nondeterministic and random [67][68][69]. Our study found that although the stability of GPT-4 has significantly improved compared to that of GPT-3.5, it still exhibits a degree of randomness in its outputs.…”
Section: Challenge Of Utilizing Chatgpt In Medical Educationmentioning
confidence: 47%
“…Previous studies have noted that the responses generated by GPT-3.5 are nondeterministic and random [67][68][69]. Our study found that although the stability of GPT-4 has significantly improved compared to that of GPT-3.5, it still exhibits a degree of randomness in its outputs.…”
Section: Challenge Of Utilizing Chatgpt In Medical Educationmentioning
confidence: 47%