Proceedings of the 5th Clinical Natural Language Processing Workshop 2023
DOI: 10.18653/v1/2023.clinicalnlp-1.26
|View full text |Cite
|
Sign up to set email alerts
|

Generating medically-accurate summaries of patient-provider dialogue: A multi-stage approach using large language models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…It is important to stress that our evaluation of the GPT-4 model was based on single queries using prompts developed by informal experimentation. Further research should investigate more sophisticated prompting strategies such as "chain of thought" 36 for gene set analyses and the use of methods that apply external tools and orchestrate multiple LLM interactions [37][38][39][40][41][42][43] , such as integrating literature searches into the LLM analysis rather than as a post hoc verification method. Our study also makes the implicit assumption that all gene sets emerging from a biological study have a coherent function to be summarized.…”
Section: Discussionmentioning
confidence: 99%
“…It is important to stress that our evaluation of the GPT-4 model was based on single queries using prompts developed by informal experimentation. Further research should investigate more sophisticated prompting strategies such as "chain of thought" 36 for gene set analyses and the use of methods that apply external tools and orchestrate multiple LLM interactions [37][38][39][40][41][42][43] , such as integrating literature searches into the LLM analysis rather than as a post hoc verification method. Our study also makes the implicit assumption that all gene sets emerging from a biological study have a coherent function to be summarized.…”
Section: Discussionmentioning
confidence: 99%
“…The predominant method for evaluating LLMs in the medical field involves medical exam-type questions, with a strong emphasis on multiple-choice formats [16][17][18] . Although there are instances where LLMs are tested on free-response and reasoning tasks 19,20,12 , or for medical conversation summarization and care plan generation 21 , these are less common. However, these assessments do not explore LLMs' ability for engaging in interactive patient conversations, a crucial aspect of their potential role in revolutionizing healthcare delivery.…”
Section: Introductionmentioning
confidence: 99%