2024
DOI: 10.1177/20552076241237678
|View full text |Cite
|
Sign up to set email alerts
|

Exploring the proficiency of ChatGPT-4: An evaluation of its performance in the Taiwan advanced medical licensing examination

Shih-Yi Lin,
Pak Ki Chan,
Wu-Huei Hsu
et al.

Abstract: Background Taiwan is well-known for its quality healthcare system. The country's medical licensing exams offer a way to evaluate ChatGPT's medical proficiency. Methods We analyzed exam data from February 2022, July 2022, February 2023, and July 2033. Each exam included four papers with 80 single-choice questions, grouped as descriptive or picture-based. We used ChatGPT-4 for evaluation. Incorrect answers prompted a “chain of thought” approach. Accuracy rates were calculated as percentages. Results ChatGPT-4's … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 36 publications
0
4
0
Order By: Relevance
“…Researchers have not analyzed or elaborated on the impact of these task understanding prompts on ChatGPT's performance. However, three studies used optimized prompts [19,26,35]. A Korean study used four kinds of optimized prompts including: annotating Chinese terms in TKM, translating the instruction and question into English, providing exam-optimized instructions, and utilizing selfconsistency in the prompt.…”
Section: Figure 5 Performance Of Chatgpt On Passing Medical Licensing...mentioning
confidence: 99%
See 2 more Smart Citations
“…Researchers have not analyzed or elaborated on the impact of these task understanding prompts on ChatGPT's performance. However, three studies used optimized prompts [19,26,35]. A Korean study used four kinds of optimized prompts including: annotating Chinese terms in TKM, translating the instruction and question into English, providing exam-optimized instructions, and utilizing selfconsistency in the prompt.…”
Section: Figure 5 Performance Of Chatgpt On Passing Medical Licensing...mentioning
confidence: 99%
“…Pretend to be a junior doctor with expertise in clinical practice and exam solving and retry" or "Could you double-check the answer?". ChatGPT could correctly answer up to 88.9% and 84% of these questions, respectively [19,35]. For task understanding prompts, we conducted a subgroup analysis and meta-regression to examine whether they affected ChatGPT's performance.…”
Section: Figure 5 Performance Of Chatgpt On Passing Medical Licensing...mentioning
confidence: 99%
See 1 more Smart Citation
“…Multiple studies have systematically evaluated ChatGPT's performance on standardized tests across various languages. Notably, it has demonstrated excellent performance on assessments such as the United States Medical Licensing Examination (USMLE) [11][12][13] , the Japanese Medical Licensing Examination (JMLE) 14 , the Saudi Medical Licensing Examination (SMLE) 15 , the Polish medical specialization licensing exam (PES) 16 and Taiwan's medical licensing exams 17 . However, over the past ve years of the Chinese National Medical Licensing Examination (NMLE), ChatGPT scores have consistently fallen below the passing threshold.…”
Section: Introductionmentioning
confidence: 99%