2023
DOI: 10.1016/j.bja.2023.04.017
|View full text |Cite
|
Sign up to set email alerts
|

Assessment of ChatGPT success with specialty medical knowledge using anaesthesiology board examination practice questions

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 31 publications
(7 citation statements)
references
References 3 publications
0
4
0
Order By: Relevance
“…Similar findings were observed in other board certification examinations. 27 In general, highly specialized models with questions that are answerable by yes or no answers are more accurate, 36 whereas highly versatile models are less accurate. 37 An LLM is a multifunctional model that has not been trained in a specific domain; however, it is possible to fine-tune these models on specific tasks or domains to improve their performance in those areas.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Similar findings were observed in other board certification examinations. 27 In general, highly specialized models with questions that are answerable by yes or no answers are more accurate, 36 whereas highly versatile models are less accurate. 37 An LLM is a multifunctional model that has not been trained in a specific domain; however, it is possible to fine-tune these models on specific tasks or domains to improve their performance in those areas.…”
Section: Discussionmentioning
confidence: 99%
“…A previous study on LLMs used similar methods, making the LLM answer the same questions twice and evaluating the agreement. 27 The primary outcome was the proportion of correct answers to the questions without images, under the same conditions as those encountered by examinees. Secondary outcomes included correct answers to all answerable questions, those with images, and those with stand-alone and scenario-based items.…”
Section: Llmmentioning
confidence: 99%
“…Several studies focused on ChatGPT's performance in medical knowledge tests, including licensing examinations for physicians, anesthesia, ophthalmology, neurology, and other specialty examinations [31][32][33][34]. Overall, ChatGPT demonstrated passing scores in most countries' licensing and specialty exams, but generally scored only slightly above the passing line, and did not achieve accuracy rates above 95% in any licensing exam.…”
Section: Medical Exam Performance and Exam Preparation With Chatgptmentioning
confidence: 99%
“…Although the sample size is insufficient to generalize the findings to other fields of expertise or users, it did provide a paradigm of ChatGPT-assisted training in new knowledge and techniques. We can foresee the flourishing development of AI chatbots being applied in medicine or health care [28][29][30][31][32][33][34][35][36]. ChatGPT represents a paradigm shift in the field of virtual assistants.…”
Section: Principal Findingsmentioning
confidence: 99%