2023
DOI: 10.2106/jbjs.oa.23.00056
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating ChatGPT Performance on the Orthopaedic In-Training Examination

Justin E. Kung,
Christopher Marshall,
Chase Gauthier
et al.

Abstract: Background: Artificial intelligence (AI) holds potential in improving medical education and healthcare delivery. ChatGPT is a state-of-the-art natural language processing AI model which has shown impressive capabilities, scoring in the top percentiles on numerous standardized examinations, including the Uniform Bar Exam and Scholastic Aptitude Test. The goal of this study was to evaluate ChatGPT performance on the Orthopaedic In-Training Examination (OITE), an assessment of medical knowledge for or… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

2
19
0
1

Year Published

2024
2024
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 37 publications
(22 citation statements)
references
References 13 publications
2
19
0
1
Order By: Relevance
“…This is particularly alarming considering that the access is free for this version, which is the most used. The substantial increase in performance of version 4 is consistent with other studies 5,[7][8][9]15 and likely stems from the foundational differences in training and algorithmic sophistication between the two versions. 1 It may partially be attributed to the integration of a rule-based reward model in version 4.…”
Section: Discussionsupporting
confidence: 87%
See 4 more Smart Citations
“…This is particularly alarming considering that the access is free for this version, which is the most used. The substantial increase in performance of version 4 is consistent with other studies 5,[7][8][9]15 and likely stems from the foundational differences in training and algorithmic sophistication between the two versions. 1 It may partially be attributed to the integration of a rule-based reward model in version 4.…”
Section: Discussionsupporting
confidence: 87%
“…Several studies have assessed the reliability of ChatGPT, each using different methodologies and subsequently reporting varying levels of performance 5–11,15 . Studies using common questions of interest to patients and laypeople usually show high accuracy rates.…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations