Is ChatGPT 3.5 smarter than Otolaryngology trainees? A comparison study of board style exam questions

dc.contributor.authorPatel, Jaimin
dc.contributor.authorRobinson, Peyton
dc.contributor.authorIlling, Elisa
dc.contributor.authorAnthony, Benjamin
dc.contributor.departmentOtolaryngology -- Head and Neck Surgery, School of Medicine
dc.date.accessioned2024-10-31T09:52:28Z
dc.date.available2024-10-31T09:52:28Z
dc.date.issued2024-09-26
dc.description.abstractObjectives: This study compares the performance of the artificial intelligence (AI) platform Chat Generative Pre-Trained Transformer (ChatGPT) to Otolaryngology trainees on board-style exam questions. Methods: We administered a set of 30 Otolaryngology board-style questions to medical students (MS) and Otolaryngology residents (OR). 31 MSs and 17 ORs completed the questionnaire. The same test was administered to ChatGPT version 3.5, five times. Comparisons of performance were achieved using a one-way ANOVA with Tukey Post Hoc test, along with a regression analysis to explore the relationship between education level and performance. Results: The average scores increased each year from MS1 to PGY5. A one-way ANOVA revealed that ChatGPT outperformed trainee years MS1, MS2, and MS3 (p = <0.001, 0.003, and 0.019, respectively). PGY4 and PGY5 otolaryngology residents outperformed ChatGPT (p = 0.033 and 0.002, respectively). For years MS4, PGY1, PGY2, and PGY3 there was no statistical difference between trainee scores and ChatGPT (p = .104, .996, and 1.000, respectively). Conclusion: ChatGPT can outperform lower-level medical trainees on Otolaryngology board-style exam but still lacks the ability to outperform higher-level trainees. These questions primarily test rote memorization of medical facts; in contrast, the art of practicing medicine is predicated on the synthesis of complex presentations of disease and multilayered application of knowledge of the healing process. Given that upper-level trainees outperform ChatGPT, it is unlikely that ChatGPT, in its current form will provide significant clinical utility over an Otolaryngologist.
dc.eprint.versionFinal published version
dc.identifier.citationPatel J, Robinson P, Illing E, Anthony B. Is ChatGPT 3.5 smarter than Otolaryngology trainees? A comparison study of board style exam questions. PLoS One. 2024;19(9):e0306233. Published 2024 Sep 26. doi:10.1371/journal.pone.0306233
dc.identifier.urihttps://hdl.handle.net/1805/44380
dc.language.isoen_US
dc.publisherPublic Library of Science
dc.relation.isversionof10.1371/journal.pone.0306233
dc.relation.journalPLoS One
dc.rightsAttribution 4.0 Internationalen
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/
dc.sourcePMC
dc.subjectArtificial intelligence
dc.subjectClinical competence
dc.subjectEducational measurement
dc.subjectOtolaryngology
dc.titleIs ChatGPT 3.5 smarter than Otolaryngology trainees? A comparison study of board style exam questions
dc.typeArticle
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Patel2024Chat-CCBY.pdf
Size:
453.1 KB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.04 KB
Format:
Item-specific license agreed upon to submission
Description: