Performance of ChatGPT on dermatology Specialty Certificate Examination multiple choice questions.
Journal
Clinical and experimental dermatology
ISSN: 1365-2230
Titre abrégé: Clin Exp Dermatol
Pays: England
ID NLM: 7606847
Informations de publication
Date de publication:
02 Jun 2023
02 Jun 2023
Historique:
received:
25
05
2023
accepted:
26
05
2023
medline:
2
6
2023
pubmed:
2
6
2023
entrez:
2
6
2023
Statut:
aheadofprint
Résumé
ChatGPT is a large language model trained on increasingly large datasets by OpenAI to perform language-based tasks. It is capable of answering multiple-choice questions, such as those posed by the dermatology SCE examination. We asked two iterations of ChatGPT: ChatGPT-3.5 and ChatGPT-4 84 multiple-choice sample questions from the sample dermatology SCE question bank. ChatGPT-3.5 achieved an overall score of 63.1%, and ChatGPT-4 scored 90.5% (a significant improvement in performance (p<0.001)). The typical pass mark for the dermatology SCE is 70-72%. ChatGPT-4 is therefore capable of answering clinical questions and achieving a passing grade in these sample questions. There are many possible educational and clinical implications for increasingly advanced artificial intelligence (AI) and its use in medicine, including in the diagnosis of dermatological conditions. Such advances should be embraced provided that patient safety is a core tenet, and the limitations of AI in the nuances of complex clinical cases are recognised.
Identifiants
pubmed: 37264670
pii: 7188526
doi: 10.1093/ced/llad197
pii:
doi:
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM
Informations de copyright
© The Author(s) 2023. Published by Oxford University Press on behalf of British Association of Dermatologists. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.