How Can the Clinical Aptitude of AI Assistants Be Assayed?
AI
ChatGPT
LLM
artificial general intelligence
artificial intelligence
barrier
barriers
challenge
challenges
chatbot
chatbots
clinical decision aid
conversational agent
conversational agents
foundation models
implementation
language model
large language models
pain point
pain points
pitfall
pitfalls
validation
Journal
Journal of medical Internet research
ISSN: 1438-8871
Titre abrégé: J Med Internet Res
Pays: Canada
ID NLM: 100959882
Informations de publication
Date de publication:
05 Dec 2023
05 Dec 2023
Historique:
received:
04
08
2023
accepted:
20
11
2023
revised:
28
09
2023
medline:
7
12
2023
pubmed:
5
12
2023
entrez:
5
12
2023
Statut:
epublish
Résumé
Large language models (LLMs) are exhibiting remarkable performance in clinical contexts, with exemplar results ranging from expert-level attainment in medical examination questions to superior accuracy and relevance when responding to patient queries compared to real doctors replying to queries on social media. The deployment of LLMs in conventional health care settings is yet to be reported, and there remains an open question as to what evidence should be required before such deployment is warranted. Early validation studies use unvalidated surrogate variables to represent clinical aptitude, and it may be necessary to conduct prospective randomized controlled trials to justify the use of an LLM for clinical advice or assistance, as potential pitfalls and pain points cannot be exhaustively predicted. This viewpoint states that as LLMs continue to revolutionize the field, there is an opportunity to improve the rigor of artificial intelligence (AI) research to reward innovation, conferring real benefits to real patients.
Identifiants
pubmed: 38051572
pii: v25i1e51603
doi: 10.2196/51603
doi:
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM
Pagination
e51603Informations de copyright
©Arun James Thirunavukarasu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 05.12.2023.