Dissociating language and thought in large language models.
cognitive neuroscience
computational modeling
language and thought
large language models
linguistic competence
Journal
Trends in cognitive sciences
ISSN: 1879-307X
Titre abrégé: Trends Cogn Sci
Pays: England
ID NLM: 9708669
Informations de publication
Date de publication:
19 Mar 2024
19 Mar 2024
Historique:
received:
06
11
2023
revised:
31
01
2024
accepted:
31
01
2024
medline:
21
3
2024
pubmed:
21
3
2024
entrez:
20
3
2024
Statut:
aheadofprint
Résumé
Large language models (LLMs) have come closest among all models to date to mastering human language, yet opinions about their linguistic and cognitive capabilities remain split. Here, we evaluate LLMs using a distinction between formal linguistic competence (knowledge of linguistic rules and patterns) and functional linguistic competence (understanding and using language in the world). We ground this distinction in human neuroscience, which has shown that formal and functional competence rely on different neural mechanisms. Although LLMs are surprisingly good at formal competence, their performance on functional competence tasks remains spotty and often requires specialized fine-tuning and/or coupling with external modules. We posit that models that use language in human-like ways would need to master both of these competence types, which, in turn, could require the emergence of separate mechanisms specialized for formal versus functional linguistic competence.
Identifiants
pubmed: 38508911
pii: S1364-6613(24)00027-5
doi: 10.1016/j.tics.2024.01.011
pii:
doi:
Types de publication
Journal Article
Review
Langues
eng
Sous-ensembles de citation
IM
Informations de copyright
Copyright © 2024 Elsevier Ltd. All rights reserved.
Déclaration de conflit d'intérêts
Declaration of interests The authors declare no conflicts of interest.