EXPRESS: Prior multisensory learning can facilitate auditory-only voice-identity and speech recognition in noise.

audio-visual multisensory learning person recognition speech in noise voice identity

Journal

Quarterly journal of experimental psychology (2006)
ISSN: 1747-0226
Titre abrégé: Q J Exp Psychol (Hove)
Pays: England
ID NLM: 101259775

Informations de publication

Date de publication:
20 Aug 2024
Historique:
medline: 21 8 2024
pubmed: 21 8 2024
entrez: 21 8 2024
Statut: aheadofprint

Résumé

Seeing the visual articulatory movements of a speaker, while hearing their voice, helps with understanding what is said. This multisensory enhancement is particularly evident in noisy listening conditions. Multisensory enhancement also occurs even in auditory-only conditions: auditory-only speech and voice-identity recognition is superior for speakers previously learned with their face, compared to control learning; an effect termed the "face-benefit". Whether the face-benefit can assist in maintaining robust perception in increasingly noisy listening conditions, similar to concurrent multisensory input, is unknown. Here, in two behavioural experiments, we examined this hypothesis. In each experiment, participants learned a series of speakers' voices together with their dynamic face, or control image. Following learning, participants listened to auditory-only sentences spoken by the same speakers and recognised the content of the sentences (speech recognition, Experiment 1) or the voice-identity of the speaker (Experiment 2) in increasing levels of auditory noise. For speech recognition, we observed that 14/30 participants (47%) showed a face-benefit. While 19/25 participants (76%) showed a face-benefit for voice-identity recognition. For those participants who demonstrated a face-benefit, the face-benefit increased with auditory noise levels. Taken together, the results support an audio-visual model of auditory communication and suggest that the brain can develop a flexible system in which learned facial characteristics are used to deal with varying auditory uncertainty.

Identifiants

pubmed: 39164830
doi: 10.1177/17470218241278649
doi:

Types de publication

Journal Article

Langues

eng

Sous-ensembles de citation

IM

Pagination

17470218241278649

Auteurs

Corrina Maguinness (C)

Chair of Cognitive and Clinical Neuroscience, Technische Universität Dresden, Dresden, Germany.
Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.

Sonja Schall (S)

Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.

Brian Mathias (B)

Chair of Cognitive and Clinical Neuroscience, Technische Universität Dresden, Dresden, Germany.
School of Psychology, University of Aberdeen, Aberdeen, United Kingdom.

Martin Schoemann (M)

Chair of Psychological Methods and Cognitive Modelling, Technische Universität Dresden, Dresden, Germany.

Katharina von Kriegstein (K)

Chair of Cognitive and Clinical Neuroscience, Technische Universität Dresden, Dresden, Germany.
Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.

Classifications MeSH