The integration of continuous audio and visual speech in a cocktail-party environment depends on attention.
Cocktail party
Hierarchical processing
Multisensory integration
Speech
Journal
NeuroImage
ISSN: 1095-9572
Titre abrégé: Neuroimage
Pays: United States
ID NLM: 9215515
Informations de publication
Date de publication:
01 07 2023
01 07 2023
Historique:
received:
17
08
2022
revised:
17
03
2023
accepted:
27
04
2023
medline:
17
5
2023
pubmed:
1
5
2023
entrez:
30
4
2023
Statut:
ppublish
Résumé
In noisy environments, our ability to understand speech benefits greatly from seeing the speaker's face. This is attributed to the brain's ability to integrate audio and visual information, a process known as multisensory integration. In addition, selective attention plays an enormous role in what we understand, the so-called cocktail-party phenomenon. But how attention and multisensory integration interact remains incompletely understood, particularly in the case of natural, continuous speech. Here, we addressed this issue by analyzing EEG data recorded from participants who undertook a multisensory cocktail-party task using natural speech. To assess multisensory integration, we modeled the EEG responses to the speech in two ways. The first assumed that audiovisual speech processing is simply a linear combination of audio speech processing and visual speech processing (i.e., an A + V model), while the second allows for the possibility of audiovisual interactions (i.e., an AV model). Applying these models to the data revealed that EEG responses to attended audiovisual speech were better explained by an AV model, providing evidence for multisensory integration. In contrast, unattended audiovisual speech responses were best captured using an A + V model, suggesting that multisensory integration is suppressed for unattended speech. Follow up analyses revealed some limited evidence for early multisensory integration of unattended AV speech, with no integration occurring at later levels of processing. We take these findings as evidence that the integration of natural audio and visual speech occurs at multiple levels of processing in the brain, each of which can be differentially affected by attention.
Identifiants
pubmed: 37121375
pii: S1053-8119(23)00294-X
doi: 10.1016/j.neuroimage.2023.120143
pii:
doi:
Types de publication
Journal Article
Research Support, N.I.H., Extramural
Research Support, Non-U.S. Gov't
Langues
eng
Sous-ensembles de citation
IM
Pagination
120143Informations de copyright
Copyright © 2023. Published by Elsevier Inc.
Déclaration de conflit d'intérêts
Declaration of Competing Interest The authors declare that they have no competing interests, financial or otherwise.