Practical, epistemic and normative implications of algorithmic bias in healthcare artificial intelligence: a qualitative study of multidisciplinary expert perspectives.

Decision Making Ethics Information Technology Policy

Journal

Journal of medical ethics
ISSN: 1473-4257
Titre abrégé: J Med Ethics
Pays: England
ID NLM: 7513619

Informations de publication

Date de publication:
23 Feb 2023
Historique:
received: 15 12 2022
accepted: 16 02 2023
entrez: 23 2 2023
pubmed: 24 2 2023
medline: 24 2 2023
Statut: aheadofprint

Résumé

There is a growing concern about artificial intelligence (AI) applications in healthcare that can disadvantage already under-represented and marginalised groups (eg, based on gender or race). Our objectives are to canvas the range of strategies stakeholders endorse in attempting to mitigate algorithmic bias, and to consider the ethical question of responsibility for algorithmic bias. The study involves in-depth, semistructured interviews with healthcare workers, screening programme managers, consumer health representatives, regulators, data scientists and developers. Findings reveal considerable divergent views on three key issues. First, views on whether bias is a problem in healthcare AI varied, with most participants agreeing bias is a problem (which we call the bias-critical view), a small number believing the opposite (the bias-denial view), and some arguing that the benefits of AI outweigh any harms or wrongs arising from the bias problem (the bias-apologist view). Second, there was a disagreement on the strategies to mitigate bias, and who is responsible for such strategies. Finally, there were divergent views on whether to include or exclude sociocultural identifiers (eg, race, ethnicity or gender-diverse identities) in the development of AI as a way to mitigate bias. Based on the views of participants, we set out responses that stakeholders might pursue, including greater interdisciplinary collaboration, tailored stakeholder engagement activities, empirical studies to understand algorithmic bias and strategies to modify dominant approaches in AI development such as the use of participatory methods, and increased diversity and inclusion in research teams and research participant recruitment and selection.

Sections du résumé

BACKGROUND BACKGROUND
There is a growing concern about artificial intelligence (AI) applications in healthcare that can disadvantage already under-represented and marginalised groups (eg, based on gender or race).
OBJECTIVES OBJECTIVE
Our objectives are to canvas the range of strategies stakeholders endorse in attempting to mitigate algorithmic bias, and to consider the ethical question of responsibility for algorithmic bias.
METHODOLOGY METHODS
The study involves in-depth, semistructured interviews with healthcare workers, screening programme managers, consumer health representatives, regulators, data scientists and developers.
RESULTS RESULTS
Findings reveal considerable divergent views on three key issues. First, views on whether bias is a problem in healthcare AI varied, with most participants agreeing bias is a problem (which we call the bias-critical view), a small number believing the opposite (the bias-denial view), and some arguing that the benefits of AI outweigh any harms or wrongs arising from the bias problem (the bias-apologist view). Second, there was a disagreement on the strategies to mitigate bias, and who is responsible for such strategies. Finally, there were divergent views on whether to include or exclude sociocultural identifiers (eg, race, ethnicity or gender-diverse identities) in the development of AI as a way to mitigate bias.
CONCLUSION/SIGNIFICANCE CONCLUSIONS
Based on the views of participants, we set out responses that stakeholders might pursue, including greater interdisciplinary collaboration, tailored stakeholder engagement activities, empirical studies to understand algorithmic bias and strategies to modify dominant approaches in AI development such as the use of participatory methods, and increased diversity and inclusion in research teams and research participant recruitment and selection.

Identifiants

pubmed: 36823101
pii: jme-2022-108850
doi: 10.1136/jme-2022-108850
pii:
doi:

Types de publication

Journal Article

Langues

eng

Sous-ensembles de citation

IM

Informations de copyright

© Author(s) (or their employer(s)) 2023. Re-use permitted under CC BY. Published by BMJ.

Déclaration de conflit d'intérêts

Competing interests: NH and SC received a grant from the Australian National Breast Cancer Foundation, a non-government organisation, to validate existing AI models on data from Australian breast cancer screening programs. NH has active research collaborations without financial arrangements with Therapixel, a developer of healthcare AI. The rest of the authors have no conflicts of interest.

Auteurs

Yves Saint James Aquino (YSJ)

Australian Centre for Health Engagement, Evidence and Values, School of Health and Society, University of Wollongong, Wollongong, New South Wales, Australia yaquino@uow.edu.au.

Stacy M Carter (SM)

Australian Centre for Health Engagement, Evidence and Values, School of Health and Society, University of Wollongong, Wollongong, New South Wales, Australia.

Nehmat Houssami (N)

School of Public Health, The University of Sydney, Sydney, New South Wales, Australia.
The Daffodil Centre, Sydney, New South Wales, Australia.

Annette Braunack-Mayer (A)

Australian Centre for Health Engagement, Evidence and Values, School of Health and Society, University of Wollongong, Wollongong, New South Wales, Australia.

Khin Than Win (KT)

Centre for Persuasive Technology and Society, Faculty of Engineering and Information Sciences, University of Wollongong, Wollongong, New South Wales, Australia.

Chris Degeling (C)

Australian Centre for Health Engagement, Evidence and Values, School of Health and Society, University of Wollongong, Wollongong, New South Wales, Australia.

Lei Wang (L)

Centre for Artificial Intelligence, School of Computing and Information Technology, University of Wollongong, Wollongong, New South Wales, Australia.

Wendy A Rogers (WA)

Department of Philosophy and School of Medicine, Macquarie University, Sydney, New South Wales, Australia.

Classifications MeSH