No Cover Image

Journal article 21 views 4 downloads

Beyond ideals: why the (medical) AI industry needs to motivate behavioural change in line with fairness and transparency values, and how it can do it

Alice Liefgreen, Netta Weinstein , Sandra Wachter, Brent Mittelstadt

AI and Society

Swansea University Author: Alice Liefgreen

  • 63376.pdf

    PDF | Version of Record

    Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

    Download (856.83KB)

Abstract

Artificial intelligence (AI) is increasingly relied upon by clinicians for making diagnostic and treatment decisions, playing an important role in imaging, diagnosis, risk analysis, lifestyle monitoring, and health information management. While research has identified biases in healthcare AI systems...

Full description

Published in: AI and Society
ISSN: 0951-5666 1435-5655
Published: Springer Nature 2023
Online Access: Check full text

URI: https://cronfa.swan.ac.uk/Record/cronfa63376
Tags: Add Tag
No Tags, Be the first to tag this record!
Abstract: Artificial intelligence (AI) is increasingly relied upon by clinicians for making diagnostic and treatment decisions, playing an important role in imaging, diagnosis, risk analysis, lifestyle monitoring, and health information management. While research has identified biases in healthcare AI systems and proposed technical solutions to address these, we argue that effective solutions require human engagement. Furthermore, there is a lack of research on how to motivate the adoption of these solutions and promote investment in designing AI systems that align with values such as transparency and fairness from the outset. Drawing on insights from psychological theories, we assert the need to understand the values that underlie decisions made by individuals involved in creating and deploying AI systems. We describe how this understanding can be leveraged to increase engagement with de-biasing and fairness-enhancing practices within the AI healthcare industry, ultimately leading to sustained behavioral change via autonomy-supportive communication strategies rooted in motivational and social psychology theories. In developing these pathways to engagement, we consider the norms and needs that govern the AI healthcare domain, and we evaluate incentives for maintaining the status quo against economic, legal, and social incentives for behavior change in line with transparency and fairness values.
College: Faculty of Humanities and Social Sciences
Funders: Wellcome Trust ( 223765/Z/21/Z), Alfred P. Sloan Foundation (G-2021-16779), Department of Health and Social Care, British Academy ( PF2\180114), Luminate Group, Miami Foundation