Artificialising whiteness? How AI normalises whiteness in theory, policy and practice
Artificialising whiteness? How AI normalises whiteness in theory, policy and practice
This chapter analyses artificial intelligence (AI) from perspectives of critical race and whiteness theory, socio-technical studies and governmentality. In her chapter ‘Artificializing whiteness? How AI normalises whiteness in theory, policy and practice’, Leonard shows how contemporary operationalisations of AI, such as automated decision making (ADM) processes, can be understood as a tool securing the privileges and power of whiteness, despite being positioned under the guise of neutral technology. Investigating how these racialising processes also intersect with gender and social class, Leonard shows how ADM systems ‘artificialise whiteness’ by consistently and routinely disadvantaging members of marginalised groups . . The chapter demonstrates how minorities, especially Black and ethnic minority women, are at risk of being subjected to unfair and biased (automated) decision making. She concludes that the US-dominated AI industry and many of AI’s design decisions collude in artificialising whiteness. She therefore advocates in favour of recognising AI as a racial, gendered and classist technology of governance which seeks to reassert, renew and normalise different facets of white privilege.
44-58
Leonard, Pauline
a2839090-eccc-4d84-ab63-c6a484c6d7c1
22 June 2023
Leonard, Pauline
a2839090-eccc-4d84-ab63-c6a484c6d7c1
Leonard, Pauline
(2023)
Artificialising whiteness? How AI normalises whiteness in theory, policy and practice.
In,
Andreassen, Rikke, Lundstrom, Catrin, Keskinen, Suvi and Tate, Shirley Anne
(eds.)
The Routledge International Handbook of New Critical Race and Whiteness Studies.
1st ed.
Taylor & Francis, .
Record type:
Book Section
Abstract
This chapter analyses artificial intelligence (AI) from perspectives of critical race and whiteness theory, socio-technical studies and governmentality. In her chapter ‘Artificializing whiteness? How AI normalises whiteness in theory, policy and practice’, Leonard shows how contemporary operationalisations of AI, such as automated decision making (ADM) processes, can be understood as a tool securing the privileges and power of whiteness, despite being positioned under the guise of neutral technology. Investigating how these racialising processes also intersect with gender and social class, Leonard shows how ADM systems ‘artificialise whiteness’ by consistently and routinely disadvantaging members of marginalised groups . . The chapter demonstrates how minorities, especially Black and ethnic minority women, are at risk of being subjected to unfair and biased (automated) decision making. She concludes that the US-dominated AI industry and many of AI’s design decisions collude in artificialising whiteness. She therefore advocates in favour of recognising AI as a racial, gendered and classist technology of governance which seeks to reassert, renew and normalise different facets of white privilege.
This record has no associated files available for download.
More information
Published date: 22 June 2023
Identifiers
Local EPrints ID: 480514
URI: http://eprints.soton.ac.uk/id/eprint/480514
PURE UUID: a2c71ee8-7646-4b24-8605-e44f50a25fb9
Catalogue record
Date deposited: 03 Aug 2023 17:21
Last modified: 18 Mar 2024 02:40
Export record
Contributors
Editor:
Rikke Andreassen
Editor:
Catrin Lundstrom
Editor:
Suvi Keskinen
Editor:
Shirley Anne Tate
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics