Fairness implications of encoding protected categorical attributes
Fairness implications of encoding protected categorical attributes
Past research has demonstrated that the explicit use of protected attributes in machine learning can improve both performance and fairness. Many machine learning algorithms, however, cannot directly process categorical attributes, such as country of birth or ethnicity. Because protected attributes frequently are categorical, they must be encoded as features that can be input to a chosen machine learning algorithm, e.g. support vector machines, gradient boosting decision trees or linear models. Thereby, encoding methods influence how and what the machine learning algorithm will learn, affecting model performance and fairness. This work compares the accuracy and fairness implications of the two most well-known encoding methods:one-hot encoding and target encoding. We distinguish between two types of induced bias that may arise from these encoding methods and may lead to unfair models. The first type, irreducible bias, is due to direct group category discrimination and the second type, reducible bias, is due to the large variance in statistically underrepresented groups. We investigate the interaction between categorical encodings and target encoding regularization methods that reduce unfairness. Furthermore, we consider the problem of intersectional unfairness that may arise when machine learning best practices improve performance measures by encoding several categorical attributes into a high-cardinality feature.
Mougan, Carlos
229c7631-f1da-4896-a06a-fd27e77e5742
Alvarez, Jose M.
19026e60-f334-482f-924c-fa7e4e9a8455
Ruggieri, Salvatore
62932453-04f0-4d2e-a7f2-1fcef460cb8d
Staab, Steffen
bf48d51b-bd11-4d58-8e1c-4e6e03b30c49
Mougan, Carlos
229c7631-f1da-4896-a06a-fd27e77e5742
Alvarez, Jose M.
19026e60-f334-482f-924c-fa7e4e9a8455
Ruggieri, Salvatore
62932453-04f0-4d2e-a7f2-1fcef460cb8d
Staab, Steffen
bf48d51b-bd11-4d58-8e1c-4e6e03b30c49
Mougan, Carlos, Alvarez, Jose M., Ruggieri, Salvatore and Staab, Steffen
(2023)
Fairness implications of encoding protected categorical attributes.
AAAI/ACM Conference on AI, Ethics, and Society, , Montréal, Canada.
08 - 10 Aug 2023.
12 pp
.
(In Press)
Record type:
Conference or Workshop Item
(Paper)
Abstract
Past research has demonstrated that the explicit use of protected attributes in machine learning can improve both performance and fairness. Many machine learning algorithms, however, cannot directly process categorical attributes, such as country of birth or ethnicity. Because protected attributes frequently are categorical, they must be encoded as features that can be input to a chosen machine learning algorithm, e.g. support vector machines, gradient boosting decision trees or linear models. Thereby, encoding methods influence how and what the machine learning algorithm will learn, affecting model performance and fairness. This work compares the accuracy and fairness implications of the two most well-known encoding methods:one-hot encoding and target encoding. We distinguish between two types of induced bias that may arise from these encoding methods and may lead to unfair models. The first type, irreducible bias, is due to direct group category discrimination and the second type, reducible bias, is due to the large variance in statistically underrepresented groups. We investigate the interaction between categorical encodings and target encoding regularization methods that reduce unfairness. Furthermore, we consider the problem of intersectional unfairness that may arise when machine learning best practices improve performance measures by encoding several categorical attributes into a high-cardinality feature.
Text
FairEncodingCRC
- Accepted Manuscript
More information
Accepted/In Press date: 21 June 2023
Venue - Dates:
AAAI/ACM Conference on AI, Ethics, and Society, , Montréal, Canada, 2023-08-08 - 2023-08-10
Identifiers
Local EPrints ID: 478460
URI: http://eprints.soton.ac.uk/id/eprint/478460
PURE UUID: 60e4a48b-88ff-448b-9b37-287c60247371
Catalogue record
Date deposited: 03 Jul 2023 16:52
Last modified: 17 Mar 2024 03:38
Export record
Contributors
Author:
Carlos Mougan
Author:
Jose M. Alvarez
Author:
Salvatore Ruggieri
Author:
Steffen Staab
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics