Investigating bias in multilingual language models: cross-lingual transfer of debiasing techniques
Investigating bias in multilingual language models: cross-lingual transfer of debiasing techniques
This paper investigates the transferability of debiasing techniques across different languages within multilingual models. We examine the applicability of these techniques in English, French, German, and Dutch. Using multilingual BERT (mBERT), we demonstrate that cross-lingual transfer of debiasing techniques is not only feasible but also yields promising results. Surprisingly, our findings reveal no performance disadvantages when applying these techniques to non-English languages. Using translations of the CrowS-Pairs dataset, our analysis identifies SentenceDebias as the best technique across different languages, reducing bias in mBERT by an average of 13%. We also find that debiasing techniques with additional pretraining exhibit enhanced cross-lingual effectiveness for the languages included in the analyses, particularly in lower-resource languages. These novel insights contribute to a deeper understanding of bias mitigation in multilingual language models and provide practical guidance for debiasing techniques in different language contexts.
2887-2896
Association for Computational Linguistics (ACL)
Reusens, Manon
3dc14c4b-793a-41d6-b7bd-64303cda1c42
Borchert, Philipp
a57c31fb-9fdc-4ca5-bbd1-b6c6fc8a8483
Mieskes, Margot
7db00e45-cddc-4b45-9f75-c71e0edbee24
De Weerdt, Jochen
1eaa177f-03d0-47e5-b8b6-4fb419d49e47
Baesens, Bart
f7c6496b-aa7f-4026-8616-ca61d9e216f0
December 2023
Reusens, Manon
3dc14c4b-793a-41d6-b7bd-64303cda1c42
Borchert, Philipp
a57c31fb-9fdc-4ca5-bbd1-b6c6fc8a8483
Mieskes, Margot
7db00e45-cddc-4b45-9f75-c71e0edbee24
De Weerdt, Jochen
1eaa177f-03d0-47e5-b8b6-4fb419d49e47
Baesens, Bart
f7c6496b-aa7f-4026-8616-ca61d9e216f0
Reusens, Manon, Borchert, Philipp, Mieskes, Margot, De Weerdt, Jochen and Baesens, Bart
(2023)
Investigating bias in multilingual language models: cross-lingual transfer of debiasing techniques.
Bouamor, Houda, Pino, Juan and Bali, Kalika
(eds.)
In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing.
Association for Computational Linguistics (ACL).
.
Record type:
Conference or Workshop Item
(Paper)
Abstract
This paper investigates the transferability of debiasing techniques across different languages within multilingual models. We examine the applicability of these techniques in English, French, German, and Dutch. Using multilingual BERT (mBERT), we demonstrate that cross-lingual transfer of debiasing techniques is not only feasible but also yields promising results. Surprisingly, our findings reveal no performance disadvantages when applying these techniques to non-English languages. Using translations of the CrowS-Pairs dataset, our analysis identifies SentenceDebias as the best technique across different languages, reducing bias in mBERT by an average of 13%. We also find that debiasing techniques with additional pretraining exhibit enhanced cross-lingual effectiveness for the languages included in the analyses, particularly in lower-resource languages. These novel insights contribute to a deeper understanding of bias mitigation in multilingual language models and provide practical guidance for debiasing techniques in different language contexts.
Text
2023.emnlp-main.175
- Accepted Manuscript
More information
Published date: December 2023
Venue - Dates:
2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, , Hybrid, Singapore, Singapore, 2023-12-06 - 2023-12-10
Identifiers
Local EPrints ID: 495211
URI: http://eprints.soton.ac.uk/id/eprint/495211
PURE UUID: f2d247c0-bbc1-44b5-b9f2-2a148bfed3e1
Catalogue record
Date deposited: 01 Nov 2024 17:51
Last modified: 02 Nov 2024 02:40
Export record
Contributors
Author:
Manon Reusens
Author:
Philipp Borchert
Author:
Margot Mieskes
Author:
Jochen De Weerdt
Editor:
Houda Bouamor
Editor:
Juan Pino
Editor:
Kalika Bali
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics