Towards an enhanced understanding of bias in pre-trained neural language models: a survey with special emphasis on affective bias
Towards an enhanced understanding of bias in pre-trained neural language models: a survey with special emphasis on affective bias
The remarkable progress in Natural Language Processing (NLP) brought about by deep learning, particularly with the recent advent of large pre-trained neural language models, is brought into scrutiny as several studies began to discuss and report potential biases in NLP applications. Bias in NLP is found to originate from latent historical biases encoded by humans into textual data which gets perpetuated or even amplified by NLP algorithm. We present a survey to comprehend bias in large pre-trained language models and analyze the stages at which they occur in these models, and various ways in which these biases could be quantified and mitigated. Considering wide applicability of textual affective computing-based downstream tasks in real-world systems such as business, health care, and education, we give a special emphasis on investigating bias in the context of affect (emotion) i.e., Affective Bias, in large pre-trained language models. We present a summary of various bias evaluation corpora that help to aid f
NLP Bias, Fairness, Large Pretrained Language Models, Affective Bias, Affective Computing
13-45
Anoop, K.
9cc17e26-a329-49fe-b73b-2fce75084966
P. Gangan, Manjary
f1f79b4a-2662-4f0c-ad33-dbb0cbf2512b
Deepak, P.
80ebb63c-91a6-4500-8e03-9d806262049d
Lajish, V.L.
034cc3e6-c98a-4e9c-ab30-4729948b55c2
15 November 2022
Anoop, K.
9cc17e26-a329-49fe-b73b-2fce75084966
P. Gangan, Manjary
f1f79b4a-2662-4f0c-ad33-dbb0cbf2512b
Deepak, P.
80ebb63c-91a6-4500-8e03-9d806262049d
Lajish, V.L.
034cc3e6-c98a-4e9c-ab30-4729948b55c2
Anoop, K., P. Gangan, Manjary, Deepak, P. and Lajish, V.L.
(2022)
Towards an enhanced understanding of bias in pre-trained neural language models: a survey with special emphasis on affective bias.
Mathew, J., Kumar, G.S., Deepak, P. and Jose, J.M.
(eds.)
In Responsible Data Science: Select Proceedings of ICDSE 2021.
vol. 940,
Springer Singapore.
.
(doi:10.1007/978-981-19-4453-6_2).
Record type:
Conference or Workshop Item
(Paper)
Abstract
The remarkable progress in Natural Language Processing (NLP) brought about by deep learning, particularly with the recent advent of large pre-trained neural language models, is brought into scrutiny as several studies began to discuss and report potential biases in NLP applications. Bias in NLP is found to originate from latent historical biases encoded by humans into textual data which gets perpetuated or even amplified by NLP algorithm. We present a survey to comprehend bias in large pre-trained language models and analyze the stages at which they occur in these models, and various ways in which these biases could be quantified and mitigated. Considering wide applicability of textual affective computing-based downstream tasks in real-world systems such as business, health care, and education, we give a special emphasis on investigating bias in the context of affect (emotion) i.e., Affective Bias, in large pre-trained language models. We present a summary of various bias evaluation corpora that help to aid f
This record has no associated files available for download.
More information
Published date: 15 November 2022
Venue - Dates:
International Conference on Data Science and Engineering, Indian Institute of Technology Patna, IIT Patna, India, 2021-12-17 - 2021-12-18
Keywords:
NLP Bias, Fairness, Large Pretrained Language Models, Affective Bias, Affective Computing
Identifiers
Local EPrints ID: 495963
URI: http://eprints.soton.ac.uk/id/eprint/495963
PURE UUID: 83a84f37-352c-400d-9703-df5b99ed8b13
Catalogue record
Date deposited: 28 Nov 2024 17:34
Last modified: 30 Nov 2024 03:16
Export record
Altmetrics
Contributors
Author:
K. Anoop
Author:
Manjary P. Gangan
Author:
P. Deepak
Author:
V.L. Lajish
Editor:
J. Mathew
Editor:
G.S. Kumar
Editor:
P. Deepak
Editor:
J.M. Jose
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics