Monitoring model deterioration with explainable uncertainty estimation via non-parametric bootstrap
Monitoring model deterioration with explainable uncertainty estimation via non-parametric bootstrap
Monitoring machine learning models once they are deployed is challenging. It is even more challenging to decide when to retrain models in real-case scenarios when labeled data is beyond reach, and monitoring performance metrics becomes unfeasible. In this work, we use non-parametric bootstrapped uncertainty estimates and SHAP values to provide explainable uncertainty estimation as a technique that aims to monitor the deterioration of machine learning models in deployment environments, as well as determine the source of model deterioration when target labels are not available. Classical methods are purely aimed at detecting distribution shift, which can lead to false positives in the sense that the model has not deteriorated despite a shift in the data distribution. To estimate model uncertainty we construct prediction intervals using a novel bootstrap method, which improves upon the work of Kumar and Srivastava (2012). We show that both our model deterioration detection system as well as our uncertainty estimation method achieve better performance than the current state-of-the-art. Finally, we use explainable AI techniques to gain an understanding of the drivers of model deterioration. We release an open source Python package, doubt, which implements our proposed methods, as well as the code used to reproduce our experiments.
Mougan, Carlos
229c7631-f1da-4896-a06a-fd27e77e5742
Saattrup Nielsen, Dan
45b4b764-1a27-435d-b8f7-ca2e7ec15ad9
7 February 2023
Mougan, Carlos
229c7631-f1da-4896-a06a-fd27e77e5742
Saattrup Nielsen, Dan
45b4b764-1a27-435d-b8f7-ca2e7ec15ad9
Mougan, Carlos and Saattrup Nielsen, Dan
(2023)
Monitoring model deterioration with explainable uncertainty estimation via non-parametric bootstrap.
In Proceedings of the AAAI Conference on Artificial Intelligence 3.
9 pp
.
Record type:
Conference or Workshop Item
(Paper)
Abstract
Monitoring machine learning models once they are deployed is challenging. It is even more challenging to decide when to retrain models in real-case scenarios when labeled data is beyond reach, and monitoring performance metrics becomes unfeasible. In this work, we use non-parametric bootstrapped uncertainty estimates and SHAP values to provide explainable uncertainty estimation as a technique that aims to monitor the deterioration of machine learning models in deployment environments, as well as determine the source of model deterioration when target labels are not available. Classical methods are purely aimed at detecting distribution shift, which can lead to false positives in the sense that the model has not deteriorated despite a shift in the data distribution. To estimate model uncertainty we construct prediction intervals using a novel bootstrap method, which improves upon the work of Kumar and Srivastava (2012). We show that both our model deterioration detection system as well as our uncertainty estimation method achieve better performance than the current state-of-the-art. Finally, we use explainable AI techniques to gain an understanding of the drivers of model deterioration. We release an open source Python package, doubt, which implements our proposed methods, as well as the code used to reproduce our experiments.
Text
26755-Article Text-30818-1-2-20230626
Restricted to Repository staff only
Request a copy
More information
Published date: 7 February 2023
Identifiers
Local EPrints ID: 479952
URI: http://eprints.soton.ac.uk/id/eprint/479952
PURE UUID: c69103ec-1577-45bd-aaab-2c8ae1dc52de
Catalogue record
Date deposited: 31 Jul 2023 16:41
Last modified: 17 Mar 2024 03:33
Export record
Contributors
Author:
Carlos Mougan
Author:
Dan Saattrup Nielsen
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics