English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Paper

Detecting and Mitigating Test-time Failure Risks via Model-agnostic Uncertainty Learning

MPS-Authors
/persons/resource/persons144524

Gummadi,  Krishna
Group K. Gummadi, Max Planck Institute for Software Systems, Max Planck Society;

External Resource
No external resources are shared
Fulltext (public)

arXiv:2109.04432.pdf
(Preprint), 999KB

Supplementary Material (public)
There is no public supplementary material available
Citation

Lahoti, P., Gummadi, K., & Weikum, G. (2021). Detecting and Mitigating Test-time Failure Risks via Model-agnostic Uncertainty Learning. Retrieved from https://arxiv.org/abs/2109.04432.


Cite as: http://hdl.handle.net/21.11116/0000-0009-6497-C
Abstract
Reliably predicting potential failure risks of machine learning (ML) systems when deployed with production data is a crucial aspect of trustworthy AI. This paper introduces Risk Advisor, a novel post-hoc meta-learner for estimating failure risks and predictive uncertainties of any already-trained black-box classification model. In addition to providing a risk score, the Risk Advisor decomposes the uncertainty estimates into aleatoric and epistemic uncertainty components, thus giving informative insights into the sources of uncertainty inducing the failures. Consequently, Risk Advisor can distinguish between failures caused by data variability, data shifts and model limitations and advise on mitigation actions (e.g., collecting more data to counter data shift). Extensive experiments on various families of black-box classification models and on real-world and synthetic datasets covering common ML failure scenarios show that the Risk Advisor reliably predicts deployment-time failure risks in all the scenarios, and outperforms strong baselines.