Uncertainty-aware Sensitivity Analysis Using Rényi Divergences

Loading...
Thumbnail Image

Access rights

openAccess
publishedVersion

URL

Journal Title

Journal ISSN

Volume Title

A4 Artikkeli konferenssijulkaisussa

Major/Subject

Mcode

Degree programme

Language

en

Pages

Series

Proceedings of the Thirty-Seventh Conference on Uncertainty in Artificial Intelligence, pp. 1185-1194, Proceedings of Machine Learning Research ; Volume 161

Abstract

For nonlinear supervised learning models, assessing the importance of predictor variables or their interactions is not straightforward because importance can vary in the domain of the variables. Importance can be assessed locally with sensitivity analysis using general methods that rely on the model's predictions or their derivatives. In this work, we extend derivative based sensitivity analysis to a Bayesian setting by differentiating the Rényi divergence of a model's predictive distribution. By utilising the predictive distribution instead of a point prediction, the model uncertainty is taken into account in a principled way. Our empirical results on simulated and real data sets demonstrate accurate and reliable identification of important variables and interaction effects compared to alternative methods.

Description

Keywords

Other note

Citation

Paananen, T, Andersen, M & Vehtari, A 2021, Uncertainty-aware Sensitivity Analysis Using Rényi Divergences. in Proceedings of the Thirty-Seventh Conference on Uncertainty in Artificial Intelligence. Proceedings of Machine Learning Research, vol. 161, JMLR, pp. 1185-1194, Conference on Uncertainty in Artificial Intelligence, Virtual, Online, 27/07/2021. < https://proceedings.mlr.press/v161/paananen21a.html >