2019-03-14T08:17:59Z
2019-03-14T08:17:59Z
2018-11-23
In this paper we propose a method to obtain global explanations for trained black-box classifiers by sampling their decision function to learn alternative interpretable models. The envisaged approach provides a unified solution to approximate non-linear decision boundaries with simpler classifiers while retaining the original classification accuracy. We use a private residential mortgage default dataset as a use case to illustrate the feasibility of this approach to ensure the decomposability of attributes during pre-processing.
Objeto de conferencia
Inglés
Neural Information Processing Systems Foundation
Comunicació a: NIPS 2018 Workshop on Challenges and Opportunities for AI in Financial Services: the Impact of Fairness, Explainability, Accuracy, and Privacy, Montréal, Canada. December 7th, 2018
(c) Unceta et al., 2018