Building uncertainty models on top of black-box predictive APIs

dc.contributor.author
Brando Guillaumes, Axel
dc.contributor.author
Torres, Damià
dc.contributor.author
Rodriguez-Serrano, José A.
dc.contributor.author
Vitrià i Marca, Jordi
dc.date.issued
2020-07-14T08:42:01Z
dc.date.issued
2020-07-14T08:42:01Z
dc.date.issued
2020-07-02
dc.date.issued
2020-07-14T08:42:01Z
dc.identifier
2169-3536
dc.identifier
https://hdl.handle.net/2445/168552
dc.identifier
702678
dc.description.abstract
With the commoditization of machine learning, more and more off-the-shelf models are available as part of code libraries or cloud services. Typically, data scientists and other users apply these models as ''black boxes'' within larger projects. In the case of regressing a scalar quantity, such APIs typically offer a predict() function, which outputs the estimated target variable (often referred to as y¿ or, in code, y_hat). However, many real-world problems may require some sort of deviation interval or uncertainty score rather than a single point-wise estimate. In other words, a mechanism is needed with which to answer the question ''How confident is the system about that prediction?'' Motivated by the lack of this characteristic in most predictive APIs designed for regression purposes, we propose a method that adds an uncertainty score to every black-box prediction. Since the underlying model is not accessible, and therefore standard Bayesian approaches are not applicable, we adopt an empirical approach and fit an uncertainty model using a labelled dataset (x, y) and the outputs y¿ of the black box. In order to be able to use any predictive system as a black box and adapt to its complex behaviours, we propose three variants of an uncertainty model based on deep networks. The first adds a heteroscedastic noise component to the black-box output, the second predicts the residuals of the black box, and the third performs quantile regression using deep networks. Experiments using real financial data that contain an in-production black-box system and two public datasets (energy forecasting and biology responses) illustrate and quantify how uncertainty scores can be added to black-box outputs.
dc.format
13 p.
dc.format
application/pdf
dc.format
application/pdf
dc.language
eng
dc.publisher
Institute of Electrical and Electronics Engineers (IEEE)
dc.relation
Reproducció del document publicat a: https://doi.org/10.1109/ACCESS.2020.3006711
dc.relation
IEEE Access, 2020, vol. 8, p. 121344 -121356
dc.relation
https://doi.org/10.1109/ACCESS.2020.3006711
dc.rights
cc-by (c) Brando, Axel et al., 2020
dc.rights
http://creativecommons.org/licenses/by/3.0/es
dc.rights
info:eu-repo/semantics/openAccess
dc.source
Articles publicats en revistes (Matemàtiques i Informàtica)
dc.subject
Aprenentatge automàtic
dc.subject
Xarxes neuronals (Informàtica)
dc.subject
Intel·ligència artificial
dc.subject
Machine learning
dc.subject
Neural networks (Computer science)
dc.subject
Artificial intelligence
dc.title
Building uncertainty models on top of black-box predictive APIs
dc.type
info:eu-repo/semantics/article
dc.type
info:eu-repo/semantics/publishedVersion


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)