Explainability of artificial intelligence methods has become a crucial issue, especially in the most regulated fields, such as health and finance. In this paper, we provide a global explainable AI method which is based on Lorenz decompositions, thus extending previous contributions based on variance decompositions. This allows the resulting Shapley-Lorenz decomposition to be more generally applicable, and provides a unifying variable importance criterion that combines predictive accuracy with explainability, using a normalised and easy to interpret metric. The proposed decomposition is illustrated within the context of a real financial problem: the prediction of bitcoin prices.
Shapley-Lorenz eXplainable Artificial Intelligence / P. Giudici, E. Raffinetti. - In: EXPERT SYSTEMS WITH APPLICATIONS. - ISSN 0957-4174. - (2020). [Epub ahead of print]
Shapley-Lorenz eXplainable Artificial Intelligence
E. RaffinettiSecondo
2020
Abstract
Explainability of artificial intelligence methods has become a crucial issue, especially in the most regulated fields, such as health and finance. In this paper, we provide a global explainable AI method which is based on Lorenz decompositions, thus extending previous contributions based on variance decompositions. This allows the resulting Shapley-Lorenz decomposition to be more generally applicable, and provides a unifying variable importance criterion that combines predictive accuracy with explainability, using a normalised and easy to interpret metric. The proposed decomposition is illustrated within the context of a real financial problem: the prediction of bitcoin prices.File | Dimensione | Formato | |
---|---|---|---|
EXPERT_SYSTEMS_WITH_APPLICATIONS_2020.pdf
accesso aperto
Tipologia:
Publisher's version/PDF
Dimensione
436.26 kB
Formato
Adobe PDF
|
436.26 kB | Adobe PDF | Visualizza/Apri |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.