TY - GEN
T1 - Can Post-hoc Explanations Effectively Detect Out-of-Distribution Samples?
AU - Martinez-Seras, Aitor
AU - Ser, Javier Del
AU - Garcia-Bringas, Pablo
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Today there is consensus around the importance of explainability as a mandatory feature in practical deployments of Artificial Intelligence (AI) models. Most research activity reported so far in the eXplainable AI (XAI) research arena has stressed on proposing new techniques for eliciting such explanations, together with different approaches for measuring their effectivity to increase the trustworthiness of the audience for which explanations are furnished. However, alternative uses of explanations beyond their original purpose have been very scarcely explored. In this work we investigate whether local explanations can be utilized for detecting Out-of-Distribution (OoD) test samples in machine learning classifiers, i.e., to identify whether query examples of an already trained classification model can be thought to belong to the distribution of the training data. To this end, we devise and assess the performance of a clustering-based OoD detection approach that exemplifies how heatmaps produced by well-established local explanation methods can be of further use than explaining individual predictions issued by the model under analysis. The overarching purpose of this work is not only to expose the benefits and the limits of the proposed XAI-based OoD approach, but also to point out the enormous potential of post-hoc explanations beyond easing the interpretability of black-box models themselves.
AB - Today there is consensus around the importance of explainability as a mandatory feature in practical deployments of Artificial Intelligence (AI) models. Most research activity reported so far in the eXplainable AI (XAI) research arena has stressed on proposing new techniques for eliciting such explanations, together with different approaches for measuring their effectivity to increase the trustworthiness of the audience for which explanations are furnished. However, alternative uses of explanations beyond their original purpose have been very scarcely explored. In this work we investigate whether local explanations can be utilized for detecting Out-of-Distribution (OoD) test samples in machine learning classifiers, i.e., to identify whether query examples of an already trained classification model can be thought to belong to the distribution of the training data. To this end, we devise and assess the performance of a clustering-based OoD detection approach that exemplifies how heatmaps produced by well-established local explanation methods can be of further use than explaining individual predictions issued by the model under analysis. The overarching purpose of this work is not only to expose the benefits and the limits of the proposed XAI-based OoD approach, but also to point out the enormous potential of post-hoc explanations beyond easing the interpretability of black-box models themselves.
KW - Explainable Artificial Intelligence
KW - local explanations
KW - Out-of-Distribution (OoD) detection
UR - http://www.scopus.com/inward/record.url?scp=85138812137&partnerID=8YFLogxK
U2 - 10.1109/FUZZ-IEEE55066.2022.9882726
DO - 10.1109/FUZZ-IEEE55066.2022.9882726
M3 - Conference contribution
AN - SCOPUS:85138812137
T3 - IEEE International Conference on Fuzzy Systems
BT - 2022 IEEE International Conference on Fuzzy Systems, FUZZ 2022 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2022 IEEE International Conference on Fuzzy Systems, FUZZ 2022
Y2 - 18 July 2022 through 23 July 2022
ER -