Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
TransLIME: Towards transfer explainability to explain black-box models on tabular datasets
2
Zitationen
5
Autoren
2025
Jahr
Abstract
Explainable Artificial Intelligence methods have gained significant traction for their ability to elucidate the decision-making processes of black-box models, particularly in high-stakes fields such as healthcare and finance. Among these, Local Interpretable Model-agnostic Explanations (LIME) stands out as a widely adopted post-hoc, model-agnostic approach that interprets black-box predictions by constructing an interpretable surrogate model on perturbed instances to approximate the local behavior of the original model around a given instance. However, the effectiveness of LIME can depend on the quality of the training data used by the black-box model. When trained on limited or low-quality data, the black-box model may yield inaccurate predictions for perturbed samples, resulting in poorly defined local decision boundaries and consequently unreliable explanations. This limitation is especially problematic in data-scarce settings. To overcome this challenge, we propose TransLIME, a novel end-to-end explainable transfer learning framework that improves the local fidelity and stability of LIME on limited tabular datasets by transferring relevant explainability knowledge from a related auxiliary source domain with a shifted distribution. Also, in TransLIME, only representative source prototype explanations obtained through clustering are transferred to the target domain, thereby reducing cross-domain exposure of both data and explanatory information during transfer. Experimental evaluations on real-world datasets demonstrate the effectiveness of the proposed framework in improving explanation quality in target domains with limited data.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.657 Zit.
Generative Adversarial Nets
2023 · 19.894 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.315 Zit.
"Why Should I Trust You?"
2016 · 14.512 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.188 Zit.