Density-based reliable and robust explainer for counterfactual explanation.

Expert Syst. Appl.(2023)

引用 0|浏览0
暂无评分
摘要
As an essential post-hoc explanatory method, counterfactual explanation enables people to understand and react to machine learning models. Works on counterfactual explanation generally aim at generating high -quality results, which means providing close and detailed explanations to users. However, the counterfactual explainer trained on data is fragile in practice, i.e., even a small perturbation to samples can lead to large differences in explanation. In this work, we address this issue by analyzing and formalizing the robustness of counterfactual explainer with practical considerations. An explainer is considered robust if it can generate relatively stable counterfactuals under various settings. To this end, we propose a robust and reliable explainer for searching counterfactuals of classifier predictions by using density gravity. To evaluate the performance, we provide metrics that allow comparison of our proposed explainer with others and further demonstrate the importance of density in enhancing robustness. Extensive experiments on real-world datasets show that our method offers a significant improvement in explainer reliability and stability.
更多
查看译文
关键词
Counterfactual explanation,Robustness,Density,Interpretability
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要