Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

S-LIME: Stabilized-LIME for Model Explanation

About

An increasing number of machine learning models have been deployed in domains with high stakes such as finance and healthcare. Despite their superior performances, many models are black boxes in nature which are hard to explain. There are growing efforts for researchers to develop methods to interpret these black-box models. Post hoc explanations based on perturbations, such as LIME, are widely used approaches to interpret a machine learning model after it has been built. This class of methods has been shown to exhibit large instability, posing serious challenges to the effectiveness of the method itself and harming user trust. In this paper, we propose S-LIME, which utilizes a hypothesis testing framework based on central limit theorem for determining the number of perturbation points needed to guarantee stability of the resulting explanation. Experiments on both simulated and real world data sets are provided to demonstrate the effectiveness of our method.

Zhengze Zhou, Giles Hooker, Fei Wang• 2021

Related benchmarks

TaskDatasetResultRank
Local Explanation GenerationCovType
Stability99.9
14
Explanation RegularityDigits
Regularity73.1
11
Explanation Fidelity EstimationCA Housing
R2 Score (Fidelity)0.502
11
Explanation RegularityWine
Regularity89.2
11
Explanation RegularityDiabetes
Regularity95.8
11
Explanation StabilityWine
Stability99.9
11
Explanation StabilityAmes Housing
Stability0.998
11
Explanation Fidelity EstimationCovType
Fidelity (R2)0.468
11
Explanation StabilityBreast cancer
Stability99.5
11
Explanation StabilityDigits
Stability98.5
11
Showing 10 of 44 rows

Other info

Follow for update