Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Explanations Go Linear: Post-hoc Explainability for Tabular Data with Interpretable Meta-Encoding

About

Post-hoc explainability is essential for understanding black-box machine learning models. Surrogate-based techniques are widely used for local and global model-agnostic explanations but have significant limitations. Local surrogates capture non-linearities but are computationally expensive and sensitive to parameters, while global surrogates are more efficient but struggle with complex local behaviors. In this paper, we present ILLUME, a flexible and interpretable framework grounded in representation learning, that can be integrated with various surrogate models to provide explanations for any black-box classifier. Specifically, our approach combines a globally trained surrogate with instance-specific linear transformations learned with a meta-encoder to generate both local and global explanations. Through extensive empirical evaluations, we demonstrate the effectiveness of ILLUME in producing feature attributions and decision rules that are not only accurate but also robust and computationally efficient, thus providing a unified explanation framework that effectively addresses the limitations of traditional surrogate methods.

Simone Piaggesi, Riccardo Guidotti, Fosca Giannotti, Dino Pedreschi• 2025

Related benchmarks

TaskDatasetResultRank
Feature Importance Correctnessint2-3c
Prediction Accuracy97.2
30
Feature Importance Correctnessssin-2c
CS-Score0.998
21
Feature Importance Correctnessint2-8p
CS-Score0.305
21
Feature Importance Rankingssin-2c
Spearman Correlation1
21
Feature Importance Rankingint2-8p
Spearman Rank Correlation0.458
21
Explanation GenerationSENECA-RC synthetic
Per-instance Explanation Time (s)4.4
15
Feature Importance ExplanationReal-world datasets
Best Score95.9
12
Showing 7 of 7 rows

Other info

Follow for update