Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

AEM: Attention Entropy Maximization for Multiple Instance Learning based Whole Slide Image Classification

About

Multiple Instance Learning (MIL) effectively analyzes whole slide images but faces overfitting due to attention over-concentration. While existing solutions rely on complex architectural modifications or additional processing steps, we introduce Attention Entropy Maximization (AEM), a simple yet effective regularization technique. Our investigation reveals the positive correlation between attention entropy and model performance. Building on this insight, we integrate AEM regularization into the MIL framework to penalize excessive attention concentration. To address sensitivity to the AEM weight parameter, we implement Cosine Weight Annealing, reducing parameter dependency. Extensive evaluations demonstrate AEM's superior performance across diverse feature extractors, MIL frameworks, attention mechanisms, and augmentation techniques. Here is our anonymous code: https://github.com/dazhangyu123/AEM.

Yunlong Zhang, Honglin Li, Yunxuan Sun, Zhongyi Shui, Jingxiong Li, Chenglu Zhu, Lin Yang• 2024

Related benchmarks

TaskDatasetResultRank
Whole Slide Image classificationCAMELYON16 (test)
AUC0.998
163
Tumor localizationCAMELYON16 (test)
AUC96.7
65
WSI ClassificationCAMELYON17 (test)
AUC90.5
33
Whole Slide Image classificationLBC (test)
F1 Score0.691
24
WSI subtypingCAMELYON 17
F1 Score64.7
24
WSI subtypingBRACS
F1 Score74.2
24
WSI subtypingCAMELYON-16
F1 Score94.7
24
WSI subtypingCAMELYON-16 v1 (test)
F1 Score97.5
11
WSI subtypingCAMELYON-17 v1 (test)
F1 Score68.8
11
Showing 9 of 9 rows

Other info

Code

Follow for update