TICON: A Slide-Level Tile Contextualizer for Histopathology Representation Learning
About
The interpretation of small tiles in large whole slide images (WSI) often needs a larger image context. We introduce TICON, a transformer-based tile representation contextualizer that produces rich, contextualized embeddings for ''any'' application in computational pathology. Standard tile encoder-based pipelines, which extract embeddings of tiles stripped from their context, fail to model the rich slide-level information essential for both local and global tasks. Furthermore, different tile-encoders excel at different downstream tasks. Therefore, a unified model is needed to contextualize embeddings derived from ''any'' tile-level foundation model. TICON addresses this need with a single, shared encoder, pretrained using a masked modeling objective to simultaneously unify and contextualize representations from diverse tile-level pathology foundation models. Our experiments demonstrate that TICON-contextualized embeddings significantly improve performance across many different tasks, establishing new state-of-the-art results on tile-level benchmarks (i.e., HEST-Bench, THUNDER, CATCH) and slide-level benchmarks (i.e., Patho-Bench). Finally, we pretrain an aggregator on TICON to form a slide-level foundation model, using only 11K WSIs, outperforming SoTA slide-level foundation models pretrained with up to 350K WSIs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mutation Prediction | Pathobench CPTAC mutation prediction | AUC72.7 | 13 | |
| Subtyping | Pathobench BRACS subtyping | Balanced Accuracy63.8 | 13 | |
| Slide-level classification | BRACS | F1 Score63.8 | 10 | |
| Slide-level classification | CPTAC | F1 Score72.7 | 10 | |
| Tile-level classification | CATCH | F1 Score87.6 | 10 | |
| Tile classification | THUNDER-SPIDER 4 tasks | F1 Score89 | 6 | |
| Tile classification | THUNDER ORIGINAL 12 tasks | F1 Score0.821 | 6 |