Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

OV-Stitcher: A Global Context-Aware Framework for Training-Free Open-Vocabulary Semantic Segmentation

About

Training-free open-vocabulary semantic segmentation(TF-OVSS) has recently attracted attention for its ability to perform dense prediction by leveraging the pretrained knowledge of large vision and vision-language models, without requiring additional training. However, due to the limited input resolution of these pretrained encoders, existing TF-OVSS methods commonly adopt a sliding-window strategy that processes cropped sub-images independently. While effective for managing high-resolution inputs, this approach prevents global attention over the full image, leading to fragmented feature representations and limited contextual reasoning. We propose OV-Stitcher, a training-free framework that addresses this limitation by stitching fragmented sub-image features directly within the final encoder block. By reconstructing attention representations from fragmented sub-image features, OV-Stitcher enables global attention within the final encoder block, producing coherent context aggregation and spatially consistent, semantically aligned segmentation maps. Extensive evaluations across eight benchmarks demonstrate that OV-Stitcher establishes a scalable and effective solution for open-vocabulary segmentation, achieving a notable improvement in mean Intersection over Union(mIoU) from 48.7 to 50.7 compared with prior training-free baselines.

Seungjae Moon, Seunghyun Oh, Youngmin Ro• 2026

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K
mIoU27.8
366
Semantic segmentationPC-59
mIoU49.1
148
Semantic segmentationPascal Context 60
mIoU43.9
139
Semantic segmentationVOC-20
mIoU90.2
118
Semantic segmentationStuff
mIoU32.1
50
Semantic segmentationObject
mIoU46.5
50
Semantic segmentationCity*
mIoU52.3
43
Semantic segmentationVOC 21 (val)
mIoU76.4
28
Semantic segmentationAverage Overall
mIoU51.9
28
Showing 9 of 9 rows

Other info

Follow for update