PMT: Plain Mask Transformer for Image and Video Segmentation with Frozen Vision Encoders
About
Vision Foundation Models (VFMs) pre-trained at scale enable a single frozen encoder to serve multiple downstream tasks simultaneously. Recent VFM-based encoder-only models for image and video segmentation, such as EoMT and VidEoMT, achieve competitive accuracy with remarkably low latency, yet they require finetuning the encoder, sacrificing the multi-task encoder sharing that makes VFMs practically attractive for large-scale deployment. To reconcile encoder-only simplicity and speed with frozen VFM features, we propose the Plain Mask Decoder (PMD), a fast Transformer-based segmentation decoder that operates on top of frozen VFM features. The resulting model, the Plain Mask Transformer (PMT), preserves the architectural simplicity and low latency of encoder-only designs while keeping the encoder representation unchanged and shareable. The design seamlessly applies to both image and video segmentation, inheriting the generality of the encoder-only framework. On standard image segmentation benchmarks, PMT matches the frozen-encoder state of the art while running up to ~3x faster. For video segmentation, it even performs on par with fully finetuned methods, while being up to 8x faster than state-of-the-art frozen-encoder models. Code: https://github.com/tue-mps/pmt.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Video Instance Segmentation | YouTube-VIS 2019 (val) | AP69.2 | 604 | |
| Video Instance Segmentation | YouTube-VIS 2021 (val) | AP64.3 | 356 | |
| Video Semantic Segmentation | VSPW (val) | mIoU65.7 | 121 | |
| Video Panoptic Segmentation | VIPSeg (val) | VPQ55.5 | 83 | |
| Semantic segmentation | ADE20K 2016 (val) | mIoU58.5 | 10 |