Temporally Consistent Referring Video Object Segmentation with Hybrid Memory
About
Referring Video Object Segmentation (R-VOS) methods face challenges in maintaining consistent object segmentation due to temporal context variability and the presence of other visually similar objects. We propose an end-to-end R-VOS paradigm that explicitly models temporal instance consistency alongside the referring segmentation. Specifically, we introduce a novel hybrid memory that facilitates inter-frame collaboration for robust spatio-temporal matching and propagation. Features of frames with automatically generated high-quality reference masks are propagated to segment the remaining frames based on multi-granularity association to achieve temporally consistent R-VOS. Furthermore, we propose a new Mask Consistency Score (MCS) metric to evaluate the temporal consistency of video segmentation. Extensive experiments demonstrate that our approach enhances temporal consistency by a significant margin, leading to top-ranked performance on popular R-VOS benchmarks, i.e., Ref-YouTube-VOS (67.1%) and Ref-DAVIS17 (65.6%). The code is available at https://github.com/bo-miao/HTR.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Referring Video Object Segmentation | Ref-YouTube-VOS (val) | J&F Score67.1 | 200 | |
| Referring Video Object Segmentation | Ref-DAVIS 2017 (val) | J&F65.6 | 178 | |
| Referring Video Object Segmentation | Ref-DAVIS 17 | J&F Score65.6 | 131 | |
| Referring Video Object Segmentation | MeViS (val) | J&F Score0.427 | 122 | |
| Referring Video Object Segmentation | Ref-YouTube-VOS | J&F67.5 | 85 | |
| Referring Video Object Segmentation | A2D-Sentences | oIoU80.1 | 57 | |
| Referring Video Object Segmentation | JHMDB Sentences | Overall IoU73.9 | 56 |