Social Fabric: Tubelet Compositions for Video Relation Detection
About
This paper strives to classify and detect the relationship between object tubelets appearing within a video as a <subject-predicate-object> triplet. Where existing works treat object proposals or tubelets as single entities and model their relations a posteriori, we propose to classify and detect predicates for pairs of object tubelets a priori. We also propose Social Fabric: an encoding that represents a pair of object tubelets as a composition of interaction primitives. These primitives are learned over all relations, resulting in a compact representation able to localize and classify relations from the pool of co-occurring object tubelets across all timespans in a video. The encoding enables our two-stage network. In the first stage, we train Social Fabric to suggest proposals that are likely interacting. We use the Social Fabric in the second stage to simultaneously fine-tune and predict predicate labels for the tubelets. Experiments demonstrate the benefit of early video relation modeling, our encoding and the two-stage architecture, leading to a new state-of-the-art on two benchmarks. We also show how the encoding enables query-by-primitive-example to search for spatio-temporal video relations. Code: https://github.com/shanshuo/Social-Fabric.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Relation Detection | VRD (test) | R@5013.73 | 75 | |
| Relation Detection | VidVRD v1.0 (test) | R@5013.73 | 18 | |
| Relation Tagging | VidVRD v1.0 (test) | P@549.2 | 18 | |
| Relation Detection | VidOR SOTA comparisons (val) | mAP11.21 | 14 | |
| Relation Tagging | VidOR SOTA comparisons (val) | P@168.86 | 14 | |
| Visual Relation Tagging | VidOR (val) | P@555.16 | 14 | |
| Relation Tagging | VidVRD (test) | P@162.5 | 14 | |
| Visual Relation Detection | VidOR (val) | R@500.0999 | 13 | |
| Relation Tagging | ImageNet-VidVRD (test) | P@162.5 | 9 | |
| Relation Detection | ImageNet-VidVRD (test) | mAP20.08 | 8 |