Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Detecting and Recognizing Human-Object Interactions

About

To understand the visual world, a machine must not only recognize individual object instances but also how they interact. Humans are often at the center of such interactions and detecting human-object interactions is an important practical and scientific problem. In this paper, we address the task of detecting <human, verb, object> triplets in challenging everyday photos. We propose a novel model that is driven by a human-centric approach. Our hypothesis is that the appearance of a person -- their pose, clothing, action -- is a powerful cue for localizing the objects they are interacting with. To exploit this cue, our model learns to predict an action-specific density over target object locations based on the appearance of a detected person. Our model also jointly learns to detect people and objects, and by fusing these predictions it efficiently infers interaction triplets in a clean, jointly trained end-to-end system we call InteractNet. We validate our approach on the recently introduced Verbs in COCO (V-COCO) and HICO-DET datasets, where we show quantitatively compelling results.

Georgia Gkioxari, Ross Girshick, Piotr Doll\'ar, Kaiming He• 2017

Related benchmarks

TaskDatasetResultRank
Human-Object Interaction DetectionHICO-DET (test)
mAP (full)9.94
493
Human-Object Interaction DetectionV-COCO (test)
AP (Role, Scenario 1)40
270
Human-Object Interaction DetectionHICO-DET
mAP (Full)9.94
233
Human-Object Interaction DetectionV-COCO 1.0 (test)
AP_role (#1)40
76
Human-Object Interaction DetectionV-COCO
AP^1 Role40
65
HOI DetectionV-COCO
AP Role 140
40
HOI DetectionHICO-DET
mAP (Rare)7.16
34
HOI DetectionHICO-DET (test)
Box mAP (Full)9.94
32
Human-Object Interaction DetectionV-COCO
Box mAP (Scenario 1)40
32
HOI DetectionV-COCO v1 (test)
AP Role (Scenario 1)40
25
Showing 10 of 16 rows

Other info

Follow for update